عکس nghorbani
Solving Optical MoCap AutomaticallyJupyter Notebook
موضوع‌ها
۵
فورک‌ها
۱۵
ستاره‌ها
۹۷
تاریخ ایجاد
۷ مهر ۱۴۰۰
آخرین بروزرسانی
۵ ماه قبل
لایسنس
Other

SOMA: Solving Optical Marker-Based MoCap Automatically, ICCV'21

This repository contains the official PyTorch implementation of:

SOMA: Solving Optical Marker-Based MoCap Automatically
Nima Ghorbani and Michael J. Black
Paper | Supp.Mat. | Video | Project website | Poster

alt text

SOMA automatically transforms raw marker-based mocap point clouds (black dots in the back) into solved SMPL-X bodies and labeled markers (colored dots).

Installation

SOMA is originally developed in Python 3.7, PyTorch 1.8.2 LTS, for Ubuntu 20.04.2 LTS. Below we prepare the python environment using Anaconda, however, we opt for a simple pip package manager for installing dependencies.

sudo apt install libatlas-base-dev
sudo apt install libpython3.7
sudo apt install libtbb2

conda create -n soma python=3.7 
conda install -c conda-forge ezc3d

pip3 install torch==1.8.2+cu102 torchvision==0.9.2+cu102 torchaudio==0.8.2 -f https://download.pytorch.org/whl/lts/1.8/torch_lts.html

ezc3d installation is currently not supported by pip.

Assuming that you have already cloned this repository to your local drive go to the root directory of SOMA code and run

pip install -r requirements.txt
python setup.py develop

Copy the precompiled smpl-fast-derivatives into your python site-packages folder, i.e. anaconda3/envs/soma/lib/python3.7/site-packages. The final directory should look like anaconda3/envs/soma/lib/python3.7/site-packages/psbody/smpl.

Install the psbody.mesh library following the instructions in https://github.com/MPI-IS/mesh. Hint: clone the mesh repository and run the following from the anaconda environment: python setup.py install .

To use the rendering capabilities first install an instance of Blender-2.83 LTS on your machine. Afterward uncompress contents of the precompiled bpy-2.83 into your python site-packages folder, i.e. anaconda3/envs/soma/lib/python3.7/site-packages.

Last but not least, the current SOMA code relies on MoSh++ mocap solver. Please install MoSh++ following the guidelines in its repository.

Using SOMA

There are multiple main parts of the codebase that we try to explain in the Tutorials:

Citation

Please cite the following paper if you use this code directly or indirectly in your research/projects:

@inproceedings{SOMA:ICCV:2021,
  title = {{SOMA}: Solving Optical Marker-Based MoCap Automatically},
  author = {Ghorbani, Nima and Black, Michael J.},
  booktitle = {Proceedings of IEEE/CVF International Conference on Computer Vision (ICCV)},
  month = oct,
  year = {2021},
  doi = {},
  month_numeric = {10}}

License

Software Copyright License for non-commercial scientific research purposes. Please read carefully the terms and conditions and any accompanying documentation before you download and/or use the SOMA data and software, (the "Data & Software"), software, scripts, and animations. By downloading and/or using the Data & Software (including downloading, cloning, installing, and any other use of this repository), you acknowledge that you have read these terms and conditions, understand them, and agree to be bound by them. If you do not agree with these terms and conditions, you must not download and/or use the Data & Software. Any infringement of the terms of this agreement will automatically terminate your rights under this License.

Contact

The code in this repository is developed by Nima Ghorbani while at Max-Planck Institute for Intelligent Systems, Tübingen, Germany.

If you have any questions you can contact us at soma@tuebingen.mpg.de.

For commercial licensing, contact ps-licensing@tue.mpg.de