Skip to content

Latest commit

 

History

History
142 lines (101 loc) · 4.59 KB

README.md

File metadata and controls

142 lines (101 loc) · 4.59 KB

Dynamic Inertial Poser (DynaIP): Part-Based Motion Dynamics Learning for Enhanced Human Pose Estimation with Sparse Inertial Sensors

Yu ZhangSongpengcheng XiaLei Chu*  Jiarui YangQi WuLing Pei*
Shanghai Jiao Tong University
Equal contribution   *Corresponding author
🤩 Accepted to CVPR 2024

This is the pytorch implementation of our paper DynaIP at CVPR 2024.

Environment Setup

We tested our code on Windows with Python 3.8.15, Pytorch 1.10.2 with cuda11.1, other dependencies are specified in requirements.txt.

conda create -n dynaip python==3.8.15
conda activate dynaip
pip install torch==1.10.2+cu111 torchvision==0.11.3+cu111 torchaudio==0.10.2 -f https://download.pytorch.org/whl/cu111/torch_stable.html
pip install -r requirements.txt

Datasets and Models

Datasets

We used publicly available Xsens Mocap datasets and DIP-IMU to train and evaluate our model. These Xsens datasets include AnDy, UNIPD-BPE, Emokine, CIP and Virginia Natural Motion. You can download the raw data from:

Clone this repo, download the above datasets, extract and place them in ./datasets/raw/.

datasets
├─extract
├─raw
│  ├─andy
│  ├─cip
│  ├─dip
│  ├─emokine
│  ├─unipd
│  └─virginia
└─work

SMPL Models

We used the smpl model, download it from here and place in ./smpl_models/.

smpl_models
    smpl_female.pkl
    smpl_male.pkl
    smpl_neutral.pkl

Data Processing

  1. Run extract.py , this will extract imu and pose data from raw .mvnx files, downsampling them to 60Hz. After data extraction, you can use scripts in extract.py for visualization.
python ./datasets/extract.py

Note that Virginia Natural Motion has pose drifts due to long time tracking, we visualized part of its sequence and manually selected clean frames as training and evaluation data, those selected frames are also stored in extract.py.

  1. Run process.py to preprocess IMU data from extracted Xsens datasets and raw DIP-IMU.
python ./datasets/process.py

Since DIP-IMU has no root trajectory, we generate pseudo-root trajectory by forcing the lowest foot to touch the ground. The training and test split information of each dataset is stored in ./datasets/split_info/.

Training and Evaluation

For training and evaluation, simply run:

python train.py
python evaluation.py

The pretrained weights are stored in ./weights folder.

Visualization

We use aitviewer for visualization, run:

python vis.py

for visualizing the predicted results.

Acknowledgement

Some of our codes are adapted from PIP and VT-Natural-Motion-Processing.

Citation

If you find this project helpful, please consider citing us:

@article{zhang2023dynamic,
  title={Dynamic Inertial Poser (DynaIP): Part-Based Motion Dynamics Learning for Enhanced Human Pose Estimation with Sparse Inertial Sensors},
  author={Zhang, Yu and Xia, Songpengcheng and Chu, Lei and Yang, Jiarui and Wu, Qi and Pei, Ling},
  journal={ IEEE / CVF Computer Vision and Pattern Recognition Conference (CVPR)},
  year={2024},
  publisher={IEEE},
  booktitle={cvpr}
}