This package contains code for the Deep Multitask Human Sensing (DMHS) method, published in the CVPR 2017 paper Deep Multitask Architecture for Integrated 2D and 3D Human Sensing.
By using the software, you are agreeing to the terms of the license agreement.
Our software is built on top of the Caffe deep learning library used by the Convolutional Pose Machines method. The current version was developed by:
Alin-Ionut Popa , Mihai Zanfir and Cristian Sminchisescu
We provide a deep multitask architecture for fully automatic 2d and 3d human sensing (DMHS), including recognition and reconstruction, in monocular images. The system computes the figure-ground segmentation, semantically identifies the human body parts at pixel level, and estimates the 2d and 3d pose of the person. This software allows you to test our algorithm on your own images.
If you use this code/model for your research, please cite the following paper:
@inproceedings{dmhs_cvpr17,
author = {Alin-Ionut Popa and Mihai Zanfir and Cristian Sminchisescu},
title = {Deep Multitask Architecture for Integrated 2D and 3D Human Sensing},
booktitle = {IEEE International Conference on Computer Vision and Pattern Recognition},
year = {2017}
}
First, clone the project by running:
git clone --recursive https://github.com/alinionutpopa/dmhs.git
You need to compile the modified Caffe library in this repository. Instructions for Ubuntu 14.04 are included below. You can also consult the generic Caffe installation guide for further help.
sudo apt-get install libprotobuf-dev libleveldb-dev libsnappy-dev libopencv-dev libhdf5-serial-dev protobuf-compiler
sudo apt-get install --no-install-recommends libboost-all-dev
Install the correct CUDA driver and its SDK. Download CUDA SDK from Nvidia website.
You might need to blacklist some modules so that they do not interfere with the driver installation. You also need to uninstall your default Nvidia Driver first.
sudo apt-get install freeglut3-dev build-essential libx11-dev libxmu-dev libxi-dev libgl1-mesa-glx libglu1-mesa libglu1-mesa-dev
Open /etc/modprobe.d/blacklist.conf and add:
blacklist amd76x_edac
blacklist vga16fb
blacklist nouveau
blacklist rivafb
blacklist nvidiafb
blacklist rivatv
sudo apt-get remove --purge nvidia*
When you restart your PC, before loging in, try "Ctrl + Alt + F1" to switch to a text-based login. Try:
sudo service lightdm stop
chmod +x cuda*.run
sudo ./cuda*.run
Install a BLAS library such as ATLAS, OpenBLAS or MKL. To install BLAS:
sudo apt-get install libatlas-base-dev
Install Anaconda Python distribution or install the default Python distribution with numpy, scipy, etc.
Install MATLAB using a standard distribution.
Set the path correctly in the Makefile.config
. You can rename the Makefile.config.example
to Makefile.config
, as most common parts are filled already. You may need to change it a bit according to your environment.
After this, in Ubuntu 14.04, try:
make
If there are no error messages, you can then compile and install the Python and Matlab wrappers: To install the MATLAB wrapper (at the moment only for MATLAB versions prior to 2017):
make matcaffe
All done! Try our method!
First download the model that includes the trained weights from this link into the model
folder. Also, change the caffepath
variable from code/config.m
file accordingly.
The MATLAB script for running the demo on all three tasks (i.e. 2D pose estimation, body part labeling and 3D pose estimation) is demoDMHS.m
. The MATLAB script for running the demo on the 3D pose estimation task alone (much faster than demoDMHS.m
) is demoDMHS_3D.m
. Change the displayMode
variable to 1
in order to visualize the results.
Please note that our method requires an image cropped around the bounding box of the person. Also, please validate the scales for the 3D pose estimation and body part labeling tasks for other datasets than Human3.6M.
Contact: alin.popa@imar.ro , mihai.zanfir@imar.ro , cristian.sminchisescu@math.lth.se
This work was supported in part by the European Research Council Consolidator grant SEED, CNCS-UEFISCDI under JRP-RO-FR-2014-16, the EU Horizon 2020 Grant #688835 (DE-ENIGMA), and SSF.