Skip to content

PyTorch implementation for the paper GanHand: Predicting Human Grasp Affordances in Multi-Object Scenes (CVPR 2020 Oral)

Notifications You must be signed in to change notification settings

zhangxuelei86/GanHand

 
 

Repository files navigation

GanHand

[Project] [Paper] [Dataset]

Dataset

Checkout the github repository to download the YCB-Affordance dataset. This contains the 3D models of objects from the YCB benchmark, the videos from the YCB-Video Dataset, and the human hand grasps from the YCB-Affordance dataset.

Requirements

  • conda create -n ganhand python=3.6
  • conda activate ganhand
  • Python requirements: Run pip install -r requirements.txt.
  • MANO layer: Follow instructions from the MANO layer project in here.

Data

Model

GanHand takes a single RGB image of one or several objects and predicts how a human would grasp these objects naturally. Our architecture consists of three stages. First, the objects' shapes and locations are estimated in the scene using an object 6D pose estimator or a reconstruction network (red). The predicted shape is then projected onto the image plane to obtaina segmentation mask that is concatenated with the input image and fed to the second sub-network for grasp prediction (blue). Finally, werefine the hand parameters and obtain hand final shapes and poses using a differentiable parametric model MANO (yellow). The model is trained using adversarial, interpenetration, classification and optimization losses, indicated in bold.

Test

Videos like those in the teaser of the paper can be obtained running the following command. The pretrained model can be downloaded from this link and placed under a folder named checkpoints. So the main folder should contain the model checkpoints in /checkpoints/ganhand_pretrained/

python test.py --dataset_mode ycb_affordances_complete_scene --name ganhand_pretrained --load_epoch 13

Acknowledgements

Citation

If this dataset is useful in your research, please cite:

@inproceedings{corona2020ganhand,
  title={Ganhand: Predicting human grasp affordances in multi-object scenes},
  author={Corona, Enric and Pumarola, Albert and Alenya, Guillem and Moreno-Noguer, Francesc and Rogez, Gr{\'e}gory},
  booktitle={Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition},
  pages={5031--5041},
  year={2020}
}

License

The YCB-Affordance dataset is released only for research purposes

About

PyTorch implementation for the paper GanHand: Predicting Human Grasp Affordances in Multi-Object Scenes (CVPR 2020 Oral)

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • Python 100.0%