Learning Joint Reconstruction of Hands and Manipulated Objects

  • Yana Hasson ,
  • Gul Varol ,
  • Dimitrios Tzionas ,
  • Igor Kalevatykh ,
  • Michael J. Black ,
  • Ivan Laptev ,
  • Cordelia Schmid

2019 Computer Vision and Pattern Recognition |

Published by IEEE

PDF | PDF | PDF | Publication | Publication | Publication | Publication | Publication | Publication

Estimating hand-object manipulations is essential for in- terpreting and imitating human actions. Previous work has made significant progress towards reconstruction of hand poses and object shapes in isolation. Yet, reconstructing hands and objects during manipulation is a more challeng- ing task due to significant occlusions of both the hand and object. While presenting challenges, manipulations may also simplify the problem since the physics of contact re- stricts the space of valid hand-object configurations. For example, during manipulation, the hand and object should be in contact but not interpenetrate. In this work, we regu- larize the joint reconstruction of hands and objects with ma- nipulation constraints. We present an end-to-end learnable model that exploits a novel contact loss that favors phys- ically plausible hand-object constellations. Our approach improves grasp quality metrics over baselines, using RGB images as input. To train and evaluate the model, we also propose a new large-scale synthetic dataset, ObMan, with hand-object manipulations. We demonstrate the transfer- ability of ObMan-trained models to real data.