Perceiving Systems Conference Paper 2019

Learning Joint Reconstruction of Hands and Manipulated Objects

Obman new

Estimating hand-object manipulations is essential for interpreting and imitating human actions. Previous work has made significant progress towards reconstruction of hand poses and object shapes in isolation. Yet, reconstructing hands and objects during manipulation is a more challenging task due to significant occlusions of both the hand and object. While presenting challenges, manipulations may also simplify the problem since the physics of contact restricts the space of valid hand-object configurations. For example, during manipulation, the hand and object should be in contact but not interpenetrate. In this work, we regularize the joint reconstruction of hands and objects with manipulation constraints. We present an end-to-end learnable model that exploits a novel contact loss that favors physically plausible hand-object constellations. Our approach improves grasp quality metrics over baselines, using RGB images as input. To train and evaluate the model, we also propose a new large-scale synthetic dataset, ObMan, with hand-object manipulations. We demonstrate the transferability of ObMan-trained models to real data.

Author(s): Yana Hasson and Gül Varol and Dimitrios Tzionas and Igor Kalevatykh and Michael J. Black and Ivan Laptev and Cordelia Schmid
Book Title: IEEE/CVF Conf. on Computer Vision and Pattern Recognition (CVPR)
Pages: 11807--11816
Year: 2019
Month: June
Project(s):
Bibtex Type: Conference Paper (inproceedings)
DOI: 10.1109/CVPR.2019.01208
Event Name: IEEE International Conference on Computer Vision and Pattern Recognition (CVPR) 2019
Event Place: Long Beach, USA
URL: https://www.di.ens.fr/willow/research/obman
Electronic Archiving: grant_archive
Attachments:

BibTex

@inproceedings{Hasson:CVPR:2019,
  title = {Learning Joint Reconstruction of Hands and Manipulated Objects},
  booktitle = {IEEE/CVF Conf.~on Computer Vision and Pattern Recognition (CVPR) },
  abstract = {Estimating hand-object manipulations is essential for interpreting and imitating human actions. Previous work has made significant progress towards reconstruction of hand poses and object shapes in isolation. Yet, reconstructing hands and objects during manipulation is a more challenging task due to significant occlusions of both the hand and object. While presenting challenges, manipulations may also simplify the problem since the physics of contact restricts the space of valid hand-object configurations. For example, during manipulation, the hand and object should be in contact but not interpenetrate. In this work, we regularize the joint reconstruction of hands and objects with manipulation constraints. We present an end-to-end learnable model that exploits a novel contact loss that favors physically plausible hand-object constellations. Our approach improves grasp quality metrics over baselines, using RGB images as input. To train and evaluate the model, we also propose a new large-scale synthetic dataset, ObMan, with hand-object manipulations. We demonstrate the transferability of ObMan-trained models to real data.},
  pages = {11807--11816},
  month = jun,
  year = {2019},
  slug = {hasson-cvpr-2019},
  author = {Hasson, Yana and Varol, G{\"u}l and Tzionas, Dimitrios and Kalevatykh, Igor and Black, Michael J. and Laptev, Ivan and Schmid, Cordelia},
  url = {https://www.di.ens.fr/willow/research/obman},
  month_numeric = {6}
}