InterCap: Joint Markerless 3D Tracking of Humans and Objects in Interaction

Open Access
Authors
Publication date 2022
Host editors
  • B. Andres
  • F. Bernard
  • D. Cremers
  • S. Frintrop
  • B. Goldlücke
  • I. Ihrke
Book title Pattern Recognition
Book subtitle 44th DAGM German Conference, DAGM GCPR 2022, Konstanz, Germany, September 27–30, 2022 : proceedings
ISBN
  • 9783031167874
ISBN (electronic)
  • 9783031167881
Series Lecture Notes in Computer Science
Event Pattern Recognition: 44th DAGM German Conference
Pages (from-to) 281-299
Publisher Cham: Springer
Organisations
  • Faculty of Science (FNWI) - Informatics Institute (IVI)
Abstract
Humans constantly interact with daily objects to accomplish tasks. To understand such interactions, computers need to reconstruct these from cameras observing whole-body interaction with scenes. This is challenging due to occlusion between the body and objects, motion blur, depth/scale ambiguities, and the low image resolution of hands and graspable object parts. To make the problem tractable, the community focuses either on interacting hands, ignoring the body, or on interacting bodies, ignoring hands. The GRAB dataset addresses dexterous whole-body interaction but uses marker-based MoCap and lacks images, while BEHAVE captures video of body object interaction but lacks hand detail. We address the limitations of prior work with InterCap, a novel method that reconstructs interacting whole-bodies and objects from multi-view RGB-D data, using the parametric whole-body model SMPL-X and known object meshes. To tackle the above challenges, InterCap uses two key observations: (i) Contact between the hand and object can be used to improve the pose estimation of both. (ii) Azure Kinect sensors allow us to set up a simple multi-view RGB-D capture system that minimizes the effect of occlusion while providing reasonable inter-camera synchronization. With this method we capture the InterCap dataset, which contains 10 subjects (5 males and 5 females) interacting with 10 objects of various sizes and affordances, including contact with the hands or feet. In total, InterCap has 223 RGB-D videos, resulting in 67,357 multi-view frames, each containing 6 RGB-D images. Our method provides pseudo ground-truth body meshes and objects for each video frame. Our InterCap method and dataset fill an important gap in the literature and support many research directions. Our data and code are available for research purposes.
Document type Conference contribution
Note With supplementary material
Language English
Published at https://doi.org/10.1007/978-3-031-16788-1_18
Other links https://intercap.is.tue.mpg.de
Downloads
InterCap paper (Accepted author manuscript)
Permalink to this page
Back