Back

Perceiving Systems Publications

ARCTIC: A Dataset for Dexterous Bimanual Hand-Object Manipulation

ARCTIC [File Icon] includes highly accurate full-body motion capture together with the motion of articulated 3D objects. The data is registered to both allocentric and egocentric cameras and includes accurate contact labels on the body and objects.

Publications

Perceiving Systems Conference Paper ARCTIC: A Dataset for Dexterous Bimanual Hand-Object Manipulation Fan, Z., Taheri, O., Tzionas, D., Kocabas, M., Kaufmann, M., Black, M. J., Hilliges, O. In IEEE/CVF Conf. on Computer Vision and Pattern Recognition (CVPR), 12943-12954, CVPR, June 2023 (Published)
Humans intuitively understand that inanimate objects do not move by themselves, but that state changes are typically caused by human manipulation (e.g., the opening of a book). This is not yet the case for machines. In part this is because there exist no datasets with ground-truth 3D annotations for the study of physically consistent and synchronised motion of hands and articulated objects. To this end, we introduce ARCTIC -- a dataset of two hands that dexterously manipulate objects, containing 2.1M video frames paired with accurate 3D hand and object meshes and detailed, dynamic contact information. It contains bi-manual articulation of objects such as scissors or laptops, where hand poses and object states evolve jointly in time. We propose two novel articulated hand-object interaction tasks: (1) Consistent motion reconstruction: Given a monocular video, the goal is to reconstruct two hands and articulated objects in 3D, so that their motions are spatio-temporally consistent. (2) Interaction field estimation: Dense relative hand-object distances must be estimated from images. We introduce two baselines ArcticNet and InterField, respectively and evaluate them qualitatively and quantitatively on ARCTIC.
Project Page Code Paper arXiv Video DOI URL BibTeX