Back

Perceiving Systems Members Publications

Humans from Video

Top: VIBE regresses 3D human pose and shape from video using adversarial training by leveraging a large-scale human motion dataset (AMASS) to train a motion discriminator. Bottom left: output of VIBE. Bottom middle: SMIL estimates infant shape and motion from RGB-D videos to detect cerebral palsy. Bottom right: The 3DPW dataset combines IMU data with video to obtain high-quality pseudo ground truth 3D humans in video.

Members

Perceiving Systems
Emeritus / Acting Director
Perceiving Systems
  • Guest Scientist
Perceiving Systems
  • Guest Scientist
Perceiving Systems
  • Guest Scientist
Perceiving Systems
Perceiving Systems
Affiliated Researcher
Perceiving Systems
Perceiving Systems
  • Research Group Leader
Perceiving Systems
Affiliated Researcher
Perceiving Systems
Perceiving Systems
Guest Scientist
Perceiving Systems
Affiliated Researcher

Publications

Perceiving Systems Conference Paper VIBE: Video Inference for Human Body Pose and Shape Estimation Kocabas, M., Athanasiou, N., Black, M. J. In 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR 2020), 5252-5262, IEEE, Piscataway, NJ, IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR 2020), June 2020 (Published)
Human motion is fundamental to understanding behavior. Despite progress on single-image 3D pose and shape estimation, existing video-based state-of-the-art methodsfail to produce accurate and natural motion sequences due to a lack of ground-truth 3D motion data for training. To address this problem, we propose “Video Inference for Body Pose and Shape Estimation” (VIBE), which makes use of an existing large-scale motion capture dataset (AMASS) together with unpaired, in-the-wild, 2D keypoint annotations. Our key novelty is an adversarial learning framework that leverages AMASS to discriminate between real human motions and those produced by our temporal pose and shape regression networks. We define a temporal network architecture and show that adversarial training, at the sequence level, produces kinematically plausible motion sequences without in-the-wild ground-truth 3D labels. We perform extensive experimentation to analyze the importance of motion and demonstrate the effectiveness of VIBE on challenging 3D pose estimation datasets, achieving state-of-the-art performance. Code and pretrained models are available at https://github.com/mkocabas/VIBE
arXiv code video supplemental video pdf DOI BibTeX

Perceiving Systems Conference Paper Towards Accurate Marker-less Human Shape and Pose Estimation over Time Huang, Y., Bogo, F., Lassner, C., Kanazawa, A., Gehler, P. V., Romero, J., Akhter, I., Black, M. J. In International Conference on 3D Vision (3DV), 421-430, 2017
Existing markerless motion capture methods often assume known backgrounds, static cameras, and sequence specific motion priors, limiting their application scenarios. Here we present a fully automatic method that, given multiview videos, estimates 3D human pose and body shape. We take the recently proposed SMPLify method [12] as the base method and extend it in several ways. First we fit a 3D human body model to 2D features detected in multi-view images. Second, we use a CNN method to segment the person in each image and fit the 3D body model to the contours, further improving accuracy. Third we utilize a generic and robust DCT temporal prior to handle the left and right side swapping issue sometimes introduced by the 2D pose estimator. Validation on standard benchmarks shows our results are comparable to the state of the art and also provide a realistic 3D shape avatar. We also demonstrate accurate results on HumanEva and on challenging monocular sequences of dancing from YouTube.
Code pdf DOI BibTeX

Perceiving Systems Conference Paper Multi-Person Tracking by Multicuts and Deep Matching Tang, S., Andres, B., Andriluka, M., Schiele, B. ECCV Workshop on Benchmarking Mutliple Object Tracking, 2016 PDF BibTeX