Haptic Intelligence Perceiving Systems Conference Paper 2023

Reconstructing Signing Avatars from Video Using Linguistic Priors

pdf arXiv project code
Haptic Intelligence, Perceiving Systems
  • Postdoctoral Researcher
Perceiving Systems
  • Doctoral Researcher
Perceiving Systems
  • Guest Scientist
Perceiving Systems
Perceiving Systems
  • Guest Scientist
Haptic Intelligence
Director
Perceiving Systems
Emeritus / Acting Director

Sign language (SL) is the primary method of communication for the 70 million Deaf people around the world. Video dictionaries of isolated signs are a core SL learning tool. Replacing these with 3D avatars can aid learning and enable AR/VR applications, improving access to technology and online media. However, little work has attempted to estimate expressive 3D avatars from SL video; occlusion, noise, and motion blur make this task difficult. We address this by introducing novel linguistic priors that are universally applicable to SL and provide constraints on 3D hand pose that help resolve ambiguities within isolated signs. Our method, SGNify, captures fine-grained hand pose, facial expression, and body movement fully automatically from in-the-wild monocular SL videos. We evaluate SGNify quantitatively by using a commercial motion-capture system to compute 3D avatars synchronized with monocular video. SGNify outperforms state-of-the-art 3D body-pose- and shape-estimation methods on SL videos. A perceptual study shows that SGNify's 3D reconstructions are significantly more comprehensible and natural than those of previous methods and are on par with the source videos. Code and data are available at sgnify.is.tue.mpg.de.

Author(s): Forte, Maria-Paola and Kulits, Peter and Huang, Chun-Hao and Choutas, Vasileios and Tzionas, Dimitrios and Kuchenbecker, Katherine J. and Black, Michael J.
Links:
Book Title: IEEE/CVF Conf. on Computer Vision and Pattern Recognition (CVPR)
Pages: 12791--12801
Year: 2023
Month: June
Project(s):
BibTeX Type: Conference Paper (inproceedings)
Address: Vancouver, Canada
DOI: 10.1109/CVPR52729.2023.01230
Event Name: CVPR 2023
State: Published
URL: https://sgnify.is.tue.mpg.de/
Electronic Archiving: grant_archive

BibTeX

@inproceedings{Forte23-CVPR-SGNify,
  title = {Reconstructing Signing Avatars from Video Using Linguistic Priors},
  booktitle = {IEEE/CVF Conf.~on Computer Vision and Pattern Recognition (CVPR)},
  abstract = {Sign language (SL) is the primary method of communication for the 70 million Deaf people around the world. Video dictionaries of isolated signs are a core SL learning tool. Replacing these with 3D avatars can aid learning and enable AR/VR applications, improving access to technology and online media. However, little work has attempted to estimate expressive 3D avatars from SL video; occlusion, noise, and motion blur make this task difficult. We address this by introducing novel linguistic priors that are universally applicable to SL and provide constraints on 3D hand pose that help resolve ambiguities within isolated signs. Our method, SGNify, captures fine-grained hand pose, facial expression, and body movement fully automatically from in-the-wild monocular SL videos. We evaluate SGNify quantitatively by using a commercial motion-capture system to compute 3D avatars synchronized with monocular video. SGNify outperforms state-of-the-art 3D body-pose- and shape-estimation methods on SL videos. A perceptual study shows that SGNify's 3D reconstructions are significantly more comprehensible and natural than those of previous methods and are on par with the source videos. Code and data are available at sgnify.is.tue.mpg.de.},
  pages = {12791--12801},
  address = {Vancouver, Canada},
  month = jun,
  year = {2023},
  author = {Forte, Maria-Paola and Kulits, Peter and Huang, Chun-Hao and Choutas, Vasileios and Tzionas, Dimitrios and Kuchenbecker, Katherine J. and Black, Michael J.},
  doi = {10.1109/CVPR52729.2023.01230},
  url = {https://sgnify.is.tue.mpg.de/},
  month_numeric = {6}
}