Autonomous Vision Conference Paper 2020

Towards Unsupervised Learning of Generative Models for 3D Controllable Image Synthesis

pdf suppmat Video 2 Project Page Video 1 Slides Poster
Thumb ticker sm screen shot 2016 10 03 at 7.50.00 pm
Autonomous Vision
no image
Autonomous Vision
Thumb ticker sm photosmall
Autonomous Vision
Thumb ticker sm upper body tiny
Autonomous Vision, Perceiving Systems
Guest Scientist
Yiyicvpr

In recent years, Generative Adversarial Networks have achieved impressive results in photorealistic image synthesis. This progress nurtures hopes that one day the classical rendering pipeline can be replaced by efficient models that are learned directly from images. However, current image synthesis models operate in the 2D domain where disentangling 3D properties such as camera viewpoint or object pose is challenging. Furthermore, they lack an interpretable and controllable representation. Our key hypothesis is that the image generation process should be modeled in 3D space as the physical world surrounding us is intrinsically three-dimensional. We define the new task of 3D controllable image synthesis and propose an approach for solving it by reasoning both in 3D space and in the 2D image domain. We demonstrate that our model is able to disentangle latent 3D factors of simple multi-object scenes in an unsupervised fashion from raw images. Compared to pure 2D baselines, it allows for synthesizing scenes that are consistent wrt. changes in viewpoint or object pose. We further evaluate various 3D representations in terms of their usefulness for this challenging task.

Author(s): Yiyi Liao and Katja Schwarz and Lars Mescheder and Andreas Geiger
Links:
Book Title: Proceedings IEEE Conf. on Computer Vision and Pattern Recognition (CVPR)
Pages: 5870 -- 5879
Year: 2020
Publisher: IEEE
Bibtex Type: Conference Paper (inproceedings)
Address: Piscataway, NJ
DOI: 10.1109/CVPR42600.2020.00591
Event Name: IEEE International Conference on Computer Vision and Pattern Recognition (CVPR) 2020
Event Place: Seattle, USA
State: Published
Electronic Archiving: grant_archive

BibTex

@inproceedings{Liao2020CVPR,
  title = {Towards Unsupervised Learning of Generative Models for 3D Controllable Image Synthesis},
  booktitle = { Proceedings IEEE Conf. on Computer Vision and Pattern Recognition (CVPR)},
  abstract = {In recent years, Generative Adversarial Networks have achieved impressive results in photorealistic image synthesis. This progress nurtures hopes that one day the classical rendering pipeline can be replaced by efficient models that are learned directly from images. However, current image synthesis models operate in the 2D domain where disentangling 3D properties such as camera viewpoint or object pose is challenging. Furthermore, they lack an interpretable and controllable representation. Our key hypothesis is that the image generation process should be modeled in 3D space as the physical world surrounding us is intrinsically three-dimensional. We define the new task of 3D controllable image synthesis and propose an approach for solving it by reasoning both in 3D space and in the 2D image domain. We demonstrate that our model is able to disentangle latent 3D factors of simple multi-object scenes in an unsupervised fashion from raw images. Compared to pure 2D baselines, it allows for synthesizing scenes that are consistent wrt. changes in viewpoint or object pose. We further evaluate various 3D representations in terms of their usefulness for this challenging task.},
  pages = {5870 -- 5879},
  publisher = {IEEE},
  address = {Piscataway, NJ},
  year = {2020},
  slug = {liao2020cvpr},
  author = {Liao, Yiyi and Schwarz, Katja and Mescheder, Lars and Geiger, Andreas}
}