Back

Autonomous Learning Members Publications

Representation Learning

A comparison of the Eigenbasis retrieved through PCA and a non-linear VAE architecture on a synthetic task.

Members

Empirical Inference, Autonomous Learning
Senior Research Scientist
Autonomous Learning
Empirical Inference
  • Guest Scientist

Publications

Autonomous Learning Conference Paper Demystifying Inductive Biases for (Beta-)VAE Based Architectures Zietlow, D., Rolinek, M., Martius, G. In Proceedings of the 2021 International Conference on Machine Learning (ICML), 139:12945-12954, Proceedings of Machine Learning Research , The 38th International Conference on Machine Learning (ICML 2021), July 2021 (Published)
The performance of Beta-Variational-Autoencoders and their variants on learning semantically meaningful, disentangled representations is unparalleled. On the other hand, there are theoretical arguments suggesting the impossibility of unsupervised disentanglement. In this work, we shed light on the inductive bias responsible for the success of VAE-based architectures. We show that in classical datasets the structure of variance, induced by the generating factors, is conveniently aligned with the latent directions fostered by the VAE objective. This builds the pivotal bias on which the disentangling abilities of VAEs rely. By small, elaborate perturbations of existing datasets, we hide the convenient correlation structure that is easily exploited by a variety of architectures. To demonstrate this, we construct modified versions of standard datasets in which (i) the generative factors are perfectly preserved; (ii) each image undergoes a mild transformation causing a small change of variance; (iii) the leading VAE-based disentanglement architectures fail to produce disentangled representations whilst the performance of a non-variational method remains unchanged.
Arxiv PDF Paper @ ICML 2021 (spotlight video) URL BibTeX

Autonomous Learning Conference Paper Variational Autoencoders Pursue PCA Directions (by Accident) Rolinek, M., Zietlow, D., Martius, G. In Proceedings IEEE Conf. on Computer Vision and Pattern Recognition (CVPR), 12406-12415, IEEE International Conference on Computer Vision and Pattern Recognition (CVPR), June 2019
The Variational Autoencoder (VAE) is a powerful architecture capable of representation learning and generative modeling. When it comes to learning interpretable (disentangled) representations, VAE and its variants show unparalleled performance. However, the reasons for this are unclear, since a very particular alignment of the latent embedding is needed but the design of the VAE does not encourage it in any explicit way. We address this matter and offer the following explanation: the diagonal approximation in the encoder together with the inherent stochasticity force local orthogonality of the decoder. The local behavior of promoting both reconstruction and orthogonality matches closely how the PCA embedding is chosen. Alongside providing an intuitive understanding, we justify the statement with full theoretical analysis as well as with experiments.
Arxiv URL BibTeX