ON INCORPORATING INDUCTIVE BIASES INTO VAES

Ning Miao*, Emile Mathieu, N. Siddharth, Yee Whye Teh, Tom Rainforth*

*Corresponding author for this work

Research output: Chapters, Conference Papers, Creative and Literary WorksRGC 32 - Refereed conference paper (with host publication)peer-review

1 Citation (Scopus)

Abstract

We explain why directly changing the prior can be a surprisingly ineffective mechanism for incorporating inductive biases into variational auto-encoders (VAEs), and introduce a simple and effective alternative approach: Intermediary Latent Space VAEs (InteL-VAEs). InteL-VAEs use an intermediary set of latent variables to control the stochasticity of the encoding process, before mapping these in turn to the latent representation using a parametric function that encapsulates our desired inductive bias(es). This allows us to impose properties like sparsity or clustering on learned representations, and incorporate human knowledge into the generative model. Whereas changing the prior only indirectly encourages behavior through regularizing the encoder, InteL-VAEs are able to directly enforce desired characteristics. Moreover, they bypass the computation and encoder design issues caused by non-Gaussian priors, while allowing for additional flexibility through training of the parametric mapping function. We show that these advantages, in turn, lead to both better generative models and better representations being learned. © 2022 ICLR 2022 - 10th International Conference on Learning Representationss. All rights reserved.
Original languageEnglish
Title of host publicationThe Tenth International Conference on Learning Representations
Subtitle of host publicationICLR 2022
PublisherInternational Conference on Learning Representations, ICLR
Publication statusPublished - 2022
Externally publishedYes
Event10th International Conference on Learning Representations (ICLR 2022) - Virtual
Duration: 25 Apr 202229 Apr 2022
https://iclr.cc/Conferences/2022
https://openreview.net/group?id=ICLR.cc/2022

Publication series

NameICLR - International Conference on Learning Representations

Conference

Conference10th International Conference on Learning Representations (ICLR 2022)
Period25/04/2229/04/22
Internet address

Fingerprint

Dive into the research topics of 'ON INCORPORATING INDUCTIVE BIASES INTO VAES'. Together they form a unique fingerprint.

Cite this