Host: The Japanese Society for Artificial Intelligence
Name : 34th Annual Conference, 2020
Number : 34
Location : Online
Date : June 09, 2020 - June 12, 2020
Variational Autoencoder (VAE) training suffers from posterior collapse, which means the decoders of VAEs ignore latent variables. In this paper, we argue that {\em I-unidentifiable} data generating process, which is assumed by several existing VAEs, induces posterior collapse. This is because in such an {\em I-unidentifiable} data generating process, the information that a particular latent variable is designed to acquire is easily acquired by other latent variables without sacrificing log-likelihood. We show that this perspective gives a unified explanation for posterior collapse, using VAE with autoregressive decoder and disentangled sequential autoencoder as examples. In addition, we propose maximizing conditional mutual information with adversarial training to alleviate the unidentifiability issue, which does not require specific constraints on model architectures or latent variable structures. Empirically our method mitigated posterior collapse in the above two models and improved the rate-distortion curve.