A Hands-On Tutorial for Training Interpretable Variational Autoencoders Using siVAE
This course is available only as a part of subscription plans.
Variational autoencoders (VAEs) are one of the most widely used deep generative models with applications to computer vision, language processing, and genomics, among other fields. VAEs are typically used to perform non-linear dimensionality reduction, by mapping high dimensional samples such as images into a low-dimensional latent space for visualization and other downstream analysis. One of the key limitations of VAEs is their lack of interpretability: until now, it has been challenging to identify the relationship, or attributions, between individual latent dimensions and the original input features of the samples. Increasing the interpretability of the latent dimensions learned by the VAE will improve our understanding of what the latent space of VAEs is capturing, and help interpret their visualizations.
In this hands-on tutorial, we will introduce attendees to the siVAE (scalable, interpretable VAE) model that infers a set of factor loadings that explicitly map latent dimensions to the input features that define them, during training of the VAE model. Using standard datasets from computer vision (MNIST, Fashion-MNIST, and CIFAR-10), we will walk attendees through the process of training the siVAE model, visualizing the sample embeddings inferred by classic VAEs, and extracting and visualizing the features that contribute to individual latent dimensions. We will also teach attendees how to estimate and visualize feature awareness, a new metric for measuring the overall importance of individual features for embedding a sample in the latent space. At the end of the tutorial, attendees will be able to train an siVAE model on their own datasets and interpret and visualize the latent dimensions inferred.
Tutorial Overview and Instructor Bios
Getting Started
Resource Downloads
Tutorial Slides Presentation
Introduction to Variational Autoencoders/Generative Models and siVAE
Model setup in TensorFlow and VAE Implementations
Training Scalable and interpretable VAE (siVAE) with pre-training steps
Training Steps for siVAE and Combining Model
Gerald Quon, PhD and Yongin Choi
Gerald Quon, PhD
Yongin Choi