Visualizing population structure with variational autoencoders
نویسندگان
چکیده
Abstract Dimensionality reduction is a common tool for visualization and inference of population structure from genotypes, but popular methods either return too many dimensions easy plotting (PCA) or fail to preserve global geometry (t-SNE UMAP). Here we explore the utility variational autoencoders (VAEs)—generative machine learning models in which pair neural networks seek first compress then recreate input data—for visualizing genetic variation. VAEs incorporate nonlinear relationships, allow users define dimensionality latent space, our tests better than t-SNE UMAP. Our implementation, call popvae, available as command-line python program at github.com/kr-colab/popvae. The approach yields embeddings that capture subtle aspects humans Anopheles mosquitoes, can generate artificial genotypes characteristic given sample population.
منابع مشابه
Discrete Variational Autoencoders
Probabilistic models with discrete latent variables naturally capture datasets composed of discrete classes. However, they are difficult to train efficiently, since backpropagation through discrete variables is generally not possible. We introduce a novel class of probabilistic models, comprising an undirected discrete component and a directed hierarchical continuous component, that can be trai...
متن کاملVariational Composite Autoencoders
Learning in the latent variable model is challenging in the presence of the complex data structure or the intractable latent variable. Previous variational autoencoders can be low effective due to the straightforward encoder-decoder structure. In this paper, we propose a variational composite autoencoder to sidestep this issue by amortizing on top of the hierarchical latent variable model. The ...
متن کاملSemi-Amortized Variational Autoencoders
Amortized variational inference (AVI) replaces instance-specific local inference with a global inference network. While AVI has enabled efficient training of deep generative models such as variational autoencoders (VAE), recent empirical work suggests that inference networks can produce suboptimal variational parameters. We propose a hybrid approach, to use AVI to initialize the variational par...
متن کاملTutorial on Variational Autoencoders
In just three years, Variational Autoencoders (VAEs) have emerged as one of the most popular approaches to unsupervised learning of complicated distributions. VAEs are appealing because they are built on top of standard function approximators (neural networks), and can be trained with stochastic gradient descent. VAEs have already shown promise in generating many kinds of complicated data, incl...
متن کاملLadder Variational Autoencoders
Variational autoencoders are powerful models for unsupervised learning. However deep models with several layers of dependent stochastic variables are difficult to train which limits the improvements obtained using these highly expressive models. We propose a new inference model, the Ladder Variational Autoencoder, that recursively corrects the generative distribution by a data dependent approxi...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: G3: Genes, Genomes, Genetics
سال: 2021
ISSN: ['2160-1836']
DOI: https://doi.org/10.1093/g3journal/jkaa036