Visualizing population structure with variational autoencoders

نویسندگان

چکیده

Abstract Dimensionality reduction is a common tool for visualization and inference of population structure from genotypes, but popular methods either return too many dimensions easy plotting (PCA) or fail to preserve global geometry (t-SNE UMAP). Here we explore the utility variational autoencoders (VAEs)—generative machine learning models in which pair neural networks seek first compress then recreate input data—for visualizing genetic variation. VAEs incorporate nonlinear relationships, allow users define dimensionality latent space, our tests better than t-SNE UMAP. Our implementation, call popvae, available as command-line python program at github.com/kr-colab/popvae. The approach yields embeddings that capture subtle aspects humans Anopheles mosquitoes, can generate artificial genotypes characteristic given sample population.

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Discrete Variational Autoencoders

Probabilistic models with discrete latent variables naturally capture datasets composed of discrete classes. However, they are difficult to train efficiently, since backpropagation through discrete variables is generally not possible. We introduce a novel class of probabilistic models, comprising an undirected discrete component and a directed hierarchical continuous component, that can be trai...

متن کامل

Variational Composite Autoencoders

Learning in the latent variable model is challenging in the presence of the complex data structure or the intractable latent variable. Previous variational autoencoders can be low effective due to the straightforward encoder-decoder structure. In this paper, we propose a variational composite autoencoder to sidestep this issue by amortizing on top of the hierarchical latent variable model. The ...

متن کامل

Semi-Amortized Variational Autoencoders

Amortized variational inference (AVI) replaces instance-specific local inference with a global inference network. While AVI has enabled efficient training of deep generative models such as variational autoencoders (VAE), recent empirical work suggests that inference networks can produce suboptimal variational parameters. We propose a hybrid approach, to use AVI to initialize the variational par...

متن کامل

Tutorial on Variational Autoencoders

In just three years, Variational Autoencoders (VAEs) have emerged as one of the most popular approaches to unsupervised learning of complicated distributions. VAEs are appealing because they are built on top of standard function approximators (neural networks), and can be trained with stochastic gradient descent. VAEs have already shown promise in generating many kinds of complicated data, incl...

متن کامل

Ladder Variational Autoencoders

Variational autoencoders are powerful models for unsupervised learning. However deep models with several layers of dependent stochastic variables are difficult to train which limits the improvements obtained using these highly expressive models. We propose a new inference model, the Ladder Variational Autoencoder, that recursively corrects the generative distribution by a data dependent approxi...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: G3: Genes, Genomes, Genetics

سال: 2021

ISSN: ['2160-1836']

DOI: https://doi.org/10.1093/g3journal/jkaa036