The variational autoencoder (VAE) framework remains a popular option for
training unsupervised generative models, especially for discrete data where
generative adversarial networks (GANs) require work
The Variational Autoencoder (VAE) is a popular and powerful model applied to
text modelling to generate diverse sentences. However, an issue known as
posterior collapse (or KL loss vanishing) happens
We present a new generative autoencoder model with dual contradistinctive
losses to improve generative autoencoder that performs simultaneous inference
(reconstruction) and synthesis (sampling). Our m
We present a preliminary study on an end-to-end variational autoencoder (vae)for sound morphing.
We introduce the mel-frequency cepstrumcoefficient dynamic time warping (mfcc-dtw) deviation as a measure of how well the vae decoder projects the class center in the latent (bottleneck) layer to the center of the sounds of that class in the audio domain.
Graph autoencoders (AE) and variational autoencoders (VAE) are powerful node
embedding methods, but suffer from scalability issues. In this paper, we
introduce FastGAE, a general framework to scale gr
Variational autoencoder-based voice conversion (VAE-VC) has the advantage of
requiring only pairs of speeches and speaker labels for training. Unlike the
majority of the research in VAE-VC which focus
One noted issue of vector-quantized variational autoencoder (VQ-VAE) is that
the learned discrete representation uses only a fraction of the full capacity
of the codebook, also known as codebook colla
We present an SVQ-VAE architecture using a split vector quantizer for NTTS,
as an enhancement to the well-known VAE and VQ-VAE architectures. Compared to
these previous architectures, our proposed mod
As one of the most popular generative models, Variational Autoencoder (VAE)
approximates the posterior of latent variables based on amortized variational
inference. However, when the decoder network i
The Variational Auto-Encoder (VAE) is a simple, efficient, and popular deep
maximum likelihood model. Though usage of VAEs is widespread, the derivation of
the VAE is not as widely understood. In this
The Variational Auto-Encoder (VAE) is one of the most used unsupervised
machine learning models. But although the default choice of a Gaussian
distribution for both the prior and posterior represents
State-of-the-art Variational Auto-Encoders (VAEs) for learning disentangled
latent representations give impressive results in discovering features like
pitch, pause duration, and accent in speech data
We propose a gaussian manifold variational auto-encoder (gm-vae) whose latent space consists of a set of diagonal gaussian distributions with the fisher information metric.
To learn the vae endowed with the gaussian manifold, we first propose a pseudo-gaussian manifold normal distribution based on the kullback-leibler divergence, a local approximation of the squared fisher-rao distance, to define a densityover the latent space.
Variational autoencoders (VAE) have quickly become a central tool in machine
learning, applicable to a broad range of data types and latent variable models.
By far the most common first step, taken by
Geometric disentanglement, the separation of latent codes for intrinsic (i.e.
identity) and extrinsic(i.e. pose) geometry, is a prominent task for generative
models of non-Euclidean data such as 3D de
Variational Auto-Encoder (VAE) has become the de-facto learning paradigm in
achieving both representation learning and generation for natural language.
However, existing VAE-based language models eith
Does a Variational AutoEncoder (VAE) consistently encode typical samples
generated from its decoder? This paper shows that the perhaps surprising answer
to this question is `No'; a (nominally trained)
Many different methods to train deep generative models have been introduced
in the past. In this paper, we propose to extend the variational auto-encoder
(VAE) framework with a new type of prior which
In many imaging modalities, objects of interest can occur in a variety of
locations and poses (i.e. are subject to translations and rotations in 2d or
3d), but the location and pose of an object does
Unsupervised feature learning for point clouds has been vital for large-scale point cloud understanding. Recent deep learning based methods depend on learning global geometry from self-reconstruction.