Diffusion models produce images with high quality and customizability, enabling them to be used for commercial art and graphic designpurposes.
In this work, we study image retrieval frameworks that enable us to compare generated images with training samples and detect when content has been replicated.
Recently, Rissanen et al., (2022) have presented a new type of diffusion
process for generative modeling based on heat dissipation, or blurring, as an
alternative to isotropic Gaussian diffusion. Here
Diffusion models have recently achieved great success in synthesizing diverse
and high-fidelity images. However, sampling speed and memory constraints remain
a major barrier to the practical adoption
Generative image synthesis with diffusion models has recently achieved
excellent visual quality in several tasks such as text-based or
class-conditional image synthesis. Much of this success is due to
Denoising diffusion probabilistic models (DDPMs) have emerged as competitive
generative models yet brought challenges to efficient sampling. In this paper,
we propose novel bilateral denoising diffusi
We introduce Autoregressive Diffusion Models (ARDMs), a model class
encompassing and generalizing order-agnostic autoregressive models (Uria et
al., 2014) and absorbing discrete diffusion (Austin et a
Diffusion models are recent state-of-the-art methods for image generation and
likelihood estimation. In this work, we generalize continuous-time diffusion
models to arbitrary Riemannian manifolds and
We introduce a family of diffusion-based generative models that obtain state-of-the-art likelihoods on image density estimation benchmarks, outperforming autoregressive models that have dominated these benchmarks for many years with often significantly faster optimization.
We show that the variational lower bound simplifies to a remarkably short expression in terms of the signal-to-noise ratio of the diffused data, thereby improving our theoretical understanding of this model class.
Generating temporally coherent high fidelity video is an important milestone
in generative modeling research. We make progress towards this milestone by
proposing a diffusion model for video generatio
Vector quantized diffusion (VQ-Diffusion) is a powerful generative model for
text-to-image synthesis, but sometimes can still generate low-quality samples
or weakly correlated images with text input.
While modern machine learning models rely on increasingly large training
datasets, data is often limited in privacy-sensitive domains. Generative models
trained with differential privacy (DP) on sensi
Denoising diffusion probabilistic models (DDPM) are powerful hierarchical
latent variable models with remarkable sample generation quality and training
stability. These properties can be attributed to
We introduce a framework for automatically defining and learning deepgenerative models with problem-specific structure.
We tackle problem domains that are more traditionally solved by algorithms such as sorting, constraint satisfaction for sudoku, and matrix factorization.
We propose a family of First Hitting Diffusion Models (FHDM), deep generative
models that generate data with a diffusion process that terminates at a random
first hitting time. This yields an extensio
Standard diffusion models involve an image transform -- adding Gaussian noise
-- and an image restoration operator that inverts this degradation. We observe
that the generative behavior of diffusion m
In medical applications, weakly supervised anomaly detection methods are of
great interest, as only image-level annotations are required for training.
Current anomaly detection methods mainly rely on