2nd Workshop on Deep Generative Model in Machine Learning:
Theory, Principle and Efficacy

ICLR 2026, April 27, Rio de Janeiro, Brazil

Description

Deep Generative Models (DGMs) have significantly advanced and empowered artificial intelligence (AI) over the past decade. From variational autoencoders to generative adversarial networks, flow-based models, and most recently diffusion models, DGMs continue to revolutionize the field of AI.

Despite the tremendous progress in the development of DGMs, significant challenges remain in both theoretical foundations and practical applications. The current theoretical frameworks fall short of providing sustainable and practical principles for DGMs due to:

  1. A considerable gap between theoretical assumptions and real-world practices, making it difficult to reliably describe and predict the behavior of DGMs in practical settings.
  2. The rapid evolution of AI paradigms—such as generative adversarial networks and diffusion models—which frequently shift the focus of theoretical research, leaving gaps in the understanding of newer architectures.

On the algorithmic front, DGMs face critical issues related to computational efficiency and scalability. As models grow in complexity and size, they require increasingly large datasets and computational resources, which makes training and deployment a significant challenge.

  1. Training instability and convergence issues, especially in adversarial settings, can result in unreliable model outputs.
  2. Scaling DGMs for high-resolution or multi-modal data is computationally intensive and often leads to a trade-off between model accuracy and training time.
  3. Practical applications, such as generative tasks in high-dimensional spaces or large-scale deployments in industry, demand more robust solutions for ensuring efficiency without compromising the quality of generated data.

This workshop will center around these challenges, aiming to bring together experts from learning theory and applications.

Call for Papers

We are excited to invite submissions to the ICLR 2026 Workshop on Deep Generative Models: Theory, Principle, and Efficacy. This workshop aims to explore challenges and opportunities in advancing the theoretical foundations and practical applications of deep generative models (DGMs).

Important Dates

Submission Guidelines

Building on the success of the inaugural DeLTa 2025 workshop, DeLTa 2026 expands its scope to address new theoretical and algorithmic frontiers emerging from the rapid evolution of modern deep generative models. Discussions will be organized along two major axes — Theoretical Foundations and Algorithms & Applications.

Theoretical Foundations include, but are not limited to:

Algorithms & Applications include, but are not limited to:

Submission Format

Accepted Submissions

Accepted papers will be presented as talks or posters during the workshop. The workshop will select the best paper to recognize outstanding contributions in the field.

Non-Proceedings Policy

This workshop does not produce formal proceedings. Accepted submissions will appear on OpenReview, but authors remain free to submit and publish their work elsewhere in the future.

Submission Portal

Submit your paper through the OpenReview platform.

For inquiries, contact us at delta.workshop.ml@gmail.com.

Important Information About Tiny Papers

This year, ICLR is discontinuing the separate “Tiny Papers” track and instead requires each workshop to accept short (3–5 pages in ICLR format, exact page length to be determined by each workshop) paper submissions, with an eye towards inclusion. Authors of these papers will be earmarked for potential funding from ICLR. A separate application for Financial Assistance is required to evaluate eligibility. The application for Financial Assistance will open at the beginning of February and close on March 2, 2025. For more details, visit https://iclr.cc/Conferences/2025/CallForTinyPapers.

Schedule

April 27, 2026 — Rio de Janeiro, Brazil

Brasília Time (GMT−3) Event
9:00 — 9:10 Opening Remarks
Kenji Fukumizu
9:10 — 9:40
René Vidal
How Geometry Shapes Optimization in Deep Generative Models
Show Abstract

Deep generative models have achieved remarkable empirical success, but their theoretical foundations remain poorly understood. This talk presents recent progress on the geometry and optimization of modern generative models, focusing on three representative settings. First, generative model inversion is analyzed with linear convergence of gradient descent established under two geometric conditions on the loss landscape, avoiding unrealistic random-weight assumptions. Second, transformer-based diffusion models trained on multi-token Gaussian mixture data are studied, showing that gradient descent converges to the Bayes-optimal denoiser and that self-attention approximates the optimal MMSE estimator. Third, Parsimonious Flow Matching (PFM) is introduced, which replaces the standard isotropic Gaussian latent with a multimodal mixture aligned with data structure, yielding better-conditioned optimization and faster convergence. Together, these results highlight how geometric structure in data and latent spaces enables sharper theoretical guarantees and more efficient generative modeling.

9:40 — 10:10
Sitan Chen
Theory for Discrete Diffusions: Parallel Decoding and Variable-Length Generation
Show Abstract

Compared to autoregressive models and even to continuous diffusions, diffusion language models offer a fundamentally different design space for crafting efficient and flexible generation processes. This talk discusses work along two axes of this design space: parallel decoding and variable-length generation. In the first half, an exact characterization of the optimal inference schedule for masked diffusion models is given, which depends on a certain "information profile" specific to the data distribution. From this characterization, simple schedules are derived that enable sampling provably more efficiently than autoregressive models for any distribution with bounded correlations. In the second half, FlexMDM is presented, a theoretically principled and empirically lightweight method for equipping diffusion language models with the ability to generate sequences of arbitrary length, while provably preserving their any-order generation capabilities.

10:10 — 11:00 Poster / Break
11:00 — 11:30
Nisha Chandramoorthy
Toward Physical Generative Models
Show Abstract

In any generative model, the generated samples have a different distribution than the data distribution, due to inevitable learning errors. Moreover, this discrepancy, and metrics for evaluating the generated samples, are hard to characterize in high dimensions, motivating the need to understand how learning errors affect the reproducibility of certain "features" of the generated distributions. A first question is whether generative models produce "physical" samples, i.e., samples whose support is close to that of the true target distribution, despite algorithmic errors. A second question concerns what we term a lazy generative model: given samples from the target, we apply an arbitrary random dynamical system such that the distribution at finite time is approximately Gaussian. In principle, this noising process cannot be exactly inverted to recover target samples—but under what conditions can we approximately recover samples from a nearby distribution? The first part is joint work with Adriaan de Clercq (UChicago) and the second with Georg Gottwald (U Sydney).

11:30 — 11:40
Oral: Learning Unmasking Policies for Diffusion Language Models
Metod Jazbec, Theo X. Olausson, Louis Béthune, Pierre Ablin, Michael Kirchhof, Joao Monteiro, Victor Guilherme Turrisi da Costa, Jason Ramapuram, Marco Cuturi
11:40 — 11:50
Oral: Manifold Generalization Provably Proceeds Memorization in Diffusion Models
Zebang Shen, Ya-Ping Hsieh, Niao He
11:50 — 12:00
Oral: Spectral Condition for µP under Width–Depth Scaling
Chenyu Zheng, Rongzhen Wang, Xinyu Zhang, Chongxuan Li
12:00 — 13:30 Lunch
13:30 — 14:00
Atsushi Nitanda
Slowly Annealed Langevin Dynamics: Theory and Applications to Training-Free Guided Generation
Show Abstract

We study Slowly Annealed Langevin Dynamics (SALD), a sampler for tracking a path of moving target distributions and approximating the terminal target through time slowdown. We establish non-asymptotic convergence guarantees via a KL differential inequality, showing that slowdown improves tracking through contraction of intermediate targets and a path-complexity term captured by an entropic action. Motivated by training-free guided generation with pretrained score-based generative models, we further introduce Velocity-Aware SALD (VA-SALD), which explicitly incorporates the underlying marginal distributions of the pretrained model and uses slowdown to correct the additional deviation induced by guidance. This yields a principled framework for training-free guided generation together with convergence guarantees that clarify the roles of intermediate functional inequalities, path complexity, and guidance bias.

14:00 — 14:30
Zahra Kadkhodaie
Blind Denoising Diffusion Models (BDDM) and Adaptive Sampling Algorithm
Show Abstract

Denoising diffusion models (DDMs) are state-of-the-art for numerous tasks pertaining to image generation and inverse problems. Yet many aspects of the training and sampling pipeline remain poorly understood. For instance, the necessity of noise conditioning has been particularly elusive, forcing practitioners to incorporate unnatural noise embeddings into neural network architectures and use ad hoc noise schedules during sampling. Motivated to remove these inconsistencies, a complete theory for blind denoising diffusion models (BDDMs) is provided: a variant of DDMs where the noise amplitude is not passed into the neural network during training nor sampling. The correctness of BDDMs as a sampling algorithm is justified under the sole assumption of low intrinsic dimensionality of the underlying data distribution relative to the ambient dimension. This assumption arises through the introduction of the Bayesian problem of estimating noise levels through a single noisy sample, which might be of independent interest. The performance of BDDMs is compared to standard DDMs in a variety of settings, showcasing the benefits of an adaptive scheme rigorously justified by theoretical analysis.

14:30 — 15:20 Poster / Break
15:20 — 15:50
Arnaud Doucet
Unsupervised Source Separation via Generative Modeling
Show Abstract

The goal of single-channel source separation is to reconstruct \(K\) sources given their mixture. In supervised settings where vast amounts of clean source data are available, this challenging, ill-posed problem has been addressed successfully by generative diffusion and flow-based prior models. However, access to such clean source samples is often limited. To bridge this gap, we present an unsupervised flow matching approach for source separation that learns directly from observed mixtures. This method relies on a novel combination of state-of-the-art supervised flow matching and regression-based self-supervised techniques. We provide insights into the objectives optimized by this approach and demonstrate it on image and audio benchmarks.

15:50 — 16:20
Jannis Chemseddine
Generative Flows from 1D Processes and Adapting Noise to Data
Show Abstract

The aim of the talk is twofold. First, interesting noising processes besides Brownian motion are examined: the physics-inspired Kac process and the MMD gradient flow, leading to compactly supported measure curves and a better regularity of the flow matching velocity field. Second, a "quantile toolbox" for building generative models is presented: a unifying theory and a practical toolkit that turns latent noise selection into a data-driven design element.

16:20 — 16:30
Oral: WiSP-OSch: Solver Within-Step Parallelism and Order Scheduling for Diffusion Sampling
Víctor Lucas Rosada Canesin, Julia Gusak
16:30 — 16:40
Oral: Latent Process Generator Matching
Lukas Billera, Hedwig Nora Nordlinder, Ben Murrell
16:40 — 16:50
Oral: Query Lower Bounds for Diffusion Sampling
Zhiyang Xun, Eric Price
16:50 — 17:00 Awards & Closing Remarks
Mingyuan Bai

Confirmed Speakers [A-Z]

Nisha Chandramoorthy

Nisha Chandramoorthy

Assistant Professor, The University of Chicago

Jannis Chemseddine

Jannis Chemseddine

TU Berlin

Sitan Chen

Sitan Chen

Assistant Professor, Harvard University

Arnaud Doucet

Arnaud Doucet

Senior Staff Research Scientist, Google DeepMind

Zahra

Zahra Kadkhodaie

Research Fellow, Flatiron

Atsushi

Atsushi Nitanda

Principal Scientist, A*STAR

Rene Vidal

René Vidal

Professor, University of Pennsylvania

Organizers

Taiji Suzuki

Taiji Suzuki

Professor, University of Tokyo

Sara Fridovich-Keil

Sara Fridovich-Keil

Assistant Professor, Georgia Tech

Valentin De Bortoli

Valentin De Bortoli

Research Scientist, Google DeepMind

Wei Huang

Wei Huang

Research Scientist, RIKEN AIP

Mingyuan Bai

Mingyuan Bai

Postdoctoral Researcher, RIKEN AIP

Qing Qu

Qing Qu

Assistant Professor, University of Michigan

Andi Han

Andi Han

Lecturer, University of Sydney

Kenji Fukumizu

Kenji Fukumizu

Director, The Institute of Statistical Mathematics