publications
2026
- arXivDiffusion-based Annealed Boltzmann Generators : benefits, pitfalls and hopesLouis Grenioux* and Maxence Noble*2026
Sampling configurations at thermodynamic equilibrium is a central challenge in statistical physics. Boltzmann Generators (BGs) tackle it by combining a generative model with a Monte Carlo (MC) correction step to obtain asymptotically unbiased samples from an unnormalized target. Most current BGs use classic MC mechanisms such as importance sampling, which both require tractable likelihoods from the backbone model and scale poorly in high-dimensional, multi-modal targets. We study BGs built on annealed Monte Carlo (aMC), which is designed to overcome these limitations by bridging a simple reference to the target through a sequence of intermediate densities. Diffusion models (DMs) are powerful generative models and have already been incorporated into aMC-based recalibration schemes via the diffusion-induced density path, making them appealing backbones for aMC-BGs. We provide an empirical meta-analysis of DM-based aMC-BGs on controlled multi-modal Gaussian mixtures (varying mode separation, number of modes, and dimension), explicitly disentangling inference effects from learning effects by comparing (i) a perfectly learned DM and (ii) a DM trained from data. Even with a perfect DM, standard integrations using only first-order stochastic denoising kernels fail systematically, whereas second-order denoising kernels can substantially improve performance when covariance information is available. We further propose a deterministic aMC integration based on first-order transport maps derived from DMs, which outperforms the stochastic first-order variant at higher computational cost. Finally, in the learned-DM setting, all DM-aMC variants struggle to produce accurate BGs; we trace the main bottleneck to inaccurate DM log-density estimation.
@misc{grenioux2026diffusionbasedannealedboltzmanngenerators, title = {Diffusion-based Annealed Boltzmann Generators : benefits, pitfalls and hopes}, author = {Grenioux, Louis and Noble, Maxence}, year = {2026}, url = {https://arxiv.org/abs/2601.21026}, eprint = {2601.21026}, archiveprefix = {arXiv}, primaryclass = {stat.ML}, } - arXivA Diffusive Classification Loss for Learning Energy-based Generative ModelsLouis Grenioux*, RuiKang OuYang*, and José Miguel Hernández-Lobato2026
Score-based generative models have recently achieved remarkable success. While they are usually parameterized by the score, an alternative way is to use a series of time-dependent energy-based models (EBMs), where the score is obtained from the negative input-gradient of the energy. Crucially, EBMs can be leveraged not only for generation, but also for tasks such as compositional sampling or building Boltzmann Generators via Monte Carlo methods. However, training EBMs remains challenging. Direct maximum likelihood is computationally prohibitive due to the need for nested sampling, while score matching, though efficient, suffers from mode blindness. To address these issues, we introduce the Diffusive Classification (DiffCLF) objective, a simple method that avoids blindness while remaining computationally efficient. DiffCLF reframes EBM learning as a supervised classification problem across noise levels, and can be seamlessly combined with standard score-based objectives. We validate the effectiveness of DiffCLF by comparing the estimated energies against ground truth in analytical Gaussian mixture cases, and by applying the trained models to tasks such as model composition and Boltzmann Generator sampling. Our results show that DiffCLF enables EBMs with higher fidelity and broader applicability than existing approaches.
@misc{grenioux2026diffusiveclassificationlosslearning, title = {A Diffusive Classification Loss for Learning Energy-based Generative Models}, author = {Grenioux, Louis and OuYang, RuiKang and Hernández-Lobato, José Miguel}, year = {2026}, url = {https://arxiv.org/abs/2601.21025}, eprint = {2601.21025}, archiveprefix = {arXiv}, primaryclass = {stat.ML}, }
2025
- arXivRiemannian Stochastic Interpolants for Amorphous Particle Systems2025
Modern generative models hold great promise for accelerating diverse tasks involving the simulation of physical systems, but they must be adapted to the specific constraints of each domain. Significant progress has been made for biomolecules and crystalline materials. Here, we address amorphous materials (glasses), which are disordered particle systems lacking atomic periodicity. Sampling equilibrium configurations of glass-forming materials is a notoriously slow and difficult task. This obstacle could be overcome by developing a generative framework capable of producing equilibrium configurations with well-defined likelihoods. In this work, we address this challenge by leveraging an equivariant Riemannian stochastic interpolation framework which combines Riemannian stochastic interpolant and equivariant flow matching. Our method rigorously incorporates periodic boundary conditions and the symmetries of multi-component particle systems, adapting an equivariant graph neural network to operate directly on the torus. Our numerical experiments on model amorphous systems demonstrate that enforcing geometric and symmetry constraints significantly improves generative performance.
@misc{grenioux2025riemannian, title = {Riemannian Stochastic Interpolants for Amorphous Particle Systems}, author = {Grenioux, Louis and Galliano, Leonardo and Berthier, Ludovic and Biroli, Giulio and Gabri{\'e}, Marylou}, year = {2025}, url = {https://arxiv.org/abs/2512.16607}, eprint = {2512.16607}, archiveprefix = {arXiv}, primaryclass = {stat.ML}, } - PhDInteractions and opportunities at the crossroads of deep probabilistic modeling and statistical inference through Markov Chains Monte CarloLouis GreniouxInstitut Polytechnique de Paris, Oct 2025
This thesis advances the field of sampling, a cornerstone of Bayesian inference, computational physics, and probabilistic modeling, where the goal is to generate samples from a known probability density. A related challenge arises in generative modeling, which seeks to produce new data resembling a given dataset, a problem that has seen major breakthroughs through recent advances in deep learning. The central aim of this work is to leverage modern generative models to enhance classical sampling frameworks. The study begins by examining the inherent difficulties of multi-modal sampling, identifying key limitations of both classical and advanced Monte Carlo methods. It then explores the integration of pre-trained normalizing flows into traditional Monte Carlo schemes, providing practical guidance on their performance across diverse target distributions. Building on this, diffusion models are incorporated into advanced annealed Monte Carlo methods, revealing both their potential and their limitations. The work also investigates how diffusion models can be embedded within a variational inference framework. In parallel, it proposes a learning-free diffusion-based sampler that replaces neural approximators with Monte Carlo estimators. Finally, these enhanced sampling strategies are applied to the training of energy-based models, introducing a novel algorithm in which a normalizing flow serves as an auxiliary sampler to facilitate the training of these expressive yet challenging generative models.
@phdthesis{grenioux2025interactions, title = {Interactions and opportunities at the crossroads of deep probabilistic modeling and statistical inference through Markov Chains Monte Carlo}, author = {Grenioux, Louis}, year = {2025}, month = oct, number = {2025IPPAX091}, url = {https://theses.hal.science/tel-05438658}, school = {Institut Polytechnique de Paris}, hal_id = {tel-05438658}, hal_version = {v1}, } - ICLRImproving the evaluation of samplers on multi-modal targetsLouis Grenioux*, Maxence Noble*, and Marylou GabriéIn Frontiers in Probabilistic Inference: Learning meets Sampling, 2025
Addressing multi-modality constitutes one of the major challenges of sampling. In this reflection paper, we advocate for a more systematic evaluation of samplers towards two sources of difficulty that are mode separation and dimension. For this, we propose a synthetic experimental setting that we illustrate on a selection of samplers, focusing on the challenging criterion of recovery of the mode relative importance. These evaluations are crucial to diagnose the potential of samplers to handle multi-modality and therefore to drive progress in the field.
@inproceedings{grenioux2025improving, title = {Improving the evaluation of samplers on multi-modal targets}, author = {Grenioux, Louis and Noble, Maxence and Gabri{\'e}, Marylou}, year = {2025}, booktitle = {Frontiers in Probabilistic Inference: Learning meets Sampling}, url = {https://openreview.net/forum?id=d91E9RhVFU}, } - ICLRLearned Reference-based Diffusion Sampler for multi-modal distributionsIn The Thirteenth International Conference on Learning Representations, 2025
Over the past few years, several approaches utilizing score-based diffusion have been proposed to sample from probability distributions, that is without having access to exact samples and relying solely on evaluations of unnormalized densities. The resulting samplers approximate the time-reversal of a noising diffusion process, bridging the target distribution to an easy-to-sample base distribution. In practice, the performance of these methods heavily depends on key hyperparameters that require ground truth samples to be accurately tuned. Our work aims to highlight and address this fundamental issue, focusing in particular on multi-modal distributions, which pose significant challenges for existing sampling methods. Building on existing approaches, we introduce Learned Reference-based Diffusion Sampler (LRDS), a methodology specifically designed to leverage prior knowledge on the location of the target modes in order to bypass the obstacle of hyperparameter tuning. LRDS proceeds in two steps by (i) learning a reference diffusion model on samples located in high-density space regions and tailored for multimodality, and (ii) using this reference model to foster the training of a diffusion-based sampler. We experimentally demonstrate that LRDS best exploits prior knowledge on the target distribution compared to competing algorithms on a variety of challenging distributions.
@inproceedings{noble2025learned, title = {Learned Reference-based Diffusion Sampler for multi-modal distributions}, author = {Noble, Maxence and Grenioux, Louis and Gabri{\'e}, Marylou and Durmus, Alain Oliviero}, year = {2025}, booktitle = {The Thirteenth International Conference on Learning Representations}, url = {https://openreview.net/forum?id=fmJUYgmMbL}, }
2024
- ICMLStochastic Localization via Iterative Posterior SamplingIn Proceedings of the 41st International Conference on Machine Learning, 2024
This paper has been selected as a spotlight-designated paper at the conference. Top 3.5% acceptance rate.
Building upon score-based learning, new interest in stochastic localization techniques has recently emerged. In these models, one seeks to noise a sample from the data distribution through a stochastic process, called observation process, and progressively learns a denoiser associated to this dynamics. Apart from specific applications, the use of stochastic localization for the problem of sampling from an unnormalized target density has not been explored extensively. This work contributes to fill this gap. We consider a general stochastic localization framework and introduce an explicit class of observation processes, associated with flexible denoising schedules. We provide a complete methodology, Stochastic Localization via Iterative Posterior Sampling (SLIPS), to obtain approximate samples of these dynamics, and as a by-product, samples from the target distribution. Our scheme is based on a Markov chain Monte Carlo estimation of the denoiser and comes with detailed practical guidelines. We illustrate the benefits and applicability of SLIPS on several benchmarks of multi-modal distributions, including Gaussian mixtures in increasing dimensions, Bayesian logistic regression and a high-dimensional field system from statistical-mechanics.
@inproceedings{grenioux2024stochastic, title = {Stochastic Localization via Iterative Posterior Sampling}, author = {Grenioux, Louis and Noble, Maxence and Gabri\'{e}, Marylou and Oliviero Durmus, Alain}, year = {2024}, booktitle = {Proceedings of the 41st International Conference on Machine Learning}, publisher = {PMLR}, series = {Proceedings of Machine Learning Research}, volume = {235}, pages = {16337--16376}, url = {https://proceedings.mlr.press/v235/grenioux24a.html}, editor = {Salakhutdinov, Ruslan and Kolter, Zico and Heller, Katherine and Weller, Adrian and Oliver, Nuria and Scarlett, Jonathan and Berkenkamp, Felix}, }
2023
- ICMLBalanced Training of Energy-Based Models with Adaptive Flow SamplingLouis Grenioux, Éric Moulines, and Marylou GabriéWorkshop on Structured Probabilistic Inference & Generative Modeling, 2023
Energy-based models (EBMs) are versatile density estimation models that directly parameterize an unnormalized log density. Although very flexible, EBMs lack a specified normalization constant of the model, making the likelihood of the model computationally intractable. Several approximate samplers and variational inference techniques have been proposed to estimate the likelihood gradients for training. These techniques have shown promising results in generating samples, but little attention has been paid to the statistical accuracy of the estimated density, such as determining the relative importance of different classes in a dataset. In this work, we propose a new maximum likelihood training algorithm for EBMs that uses a different type of generative model, normalizing flows (NF), which have recently been proposed to facilitate sampling. Our method fits an NF to an EBM during training so that an NF-assisted sampling scheme provides an accurate gradient for the EBMs at all times, ultimately leading to a fast sampler for generating new data.
@article{grenioux2023balanced, title = {Balanced Training of Energy-Based Models with Adaptive Flow Sampling}, author = {Grenioux, Louis and Moulines, Éric and Gabrié, Marylou}, year = {2023}, journal = {Workshop on Structured Probabilistic Inference & Generative Modeling}, } - ICMLOn Sampling with Approximate Transport MapsLouis Grenioux, Alain Oliviero Durmus, Eric Moulines, and 1 more authorIn Proceedings of the 40th International Conference on Machine Learning, 2023
Transport maps can ease the sampling of distributions with non-trivial geometries by transforming them into distributions that are easier to handle. The potential of this approach has risen with the development of Normalizing Flows (NF) which are maps parameterized with deep neural networks trained to push a reference distribution towards a target. NF-enhanced samplers recently proposed blend (Markov chain) Monte Carlo methods with either (i) proposal draws from the flow or (ii) a flow-based reparametrization. In both cases, the quality of the learned transport conditions performance. The present work clarifies for the first time the relative strengths and weaknesses of these two approaches. Our study concludes that multimodal targets can be reliably handled with flow-based proposals up to moderately high dimensions. In contrast, methods relying on reparametrization struggle with multimodality but are more robust otherwise in high-dimensional settings and under poor training. To further illustrate the influence of target-proposal adequacy, we also derive a new quantitative bound for the mixing time of the Independent Metropolis-Hastings sampler.
@inproceedings{grenioux2023sampling, title = {On Sampling with Approximate Transport Maps}, author = {Grenioux, Louis and Oliviero Durmus, Alain and Moulines, Eric and Gabri\'{e}, Marylou}, year = {2023}, booktitle = {Proceedings of the 40th International Conference on Machine Learning}, publisher = {PMLR}, series = {Proceedings of Machine Learning Research}, volume = {202}, pages = {11698--11733}, url = {https://proceedings.mlr.press/v202/grenioux23a.html}, editor = {Krause, Andreas and Brunskill, Emma and Cho, Kyunghyun and Engelhardt, Barbara and Sabato, Sivan and Scarlett, Jonathan}, }