Excited to share our latest preprint: ๐๐ฒ๐ฎ๐ฟ๐ป๐ถ๐ป๐ด ๐๐ฎ๐บ๐ถ๐น๐๐ผ๐ป๐ถ๐ฎ๐ป ๐๐น๐ผ๐ ๐ ๐ฎ๐ฝ๐: ๐ ๐ฒ๐ฎ๐ป ๐๐น๐ผ๐ ๐๐ผ๐ป๐๐ถ๐๐๐ฒ๐ป๐ฐ๐ ๐ณ๐ผ๐ฟ ๐๐ฎ๐ฟ๐ด๐ฒ-๐ง๐ถ๐บ๐ฒ๐๐๐ฒ๐ฝ ๐ ๐ผ๐น๐ฒ๐ฐ๐๐น๐ฎ๐ฟ ๐๐๐ป๐ฎ๐บ๐ถ๐ฐ๐ ๐
19.02.2026 15:18 โ ๐ 7 ๐ 3 ๐ฌ 0 ๐ 0
Consistent Sampling and Simulation: Molecular Dynamics with Energy-Based Diffusion Models
In recent years, diffusion models trained on equilibrium molecular distributions have proven effective for sampling biomolecules. Beyond direct sampling, the score of such a model can also be used to ...
(7/n) Check out our paper and code:
Paper: arxiv.org/abs/2506.17139
Code + models: github.com/noegroup/Sco...
And also, our self-contained notebooks!
Colab (JAX):
colab.research.google.com/drive/1r3DGO...
Colab (PyTorch):
colab.research.google.com/drive/1rbcND...
#NeurIPS2025 #Diffusion #MD
06.11.2025 14:41 โ ๐ 1 ๐ 0 ๐ฌ 0 ๐ 0
(6/n)
Done with a brilliant team: Hao Wu, Leon Klein, Stephan Gรผnnemann, and @franknoe.bsky.social .
06.11.2025 14:41 โ ๐ 2 ๐ 0 ๐ฌ 1 ๐ 0
Comparison of equilibrium distributions obtained by iid sampling and Langevin simulation (sim) across different systems and methods. While classical iid sampling recovers the reference equilibrium distribution,
performing simulation with the learned score reveals inconsistencies when models are not trained with Fokker-Planck regularization, i.e., p(x) != p_0(x). Regularized models achieve consistent behavior across systems.
(5/n) With this, we can run coarse-grained Langevin dynamics directly, without the need for any priors or force labels.
This works across biomolecular systems including fast-folding proteins like Chignolin and BBA.
Here is a comparison with and without our regularization:
06.11.2025 14:41 โ ๐ 2 ๐ 0 ๐ฌ 1 ๐ 0
A model trained with FokkerโPlanck regularization is self-consistent, and aligns the learned score at t = 0 with the distribution recovered by diffusion sampling.
(4/n) Our solution:
We train an energy-based diffusion model and regularize it to satisfy the FokkerโPlanck equation.
This enforces consistency between:
- The density recovered via denoising
- The potential energy learned at t = 0
Result: the same model can be used for sampling AND simulation.
06.11.2025 14:41 โ ๐ 3 ๐ 0 ๐ฌ 1 ๐ 0
(3/n) The root issue is that at very small diffusion times, diffusion models are inaccurate.
The loss is large, and the models violate the Fokker-Planck equation, meaning the evolution of the modelโs density and its score disagree.
When that happens, the recovered energy ๐ผ(x) is not meaningful.
06.11.2025 14:41 โ ๐ 1 ๐ 0 ๐ฌ 1 ๐ 0
Training diffusion models on a 2D toy example reveals inconsistencies. While classical iid diffusion sampling (i.e., denoising) correctly reproduces both modes, evaluating the score at t = 0 to estimate the unnormalized density yields a third mode and an incorrect mass distribution. Such a diffusion model would produce incorrect dynamics while producing correct samples.
(2/n) The problem: classical diffusion models learn scores that reproduce equilibrium samples, but the corresponding energy-based parameterization is not consistent.
So if you try to use the learned energy to derive forces, the dynamics are wrong, even if the samples themselves look fine.
06.11.2025 14:41 โ ๐ 1 ๐ 0 ๐ฌ 1 ๐ 0
(1/n) Can diffusion models simulate molecular dynamics instead of just generating independent samples?
In our NeurIPS 2025 paper, we train energy-based diffusion models that can do both:
- Generate independent samples
- Learn the underlying potential ๐ผ
๐งต๐
Paper: arxiv.org/abs/2506.17139
06.11.2025 14:41 โ ๐ 26 ๐ 4 ๐ฌ 1 ๐ 1
CS student @ TU Berlin. ML for molecular simulation.
ELIZA is a Konrad Zuse Schools of Excellence in Artificial Intelligence funded by @daadworldwide.bsky.social. We train #AItalents building on @ellis.eu. #ZuseSchoolsAI
Follow us on: eliza.school
Machine Learning for Precision Medicine
PhD student at BIFOLD & TU Berlin
Data Scientist at Aignostics
https://marvinsextro.de
PhD Candidate in Computational Biology @ University of Pittsburgh. Working on deep generative models for molecular structure. iandunn.io
PhD Student at the TU Berlin ML group + BIFOLD | BUA Fellow
Model robustness/correction ๐ค๐ง
Understanding representation spaces ๐โจ
PhD student explainable AI @ ML Group TU Berlin, BIFOLD
In-depth, independent reporting to better understand the world, now on Bluesky. News tips? Share them here: http://nyti.ms/2FVHq9v
Working towards the safe development of AI for the benefit of all at Universitรฉ de Montrรฉal, LawZero and Mila.
A.M. Turing Award Recipient and most-cited AI researcher.
https://lawzero.org/en
https://yoshuabengio.org/profile/
๐ณ๏ธโ๐ NVIDIA & Duke. Was Allianz, VantAI, TUM. BioCS+ML dude.
Lab page: https://machine.learning.bio
GScholar: https://scholar.google.com/citations?user=4q0fNGAAAAAJ
Assistant Professor at @EPFL in Computer Science and Life Sciences. PostDoc at @Genentech and @Stanford. PhD at @ETH and @BroadInstitute.
www.aimm.epfl.ch
UofT CompSci PhD Student in Alรกn Aspuru-Guzik's #matterlab and Vector Institute | prev. Apple
ML + Cells + Proteins. PI @ AITHYRA https://alextong.net
International Conference on Learning Representations https://iclr.cc/
Computational physicist at https://peptone.io
PhD @GroupParrinello, PostDoc @franknoe.bsky.social
Disordered Proteins, AI for Science, Molecular Dynamics, Enhanced Sampling
๐ https://scholar.google.com/citations?user=fnJktPAAAAAJ
web @ https://argmin.xyz
interests: machine learning, ai4science, algorithms, coding
member of technical staff @ https://cusp.ai
past @ MSR, DeepMind, MPI-IS
home @ Heimbach (Gilserberg), Berlin, Europe
born @ 353 ppm
block toxicity
he/him
DeepMind Professor of AI @Oxford
Scientific Director @Aithyra
Chief Scientist @VantAI
ML Lead @ProjectCETI
geometric deep learning, graph neural networks, generative models, molecular design, proteins, bio AI, ๐ ๐ถ
Postdoctoral Researcher @ TU Berlin @BIFOLD Berlin | AI for molecular simulations
Strengthening Europe's Leadership in AI through Research Excellence | ellis.eu