's Avatar

@majhas.bsky.social

PhD Student at Mila & University of Montreal | Generative modeling, sampling, molecules majhas.github.io

123 Followers  |  109 Following  |  11 Posts  |  Joined: 19.11.2024  |  1.9735

Latest posts by majhas.bsky.social on Bluesky


Video thumbnail

The gifs didn't post properly πŸ˜…

Here is one showing the electron cloud in two stages: (1) the learning of electron density during training and (2) the predicted ground-state across conformations 😎

10.06.2025 22:06 β€” πŸ‘ 2    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Post image

(9/9)⚑ Runtime efficiency
Self-refining training reduces total runtime up to 4 times compared to the baseline
and up to 2 times compared to the fully-supervised approach!!!
Less need for large pre-generated datasets β€” training and sampling happen in parallel.

10.06.2025 19:49 β€” πŸ‘ 2    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Post image

(8/n) πŸ§ͺ Robust generalization
We simulate molecular dynamics using each model’s energy predictions and evaluate accuracy along the trajectory.
Models trained with self-refinement stay accurate even far from the training distribution β€” while baselines quickly degrade.

10.06.2025 19:49 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Post image

(7/n) πŸ“Š Performance under data scarcity
Our method achieves low energy error with as few as 25 conformations.
With 10Γ— less data, it matches or outperforms fully supervised baselines.
This is especially important in settings where labeled data is expensive or unavailable.

10.06.2025 19:49 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Post image

(6/n) This minimization leads to Self-Refining Training:
πŸ” Use the current model to sample conformations via MCMC
πŸ“‰ Use those conformations to minimize energy and update the model

Everything runs asynchronously, without need for labeled data and minimal number of conformations from a dataset!

10.06.2025 19:49 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Post image

(5/n) To get around this, we introduce a variational upper bound on the KL between any sampling distribution q(R) and the target Boltzmann distribution.

Jointly minimizing this bound wrt ΞΈ and q yields
βœ… A model that predicts the ground-state solutions
βœ… Samples that match the ground true density

10.06.2025 19:49 β€” πŸ‘ 2    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Post image

(4/n) With an amortized DFT model f_ΞΈ(R), we define the density of molecular conformations as the
Boltzmann distribution

This isn't a typical ML setup because
❌ No samples from the density - can’t train a generative model
❌ No density - can’t sample via Monte Carlo!

10.06.2025 19:49 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Post image

(3/n) DFT offers a scalable solution to the SchrΓΆdinger equation but must be solved independently for each geometry by minimizing energy wrt coefficients C for a fixed basis.

This presents a bottleneck for MD/sampling.

We want to amortize this - train a model that generalizes across geometries R.

10.06.2025 19:49 β€” πŸ‘ 2    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
GitHub - majhas/self-refining-dft Contribute to majhas/self-refining-dft development by creating an account on GitHub.

(2/n) This work is the result of an amazing collaboration with @fntwin.bsky.social Hatem Helal @dom-beaini.bsky.social @k-neklyudov.bsky.social

10.06.2025 19:49 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Post image

(1/n)🚨Train a model solving DFT for any geometry with almost no training data
Introducing Self-Refining Training for Amortized DFT: a variational method that predicts ground-state solutions across geometries and generates its own training data!
πŸ“œ arxiv.org/abs/2506.01225
πŸ’» github.com/majhas/self-...

10.06.2025 19:49 β€” πŸ‘ 12    πŸ” 4    πŸ’¬ 1    πŸ“Œ 1
Post image

New preprint! πŸ§ πŸ€–

How do we build neural decoders that are:
⚑️ fast enough for real-time use
🎯 accurate across diverse tasks
🌍 generalizable to new sessions, subjects, and even species?

We present POSSM, a hybrid SSM architecture that optimizes for all three of these axes!

🧡1/7

06.06.2025 17:40 β€” πŸ‘ 54    πŸ” 24    πŸ’¬ 2    πŸ“Œ 8
Video thumbnail

🧡(1/7) Have you ever wanted to combine different pre-trained diffusion models but don't have time or data to retrain a new, bigger model?

πŸš€ Introducing SuperDiff πŸ¦Ήβ€β™€οΈ – a principled method for efficiently combining multiple pre-trained diffusion models solely during inference!

28.12.2024 14:32 β€” πŸ‘ 44    πŸ” 7    πŸ’¬ 1    πŸ“Œ 4
Post image

πŸ”Š Super excited to announce the first ever Frontiers of Probabilistic Inference: Learning meets Sampling workshop at #ICLR2025 @iclr-conf.bsky.social!

πŸ”— website: sites.google.com/view/fpiwork...

πŸ”₯ Call for papers: sites.google.com/view/fpiwork...

more details in thread belowπŸ‘‡ 🧡

18.12.2024 19:09 β€” πŸ‘ 84    πŸ” 19    πŸ’¬ 2    πŸ“Œ 3

Now you can generate equilibrium conformations for your small molecule in 3 lines of code with ET-Flow! Awesome effort put in by @fntwin.bsky.social!

12.12.2024 16:37 β€” πŸ‘ 13    πŸ” 3    πŸ’¬ 0    πŸ“Œ 1

ET-Flow shows, once again, that equivariance is better than Transformer when physical precision matters!

come see us at @neuripsconf.bsky.social !!

07.12.2024 15:57 β€” πŸ‘ 6    πŸ” 1    πŸ’¬ 0    πŸ“Œ 0

Excited to share our work! I had a wonderful time collaborating with these brilliant people

07.12.2024 16:01 β€” πŸ‘ 4    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

@majhas is following 20 prominent accounts