GitHub - ddidacus/mol-moe: Repository for: "Training Preference-Guided Routers for Molecule Generation"
Repository for: "Training Preference-Guided Routers for Molecule Generation" - ddidacus/mol-moe
Special thanks to Biogen and CIFAR for the support, and
@proceduralia.bsky.social + @pierrelucbacon.bsky.social
for their valuable supervision, and to the entire Mila community for their feedback, discussions, and support. Code, paper, and models are public: github.com/ddidacus/mol...
20.02.2025 19:43 โ ๐ 3 ๐ 1 ๐ฌ 0 ๐ 0
Mol-MoE improves with more property experts with a larger gain than classic merging and overall, it achieves the highest scores. Simple reward scalarization here does not work. We aim at further calibrating Mol-MoE and testing the performance on larger sets of objectives.
20.02.2025 19:43 โ ๐ 1 ๐ 0 ๐ฌ 1 ๐ 0
The model we obtain does achieve a smaller mean absolute error in generating compounds according to the provided properties, surpassing the alternative methods. Arguably, the learned routing functions can tackle task interference.
20.02.2025 19:43 โ ๐ 1 ๐ 0 ๐ฌ 1 ๐ 0
But the relationship between interpolation coefficients and properties isnโt strictly linear, needing a calibration function. Mol-MoE addresses this by training only the routers to predict optimal merging weights from prompts, enabling more precise control and less interference.
20.02.2025 19:43 โ ๐ 1 ๐ 0 ๐ฌ 1 ๐ 0
Think, think, think... what if we trained experts on single properties separately and leveraged model merging techniques to obtain a multi-property model? We re-implement rewarded soups and obtain a robust baseline capable of generating high-quality, out-of-distribution samples.
20.02.2025 19:43 โ ๐ 1 ๐ 0 ๐ฌ 1 ๐ 0
In our ablation studies, instruction-tuned models struggle with higher property values due to lack of explicit optimization. Even RL fine-tuning on multiple objectives can hit performance plateaus or declines, and balancing objectives requires re-training, limiting steerability.
20.02.2025 19:43 โ ๐ 1 ๐ 0 ๐ฌ 1 ๐ 0
Drug discovery inherently involves multi-objective optimization, requiring candidate molecules to not only bind effectively to target proteins, triggering a specific function, but also to meet safety and compatibility criteria to become drugs. Is supervised learning sufficient?
20.02.2025 19:43 โ ๐ 1 ๐ 0 ๐ฌ 1 ๐ 0
Molecule sequence models learn vast molecular spaces, but how to navigate them efficiently? We explored multi-objective RL, SFT, merging, but these fall short in balancing control and diversity. We introduce **Mol-MoE**: a mixture of experts for controllable molecule generation๐งต
20.02.2025 19:43 โ ๐ 1 ๐ 1 ๐ฌ 1 ๐ 0
Logically Consistent Language Models via Neuro-Symbolic Integration
Large language models (LLMs) are a promising venue for natural language understanding and generation. However, current LLMs are far from reliable: they are prone to generating non-factual information ...
Finally, LOgically COnsistent (LoCo) LLaMas can outperform solver-based baselines and SFT! I thank @nolovedeeplearning.bsky.social and @looselycorrect.bsky.social for the guidance in realizing this project, get in touch or come to chat in Singapore!
arxiv.org/abs/2409.13724
29.01.2025 23:41 โ ๐ 6 ๐ 2 ๐ฌ 0 ๐ 0
Our method makes LLaMa's knowledge more consistent to any given knowledge graph, by seeing only a portion of it! It can transfer logical rules to similar or derived concepts. As proposed by @ekinakyurek.bsky.social et al., you can use a LLM-generated KB to reason over its knowledge.
29.01.2025 23:41 โ ๐ 7 ๐ 1 ๐ฌ 2 ๐ 0
Yes! We propose to leverage the Semantic Loss as a regularizer: it maximizes the likelihood of world (model) assignments satisfying any given logical rule. We thus include efficient solvers in the training pipeline to efficiently perform model counting on the LLM's own beliefs.
29.01.2025 23:41 โ ๐ 3 ๐ 0 ๐ฌ 1 ๐ 0
Various background works focus on instilling single consistency rules, e.g. A and not A can't be both true (negation, Burns et al.), A true and A implies B, thus B true (modus ponens). Can we derive a general objective function that combines logical rules dynamically?
29.01.2025 23:41 โ ๐ 2 ๐ 0 ๐ฌ 1 ๐ 0
๐ฅณ "Logically Consistent Language Models via Neuro-Symbolic Integration" just accepted at #ICLR2025!
We focus on instilling logical rules in LLMs with an efficient loss, leading to higher factuality & (self) consistency. How? ๐งต
29.01.2025 23:41 โ ๐ 13 ๐ 2 ๐ฌ 1 ๐ 0
RNA FISH -> a fish
03.12.2024 13:13 โ ๐ 56 ๐ 3 ๐ฌ 2 ๐ 0
used Cursor (based on claude sonnet 3.5) over VS Code for a week now. Early feedback:
โ๏ธ great to parallelize training and inference
โ๏ธ multi-file context, can easily setup hyperparam sweeps
โ๏ธ great to visualize results with high level guidance. Welcome spider plots!
29.11.2024 17:34 โ ๐ 1 ๐ 0 ๐ฌ 0 ๐ 0
Announcing the NeurIPS 2024 Test of Time Paper Awardsย โ NeurIPS Blog
Test of Time Paper Awards are out! 2014 was a wonderful year with lots of amazing papers. That's why, we decided to highlight two papers: GANs (@ian-goodfellow.bsky.social et al.) and Seq2Seq (Sutskever et al.). Both papers will be presented in person ๐
Link: blog.neurips.cc/2024/11/27/a...
27.11.2024 15:48 โ ๐ 110 ๐ 14 ๐ฌ 1 ๐ 2
I guess it also depends on the field/subfield?
23.11.2024 20:50 โ ๐ 0 ๐ 0 ๐ฌ 0 ๐ 0
researchers on cancer, message me: Iโd like to know about your work, your research questions!
23.11.2024 20:49 โ ๐ 3 ๐ 1 ๐ฌ 0 ๐ 0
YouTube video by EEML Community
[EEML'24] Sander Dieleman - Generative modelling through iterative refinement
While we're starting up over here, I suppose it's okay to reshare some old content, right?
Here's my lecture from the EEML 2024 summer school in Novi Sad๐ท๐ธ, where I tried to give an intuitive introduction to diffusion models: youtu.be/9BHQvQlsVdE
Check out other lectures on their channel as well!
19.11.2024 09:57 โ ๐ 115 ๐ 12 ๐ฌ 3 ๐ 0
I've created an initial Grumpy Machine Learners starter park. If you think you're grumpy and you "do machine learning", nominate yourself. If you're on the list, but don't think you are grumpy, then take a look in the mirror.
go.bsky.app/6ddpivr
18.11.2024 14:40 โ ๐ 418 ๐ 55 ๐ฌ 124 ๐ 15
Waitress turned Congresswoman for the Bronx and Queens. Grassroots elected, small-dollar supported. A better world is possible.
ocasiocortez.com
Parker Distinguished Professor, @UNC. Program Chair #EMNLP2024. Director http://MURGeLab.cs.unc.edu (@uncnlp). @Berkeley_AI @TTIC_Connect @IITKanpur
#NLP #CV #AI #ML
https://www.cs.unc.edu/~mbansal/
19th International conference on Neurosymbolic Learning and Reasoning
UC Santa Cruz, Santa Cruz, California
8 to 10 September 2025
https://2025.nesyconf.org
I make sure that OpenAI et al. aren't the only people who are able to study large scale AI systems.
Postdoctoral researcher at McGill in #AI #ML core developer of SpeechBrain.
Studies analysis of speech patterns for bio markers, for speech enhancement, robust ASR, continual learning, etc.
proud mediterrenean ๐งฟ open-sourceress at hugging face ๐ค multimodality, zero-shot vision, vision language models, transformers
Research Scientist at DeepMind. Opinions my own. Inventor of GANs. Lead author of http://www.deeplearningbook.org . Founding chairman of www.publichealthactionnetwork.org
Research Scientist@Google DeepMind
Assoc Prof@York University, Toronto
mbrubake.github.io
Research: Computer Vision and Machine Learning, esp generative models.
Applications: CryoEM (cryoSPARC), Statistics (Stan), Forensics, and more
Ph.D. student in Artificial Intelligence at the University of Trento.
Research Fellow from ๐ฎ๐น | Multi-Agent Reinforcement Learning
ML Team MSR Montreal. Adjunct Prof UdeM MILA. Modularity & reasoning.
Professor, University of Lausanne ๐จ๐ญ and Ludwig Institute for Cancer Research
President-Elect, European Association for Cancer Research
Mentor, Teacher, Advocate, Mum
@joycelab.bsky.social
Professor @ Oxford. We use zebrafish to study how cancer starts and spreads.
Investigating #Aneuploidy in #Cancer @TelAvivUniv
Cancer-Epi|Genetics|Genomics|Proteomics and Oncogenic Fusion Proteins
@vetmedunivienna @StAnna_CCRI @CeMM_News
Thoracic, Head and Neck oncologist #medsky #oncsky
Fred Hutch Cancer Center. @fredhutch.bsky.social
University of Washington.
Writer & mom in Florida.
> Probably cured of incurable cancer (MBC) by curative-intent treatment.
> LOVE my Aesthetic Flat Closure.
> Daughter of Oanh & Larry
Postdoc @IREC_UCLouvain ๐ง๐ช CRC, CSCs & metabolism | PhD @Unistra ๐ซ๐ท | Pharm, MSc Biotech @fbioyfunr ๐ฆ๐ท๐ฎ๐น |@FIP_NGPS ViceChair, @EACR Ambassador