π§΅ Everyone is chasing new diffusion modelsβbut what about the representations they model from?
We introduce Discrete Latent Codes (DLCs):
- Discrete representation for diffusion models
- Uncond. gen. SOTA FID (1.59 on ImageNet)
- Compositional generation
- Integrates with LLM
π§±
22.07.2025 14:41 β π 5 π 3 π¬ 1 π 0
New preprint! π§ π€
How do we build neural decoders that are:
β‘οΈ fast enough for real-time use
π― accurate across diverse tasks
π generalizable to new sessions, subjects, and even species?
We present POSSM, a hybrid SSM architecture that optimizes for all three of these axes!
π§΅1/7
06.06.2025 17:40 β π 51 π 23 π¬ 2 π 7
Preprint Alert π
Can we simultaneously learn transformation-invariant and transformation-equivariant representations with self-supervised learning?
TL;DR Yes! This is possible via simple predictive learning & architectural inductive biases β without extra loss terms and predictors!
π§΅ (1/10)
14.05.2025 12:52 β π 51 π 15 π¬ 1 π 4
This can be a game changer for embodied #NeuroAI.
Or it *could* be, if it were open source.
Just imagine the resources it takes to develop an open version of this model. Now think about how much innovation could come from building on this, rather than just trying to recreate it (at best).
04.12.2024 17:01 β π 37 π 8 π¬ 3 π 0
See my inner physicist hates the whole "doesn't matter as long as it works" sentiment in the ML community π. I want to UNDERSTAND not just accept... jokes aside though I see your point for the purposes of this discussion. I think we've identified a lot of potential in this stream of inquiry π§
22.11.2024 21:43 β π 1 π 0 π¬ 1 π 0
That's somewhat along the lines of what I was thinking as well :)
Also good point about o1. I'd be very interested to see how it performs on the ToM tests!
22.11.2024 21:31 β π 1 π 0 π¬ 0 π 0
Give the results and discussion a read as well it's super interesting! There's reason to believe perfect performance of Llama on the faux pas test was illusory (expanded upon in the discussion). That bias you mention is also elaborated upon in the discussion (and I briefly summarize above).
22.11.2024 21:30 β π 1 π 0 π¬ 1 π 0
This all now begs the question of whether this makes LLMs more or less competent as practitioners of therapy. I think good arguments could be made for both perspectives. π§΅/fin
22.11.2024 20:37 β π 2 π 0 π¬ 1 π 0
This fact is of course unsurprising (as the authors admit) since humanity's embodiment has placed evolutionary pressure on resolving these uncertainties (i.e. to fight or to flee). This dis-embodiment of LLMs could prevent their commitment to the most likely explanation. π§΅/2
22.11.2024 20:36 β π 1 π 0 π¬ 1 π 0
I stand corrected. However, LLM's failure at the faux pas test underscores the need for further discussion. The failure: "not comput[ing] [mentalistic-like] inferences spontaneously to reduce uncertainty". LLMs are good at emulating human-responses, but the underlying cognition is different. π§΅/1
22.11.2024 20:35 β π 1 π 0 π¬ 1 π 0
I'd argue that until LLMs can implement theory of mind, they'd be much better at diagnostic-oriented therapy. Being able to truly understand a human, form hypotheses, and guide a patient towards resolution is very different from recommending treatment based off a checklist made using the DSM.
22.11.2024 15:26 β π 3 π 0 π¬ 1 π 0
1/ I work in #NeuroAI, a growing field of research, which many people have only the haziest conception of...
As way of introduction to this research approach, I'll provide here a very short thread outlining the definition of the field I gave recently at our BRAIN NeuroAI workshop at the NIH.
π§ π
21.11.2024 16:20 β π 169 π 48 π¬ 8 π 12
I'm making an unofficial starter pack with some of my colleagues at Mila. WIP for now but here's the link!
go.bsky.app/BHKxoss
20.11.2024 15:19 β π 69 π 29 π¬ 7 π 1
Mind if I wiggle my way into this π
20.11.2024 16:16 β π 1 π 0 π¬ 1 π 0
From double descent to grokking, deep learning sometimes works in unpredictable ways.. or does it?
For NeurIPS(my final PhD paper!), @alanjeffares.bsky.social & I explored if&how smart linearisation can help us better understand&predict numerous odd deep learning phenomena β and learned a lot..π§΅1/n
18.11.2024 19:25 β π 175 π 35 π¬ 7 π 7
Neuroscience PhD student at McGill
Co-supervised by Adrien Peyrache & Blake Richards
Ph.D. Student @mila-quebec.bsky.social and @umontreal.ca, AI Researcher
PI @ Salk Institute β talmolab.org | PhD @Princeton | BS @ UMBC
Quantifying biological motion with sleap.ai
Graduate student, Clinical Psychology & Neuropsychology @ York University
Scientist, mentor, activist, explorer.
Professor @ UniversitΓ© de Montreal & Canada Research Chair in Comp Neuroscience & Cog Neuroimaging. Director of the Quebec Neuro-AI research center (UNIQUE Centre) | Biological & Artificial Cognition | Consciousness | Creativity | π§ π€ he/him
https://mcgill-nlp.github.io/people/
AI Scientist at Xaira Therapeutics & PhD student at Mila
graduate researcher at McGill/MILA, unlearning and privacy
PhD candidate at Mila, formerly Google eng, Brown University. Making LLMs explorative, adaptive, and goal-oriented
CS Faculty at Mila and McGill, interested in Graphs and Complex Data, AI/ML, Misinformation, Computational Social Science and Online Safety
#RL Postdoc at Mila - Quebec AI Institute and UniversitΓ© de MontrΓ©al
Graduate student @Mila_Quebec @UMontrealDIRO | RL/Deep Learning/AI | De Cali/Colombia palβ Mundo π¨π΄ | #JuntosProsperamosβ‘#TogetherWeThrive| π±π
Postdoc @McGill & Mila | Co-founder & CTO @Rubisco.ai | Core member @Climate Change AI
PhD @ Mila / UdeM | NeuroAI research