Aidan Sirbu's Avatar

Aidan Sirbu

@sirbu.bsky.social

MSc Student @ Mila and McGill ML & NeuroAI research

218 Followers  |  233 Following  |  14 Posts  |  Joined: 18.11.2024  |  2.2748

Latest posts by sirbu.bsky.social on Bluesky

Image of robots struggling with a social dilemma.

Image of robots struggling with a social dilemma.

1/ Why does RL struggle with social dilemmas? How can we ensure that AI learns to cooperate rather than compete?

Introducing our new framework: MUPI (Embedded Universal Predictive Intelligence) which provides a theoretical basis for new cooperative solutions in RL.

PreprintπŸ§΅πŸ‘‡

(Paper link below.)

03.12.2025 19:19 β€” πŸ‘ 65    πŸ” 26    πŸ’¬ 5    πŸ“Œ 5

In this piece for @thetransmitter.bsky.social, I argue that ecological neuroscience should leverage generative video and interactive models to simulate the world from animals' perspectives.

The technological building blocks are almost here - we just need to align them for this application.

πŸ§ πŸ€–

08.12.2025 15:59 β€” πŸ‘ 40    πŸ” 14    πŸ’¬ 0    πŸ“Œ 1
Post image

Hope to see you at the @neuripsconf.bsky.social @dataonbrainmind.bsky.social Workshop! I will presenting our poster β€œMassively Parallel Imitation Learning of Mouse Forelimb Musculoskeletal Reaching Dynamics” @talmo.bsky.social @eazim.bsky.social Dec 7, SDCC 🦾🧠
arxiv.org/abs/2511.21848

04.12.2025 19:44 β€” πŸ‘ 5    πŸ” 2    πŸ’¬ 0    πŸ“Œ 1
Post image

🚨Our work was accepted to the @neuripsconf.bsky.social : Data on the Brain & Mind workshop!🧠🦾"Massively Parallel Imitation Learning of Mouse Forelimb Musculoskeletal Reaching Dynamics" @talmo.bsky.social @eazim.bsky.social
An imitation learning framework for modeling mouse forelimb control. 1/3

28.11.2025 17:01 β€” πŸ‘ 11    πŸ” 3    πŸ’¬ 1    πŸ“Œ 0
Post image

Talmo Lab is at @sfn.org! Come check out our latest work!

#neuroskyence #neurosky #SfN2025 #SfN25

15.11.2025 22:13 β€” πŸ‘ 17    πŸ” 6    πŸ’¬ 0    πŸ“Œ 1
New paper titled "Tracing the Representation Geometry of Language Models from Pretraining to Post-training" by Melody Z Li, Kumar K Agrawal, Arna Ghosh, Komal K Teru, Adam Santoro, Guillaume Lajoie, Blake A Richards.

New paper titled "Tracing the Representation Geometry of Language Models from Pretraining to Post-training" by Melody Z Li, Kumar K Agrawal, Arna Ghosh, Komal K Teru, Adam Santoro, Guillaume Lajoie, Blake A Richards.

LLMs are trained to compress data by mapping sequences to high-dim representations!
How does the complexity of this mapping change across LLM training? How does it relate to the model’s capabilities? πŸ€”
Announcing our #NeurIPS2025 πŸ“„ that dives into this.

🧡below
#AIResearch #MachineLearning #LLM

31.10.2025 16:19 β€” πŸ‘ 60    πŸ” 12    πŸ’¬ 1    πŸ“Œ 4

Over the past year, my lab has been working on fleshing out theory + applications of the Platonic Representation Hypothesis.

Today I want to share two new works on this topic:

Eliciting higher alignment: arxiv.org/abs/2510.02425
Unpaired learning of unified reps: arxiv.org/abs/2510.08492

1/9

10.10.2025 22:13 β€” πŸ‘ 131    πŸ” 32    πŸ’¬ 1    πŸ“Œ 5
Preview
The curriculum effect in visual learning: the role of readout dimensionality Generalization of visual perceptual learning (VPL) to unseen conditions varies across tasks. Previous work suggests that training curriculum may be integral to generalization, yet a theoretical explan...

🚨 New preprint alert!

πŸ§ πŸ€–
We propose a theory of how learning curriculum affects generalization through neural population dimensionality. Learning curriculum is a determining factor of neural dimensionality - where you start from determines where you end up.
πŸ§ πŸ“ˆ

A 🧡:

tinyurl.com/yr8tawj3

30.09.2025 14:25 β€” πŸ‘ 77    πŸ” 25    πŸ’¬ 1    πŸ“Œ 2

A big congratulations to my supervisor for this awesome achievement. Excited to see where this will go!

23.09.2025 16:27 β€” πŸ‘ 3    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

5) Finally, I don't use it for writing as much as my peers. But its quite nice asking it how to say things like "this is best exemplified by..." in a different way so I don't repeat the same thing a million times in my paper or scholarship application.

04.09.2025 15:57 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

4) Insofar as idea synthesis, I find my conversations with LLMs about as useful as talking to a layman with decent critical thinking skills and access to google. Its nice to bounce ideas off it at 1am when my colleagues may be asleep. But conversing with other academics is still by far more useful.

04.09.2025 15:54 β€” πŸ‘ 2    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

3) I am very grateful I learned how to code before the advent of LLMs. I think there's a real concern of new students relying too heavily on LLMs for coding, foregoing the learning process. At least for the time being, in order to use LLMs effectively, one still needs a strong foundation in coding.

04.09.2025 15:51 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

2) When I start coding new projects, getting copilot to draft up the initial commit saves me loads of time. Of course there will be quite a few errors and/or silly coding structure. But I find tracing through the logic and making necessary corrections to be quicker than starting from scratch

04.09.2025 15:48 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

I'm a graduate student breaking into the field of ML from physics.

1) I find LLMs useful insofar as gaining a basic understanding of new concepts. However going past a basic understanding still requires delving deep into the literature. I find the back-and-forth tutor style conversations useful.

04.09.2025 15:40 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Post image

🧡 Everyone is chasing new diffusion modelsβ€”but what about the representations they model from?
We introduce Discrete Latent Codes (DLCs):
- Discrete representation for diffusion models
- Uncond. gen. SOTA FID (1.59 on ImageNet)
- Compositional generation
- Integrates with LLM
🧱

22.07.2025 14:41 β€” πŸ‘ 5    πŸ” 3    πŸ’¬ 1    πŸ“Œ 0
Post image

New preprint! πŸ§ πŸ€–

How do we build neural decoders that are:
⚑️ fast enough for real-time use
🎯 accurate across diverse tasks
🌍 generalizable to new sessions, subjects, and even species?

We present POSSM, a hybrid SSM architecture that optimizes for all three of these axes!

🧡1/7

06.06.2025 17:40 β€” πŸ‘ 53    πŸ” 24    πŸ’¬ 2    πŸ“Œ 8
Post image

Preprint Alert πŸš€

Can we simultaneously learn transformation-invariant and transformation-equivariant representations with self-supervised learning?

TL;DR Yes! This is possible via simple predictive learning & architectural inductive biases – without extra loss terms and predictors!

🧡 (1/10)

14.05.2025 12:52 β€” πŸ‘ 51    πŸ” 16    πŸ’¬ 1    πŸ“Œ 5

This can be a game changer for embodied #NeuroAI.

Or it *could* be, if it were open source.

Just imagine the resources it takes to develop an open version of this model. Now think about how much innovation could come from building on this, rather than just trying to recreate it (at best).

04.12.2024 17:01 β€” πŸ‘ 37    πŸ” 8    πŸ’¬ 3    πŸ“Œ 0

See my inner physicist hates the whole "doesn't matter as long as it works" sentiment in the ML community πŸ˜‚. I want to UNDERSTAND not just accept... jokes aside though I see your point for the purposes of this discussion. I think we've identified a lot of potential in this stream of inquiry 🧐

22.11.2024 21:43 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

That's somewhat along the lines of what I was thinking as well :)

Also good point about o1. I'd be very interested to see how it performs on the ToM tests!

22.11.2024 21:31 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

Give the results and discussion a read as well it's super interesting! There's reason to believe perfect performance of Llama on the faux pas test was illusory (expanded upon in the discussion). That bias you mention is also elaborated upon in the discussion (and I briefly summarize above).

22.11.2024 21:30 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

This all now begs the question of whether this makes LLMs more or less competent as practitioners of therapy. I think good arguments could be made for both perspectives. 🧡/fin

22.11.2024 20:37 β€” πŸ‘ 2    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

This fact is of course unsurprising (as the authors admit) since humanity's embodiment has placed evolutionary pressure on resolving these uncertainties (i.e. to fight or to flee). This dis-embodiment of LLMs could prevent their commitment to the most likely explanation. 🧡/2

22.11.2024 20:36 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

I stand corrected. However, LLM's failure at the faux pas test underscores the need for further discussion. The failure: "not comput[ing] [mentalistic-like] inferences spontaneously to reduce uncertainty". LLMs are good at emulating human-responses, but the underlying cognition is different. 🧡/1

22.11.2024 20:35 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Preview
Uniform manifold approximation and projection Nature Reviews Methods Primers - Uniform manifold approximation and projection is a dimensionality reduction technique used to visualize and understand high-dimensional data. In this Primer, Healy...

I recently wrote a primer on UMAP for Nature Reviews Primers. If you are looking for an overview of the method, a getting started primer, or best practices it is a good place to start.

rdcu.be/d0YZT

22.11.2024 00:02 β€” πŸ‘ 111    πŸ” 36    πŸ’¬ 2    πŸ“Œ 2

I'd argue that until LLMs can implement theory of mind, they'd be much better at diagnostic-oriented therapy. Being able to truly understand a human, form hypotheses, and guide a patient towards resolution is very different from recommending treatment based off a checklist made using the DSM.

22.11.2024 15:26 β€” πŸ‘ 3    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

1/ I work in #NeuroAI, a growing field of research, which many people have only the haziest conception of...

As way of introduction to this research approach, I'll provide here a very short thread outlining the definition of the field I gave recently at our BRAIN NeuroAI workshop at the NIH.

πŸ§ πŸ“ˆ

21.11.2024 16:20 β€” πŸ‘ 167    πŸ” 48    πŸ’¬ 8    πŸ“Œ 12

I'm making an unofficial starter pack with some of my colleagues at Mila. WIP for now but here's the link!

go.bsky.app/BHKxoss

20.11.2024 15:19 β€” πŸ‘ 70    πŸ” 29    πŸ’¬ 8    πŸ“Œ 1

Mind if I wiggle my way into this πŸ›

20.11.2024 16:16 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Post image

From double descent to grokking, deep learning sometimes works in unpredictable ways.. or does it?

For NeurIPS(my final PhD paper!), @alanjeffares.bsky.social & I explored if&how smart linearisation can help us better understand&predict numerous odd deep learning phenomena β€” and learned a lot..🧡1/n

18.11.2024 19:25 β€” πŸ‘ 174    πŸ” 35    πŸ’¬ 7    πŸ“Œ 7

@sirbu is following 20 prominent accounts