Shahab Bakhtiari's Avatar

Shahab Bakhtiari

@shahabbakht.bsky.social

|| assistant prof at University of Montreal || leading the systems neuroscience and AI lab (SNAIL: https://www.snailab.ca/) 🐌 || associate academic member of Mila (Quebec AI Institute) || #NeuroAI || vision and learning in brains and machines

6,342 Followers  |  1,117 Following  |  1,386 Posts  |  Joined: 05.08.2023  |  2.0746

Latest posts by shahabbakht.bsky.social on Bluesky


Preview
Vectorized instructive signals in cortical dendrites - Nature Mice learning a neurofeedback brain–computer interface task show neuron-specific teaching signals in cortical dendrites, consistent with a vectorized solution for credit assignment in the brain.

This paper on how the brain may do gradient descent is very cool: www.nature.com/articles/s41...

26.02.2026 03:02 β€” πŸ‘ 136    πŸ” 41    πŸ’¬ 2    πŸ“Œ 2
Preview
Pace of ecology drives the tempo of visual perception across the animal kingdom Nature Ecology & Evolution - Using phylogenetic comparative methods across 237 species from disparate phyla, the authors show that species with fast-paced ecologies have higher temporal...

Our new paper is now out showing how time perception in animals is linked to their ecology. Using data from 237 species we show temporal perception is faster in species that fly and pursuit predators www.nature.com/articles/s41... 🌐

24.02.2026 13:22 β€” πŸ‘ 119    πŸ” 53    πŸ’¬ 3    πŸ“Œ 2

This study is super cool (connecting ecology and perception), that suggest some aspects of animal's perception (temporal precision) is shaped by their environment (which somehow resonates w our proposal on internal foraging perspectives on perceptual selection www.sciencedirect.com/science/arti...)

24.02.2026 13:51 β€” πŸ‘ 28    πŸ” 10    πŸ’¬ 0    πŸ“Œ 0

This is a critical methodological point about the Platonic Representation Hypothesis paper.

I mistakenly thought the PRH paper used CKA as its main similarity metric.

Another motivation for thinking more deeply about metrics of similarity and alignment.

25.02.2026 04:12 β€” πŸ‘ 4    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

Though it’s quite interesting that this subtle methodological detail turned out to be so important in the main message.

25.02.2026 03:10 β€” πŸ‘ 3    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

Thanks for the correction!

Yes, in the main text, your paper mainly relied on local similarity.

Now actually I rememberer, my first reaction reading your paper was why CKA results weren’t used in the main text.

25.02.2026 03:10 β€” πŸ‘ 3    πŸ” 0    πŸ’¬ 2    πŸ“Œ 0
Preview
Humans and neural networks show similar patterns of transfer and interference during continual learning - Nature Human Behaviour When learning new tasks, both humans and artificial neural networks face a trade-off between reusing prior knowledge to learn faster and avoiding the disruption of earlier learning. This study shows t...

What you describe sounds like the "lumpers vs. splitters" in this paper from @summerfieldlab.bsky.social lab: lumpers generalize more/retain less, and splitters generalize less/forget more. They gave a nice explanation based on rich vs lazy training regimes in ANNs.

www.nature.com/articles/s41...

24.02.2026 01:25 β€” πŸ‘ 4    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

Looking forward to reading the promised post on continual learning, @lampinen.bsky.social :)

24.02.2026 01:15 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

Cool! This is generative inference’s prediction for human perception in this illusion: the squares are no longer squares!
Try it for yourself here:
huggingface.co/spaces/ttoos...

22.02.2026 04:50 β€” πŸ‘ 21    πŸ” 6    πŸ’¬ 2    πŸ“Œ 0
Preview
Memorization vs. generalization in deep learning: implicit biases, benign overfitting, and more Or: how I learned to stop worrying and love the memorization

What is the relationship between memorization and generalization in AI? Is there a fundamental tradeoff? In infinitefaculty.substack.com/p/memorizati... I’ve reviewed some of the evolving perspectives on memorization & generalization in machine learning, from classic perspectives through LLMs.

18.02.2026 15:54 β€” πŸ‘ 132    πŸ” 27    πŸ’¬ 4    πŸ“Œ 5
Preview
Revisiting the Platonic Representation Hypothesis: An Aristotelian View The Platonic Representation Hypothesis suggests that representations from neural networks are converging to a common statistical model of reality. We show that the existing metrics used to measure rep...

Another good reason for being cautious with representational similarity analysis: arxiv.org/abs/2602.14486

The famous Platonic Representation Hypothesis was largely driven by CKA’s bias.

But the hypothesis still holds for shared local relationships.

18.02.2026 02:03 β€” πŸ‘ 8    πŸ” 0    πŸ’¬ 1    πŸ“Œ 1

The revised version of our paper on the impact of top-down feedback is now out @elife.bsky.social:

doi.org/10.7554/eLif...

tl;dr: we show that using human-brain-like feedback/anatomy in a deep RNN leads to human-like visual biases!

This work was led by @tmshbr.bsky.social

#NeuroAI πŸ§ πŸ“ˆ πŸ§ͺ

17.02.2026 22:54 β€” πŸ‘ 54    πŸ” 14    πŸ’¬ 0    πŸ“Œ 0

Excited to launch Principia, a nonprofit research organisation at the intersection of deep learning theory and AI safety.

Our goal is to develop theory for modern machine learning systems that can help us understand complex network behaviors, including those critical for AI safety and alignment.

1

16.02.2026 09:27 β€” πŸ‘ 91    πŸ” 26    πŸ’¬ 1    πŸ“Œ 1

Thrilled to finally share this work! πŸ§ πŸ”Š

Using a new reinforcement-free task we show mice (like humans) extract abstract structure from sound (unsupervised) & dCA1 is causally required by building factorised, orthogonal subspaces of abstract rules.

Led by Dammy Onih!
www.biorxiv.org/content/10.6...

16.02.2026 13:01 β€” πŸ‘ 149    πŸ” 52    πŸ’¬ 3    πŸ“Œ 2

I don’t think AI’s success in coding will automatically translate to other fields. That level of performance only works where the output is as easily verifiable as code; and not many domains fit that bill. 2/2

11.02.2026 16:16 β€” πŸ‘ 5    πŸ” 1    πŸ’¬ 0    πŸ“Œ 0
Preview
Something big is happening in AI β€” and most people will be blindsided | Fortune It’s not like a light switch... more like the moment you realize the water has been rising around you and is now at your chest.

"The experience that tech workers have had over the past year, of watching AI go from β€œhelpful tool” to β€œdoes my job better than I do”, is the experience everyone else is about to have. Law, finance, medicine, accounting, …"

I’m not sure … 1/2

fortune.com/2026/02/11/s...

11.02.2026 16:16 β€” πŸ‘ 2    πŸ” 1    πŸ’¬ 1    πŸ“Œ 0

… especially whenever controversies around representational similarity resurface.

11.02.2026 15:06 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

You’re comparing two fields at very different stages of theoretical maturity. Neuroscience (and NeuroAI) are still largely pre-theoretic. I often return to Hasok Chang’s Inventing Temperature as a parallel for where we actually stand in theoretical neuroscience, …

11.02.2026 15:06 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

Definitely not enough.

10.02.2026 22:05 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

My bet is on the ecological relevance of training data and temporal prediction as the core objective.

Architecture is difficult to constrain, given that ANNs and brains rely on substantially different functional mechanisms.

That’s just my view, though; I could be wrong.

10.02.2026 21:50 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

Exactly my point. The emerging view seems to be that, assuming equal trainability (big assumption though), the architectures may not play as big of a role as the training objective and data.

10.02.2026 21:28 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 2    πŸ“Œ 0

The problem is the training objective and lack of recurrence, not the 50-layer architecture.

10.02.2026 21:11 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

Also see @mschrimpf.bsky.social thread here: bsky.app/profile/msch...

10.02.2026 18:43 β€” πŸ‘ 3    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

My two cents: as with any discipline, we use tools to probe phenomena scientifically while simultaneously striving to understand those tools better

An iterative process that sees the works like the one above within its paradigm not standing outside of it.

10.02.2026 18:42 β€” πŸ‘ 3    πŸ” 1    πŸ’¬ 2    πŸ“Œ 0

I encourage everyone to read the paper and rather than relying on social media impressions.

It’s an excellent paper speaking about a specific category of work in NeuroAI that is still hard to generalize to the whole field.

10.02.2026 18:42 β€” πŸ‘ 6    πŸ” 1    πŸ’¬ 1    πŸ“Œ 0

You mean a framework that could build models of artificial stimuli would be sufficient?

10.02.2026 18:25 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Preview
In praise of artifice - Nature Neuroscience Nature Neuroscience - In praise of artifice

Great point.

I’d use classical stimuli for testing out-of-distribution generalization instead of model development.

This would actually be the exact opposite of what was proposed in "In praise of artifice"

www.nature.com/articles/nn1...

10.02.2026 17:44 β€” πŸ‘ 2    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Post image

Our paper is out in @natneuro.nature.com!

www.nature.com/articles/s41...

We develop a geometric theory of how neural populations support generalization across many tasks.

@zuckermanbrain.bsky.social
@flatironinstitute.org
@kempnerinstitute.bsky.social

1/14

10.02.2026 15:56 β€” πŸ‘ 271    πŸ” 99    πŸ’¬ 7    πŸ“Œ 1

I agree. My main point is that decompositionality (whether or not it supports modularity) is baked into classical stimuli a priori. These stimuli then act as an inductive bias in models developed to capture the resulting neural or behavioral responses.

10.02.2026 16:25 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

n other words, classic stimuli rely on a strong notion of modularity that may or may not hold for naturalistic stimuli, where visual features are inherently intermixed.

10.02.2026 16:12 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

@shahabbakht is following 20 prominent accounts