But this example again appears not to involve the reasoning mode. I agree that βthinkingβ is confusing nomenclature but itβs notable that most (not all) of the stupidest mistakes come from the feedforward / parallel processing mode.
22.02.2026 21:21 β
π 0
π 0
π¬ 0
π 0
Understanding the Limits of Vision Language Models Through the Lens of the Binding Problem
Recent work has documented striking heterogeneity in the performance of state-of-the-art vision language models (VLMs), including both multimodal language models and text-to-image models. These models...
We also have a paper on this arxiv.org/abs/2411.00238 but it doesnβt seem to be an arbitrary failure. Instead they seem to fail in precisely the ways that human vision fails under time pressure (including with counting), and increasingly the models seem to resolve this via sequential processing.
22.02.2026 21:17 β
π 2
π 0
π¬ 1
π 0
Re: seeing vs. thinking, 'thinking' is arguably a bad term for this, in the vision setting the models depend on sequential processing to individuate objects but that wouldn't commonly be referred to as 'thinking' in the colloquial sense.
22.02.2026 20:27 β
π 0
π 0
π¬ 1
π 0
Huh, seems to be the result of different prompts, although it's arguably confusing to say that no liquid can be poured into it (at all or only in the current configuration?). In general most of the comically stupid mistakes (e.g. how many b's in blueberry) seem to be from the non-thinking models.
22.02.2026 20:23 β
π 0
π 0
π¬ 1
π 0
Funny, but these demos always seem to be the free / instant model. With thinking turned on it gets this correct.
22.02.2026 18:48 β
π 0
π 0
π¬ 1
π 0
very interesting! Consistent with this, we found that induction heads seem to be completely distinct from what we called 'symbolic induction heads' i.e. function vector heads arxiv.org/abs/2502.20332
22.02.2026 01:36 β
π 2
π 0
π¬ 0
π 0
How do you knock the induction heads out of an LM while preserving its ability to think? Is it even possible?
@keremsahin22.bsky.social's work is worth reading if you haven't seen it yet.
hapax.baulab.info
21.02.2026 21:31 β
π 26
π 5
π¬ 1
π 1
Memorization vs. generalization in deep learning: implicit biases, benign overfitting, and more
Or: how I learned to stop worrying and love the memorization
What is the relationship between memorization and generalization in AI? Is there a fundamental tradeoff? In infinitefaculty.substack.com/p/memorizati... Iβve reviewed some of the evolving perspectives on memorization & generalization in machine learning, from classic perspectives through LLMs.
18.02.2026 15:54 β
π 132
π 27
π¬ 4
π 5
Unfortunately the event is in-person only.
16.02.2026 16:43 β
π 2
π 0
π¬ 0
π 0
Mechanistic Basis of Reasoning (in Brains and AI) | IVADO
Very excited for our second workshop on the computational ingredients of reasoning (Feb 24-27), this one focused on mechanisms of reasoning in both AI and the brain. Check out the program to see our amazing lineup of speakers, and please consider attending! ivado.ca/en/events/me...
16.02.2026 16:29 β
π 9
π 2
π¬ 1
π 0
Building compositional tasks with shared neural subspaces
Nature - The brain can flexibly perform multiple tasks by compositionally combining task-relevant neural representations.
Thrilled that my paper is out in the @nature.com. We explored how the brain builds complex tasks by compositionally combining simpler sub-task representations. The brain flexibly performs multiple tasks by dynamically reusing neural subspaces for sensory inputs and motor actions
rdcu.be/eRVUk
11.02.2026 22:40 β
π 130
π 47
π¬ 4
π 1
Excited to announce a new book telling the story of mathematical approaches to studying the mind, from the origins of cognitive science to modern AI! The Laws of Thought will be published in February and is available for pre-order now.
18.12.2025 15:59 β
π 166
π 39
π¬ 2
π 5
That is, in order to do the kinds of things that are supposed to require algebraic / rule-based operations, these models actually do something that is algebraic, which both affirms the importance of algebraic operations for human-like reasoning and also shows it doesn't need to be innate.
08.02.2026 02:11 β
π 0
π 0
π¬ 1
π 0
Can you elaborate?
08.02.2026 01:45 β
π 0
π 0
π¬ 1
π 0
Very cool! Yeah would love to chat at some point.
08.02.2026 00:24 β
π 2
π 0
π¬ 0
π 0
Yes!
08.02.2026 00:12 β
π 0
π 0
π¬ 1
π 0
I'm curious, would you say that this sort of thing is consistent with the predictions of the LoT hypothesis, i.e. these models may emergently be implementing a LoT?
07.02.2026 23:13 β
π 4
π 0
π¬ 1
π 0
amazing summary of our work on visual symbolic mechanisms bsky.app/profile/this...
07.02.2026 23:05 β
π 4
π 0
π¬ 1
π 0
06.02.2026 21:07 β
π 10
π 3
π¬ 0
π 2
Thanks Shahab!
07.02.2026 14:28 β
π 1
π 0
π¬ 0
π 0
Excited to share that our work on Visual Symbolic Mechanisms has been accepted to ICLR! π΄π§π·
06.02.2026 18:46 β
π 11
π 2
π¬ 0
π 0
YouTube video by Johns Hopkins University
Apes Share Human Ability to Imagine
Imagination in bonobos!
I am thrilled to share a new paper w/ Amalia Bastos, out now in @science.org
We provide the first experimental evidence that a nonhuman animal can follow along a pretend scenario & track imaginary objects. Work w/ Kanzi, the bonobo, at Ape Initiative
youtu.be/NUSHcQQz2Ko
05.02.2026 19:18 β
π 289
π 110
π¬ 10
π 10
Very excited about this work looking at the emergent mechanisms that vision language models use to perform structured visual processing, mirroring a computational strategy (visual indexing) proposed in cognitive science, but here learned by VLMs. Check out the paper/thread for more details!
05.02.2026 21:24 β
π 25
π 2
π¬ 1
π 0
The visual world is composed of objects, and those objects are composed of features. But do VLMs exploit this compositional structure when processing multi-object scenes? In our ππ #ICLR2026 paper, we find they do β via emergent symbolic mechanisms for visual binding. π§΅π
05.02.2026 20:54 β
π 83
π 25
π¬ 1
π 3
This was a blast! Thanks for joining us!
31.01.2026 23:49 β
π 9
π 0
π¬ 0
π 0
Federal agents with weapons drawn, moments before murdering American citizens on the streets of Minneapolis at the dawn of 2026.
What should academics be doing right now?
I have been writing up some thoughts on what the research says about effective action, and what universities specifically can do.
davidbau.github.io/poetsandnurs...
It's on GitHub. Suggestions and pull requests welcome.
github.com/davidbau/poe...
26.01.2026 03:27 β
π 37
π 16
π¬ 0
π 4
Can you solve this algebra puzzle? π§©
cb=c, ac=b, ab=?
A small transformer can learn to solve problems like this!
And since the letters don't have inherent meaning, this lets us study how context alone imparts meaning. Here's what we found:π§΅β¬οΈ
22.01.2026 16:09 β
π 48
π 10
π¬ 2
π 2