Much as "cupboard" is pronounced "cubberd", I think that "clipboard" should be pronounced "clibberd"
11.10.2025 00:24 β π 6 π 0 π¬ 1 π 0@rtommccoy.bsky.social
Assistant professor at Yale Linguistics. Studying computational linguistics, cognitive science, and AI. He/him.
Much as "cupboard" is pronounced "cubberd", I think that "clipboard" should be pronounced "clibberd"
11.10.2025 00:24 β π 6 π 0 π¬ 1 π 0Beginning a Grand Tour of California!
- Oct 6: Colloquium at Berkeley Linguistics
- Oct 9: Workshop at Google Mountain View
- Oct 14: Talk at UC Irvine Center for Lg, Intelligence & Computation
- Oct 16: NLP / Text-as-Data talk at NYU
Say hi if you'll be around!
Exciting talk in the linguistics department at UC Berkeley tomorrow!
@rtommccoy.bsky.social
Yes!! An excellent point!!
30.09.2025 15:41 β π 0 π 0 π¬ 0 π 0Illustration of the blog post's main argument, summarized as: "Theory of Mind as a Central Skill for Researchers: Research involves many skills.If each skill is viewed separately, each one takes a long time to learn. These skills can instead be connected via theory of mind β the ability to reason about the mental states of others. This allows you to transfer your abilities across areas, making it easier to gain new skills."
π€ π§ NEW BLOG POST π§ π€
What skills do you need to be a successful researcher?
The list seems long: collaborating, writing, presenting, reviewing, etc
But I argue that many of these skills can be unified under a single overarching ability: theory of mind
rtmccoy.com/posts/theory...
Totally. I think one key question is whether you want to model the whole developmental process or just the end state. If just the end state, LLMs have a lot to offer; but if the whole development (which is what we ultimately should aim for!) there are many issues in how LLMs get there
01.09.2025 00:48 β π 1 π 0 π¬ 1 π 0The conversation that frequently plays out is:
A: "LLMs do lots of compositional things!"
B: "But they also make lots of mistakes!"
A: "But so do humans!"
I don't find that very productive, so would love to see the field move toward more detailed/contentful comparisons.
They're definitely not fully systematic, so currently it kinda comes down to personal opinion about how systematic is systematic enough. And one thing I would love to see is more systematic head-to-head comparisons of humans and neural networks so that we don't need to rely on intuitions.
01.09.2025 00:45 β π 0 π 0 π¬ 1 π 0Yeah, I think that's a good definition! I also believe that some LLM behaviors qualify as this - they routinely generate sentences with a syntactic structure that never appeared in the training set.
01.09.2025 00:44 β π 1 π 0 π¬ 1 π 0"Hello world!" sounds like a word followed by a crossword clue for that word: "Hell = Low world"
31.08.2025 22:55 β π 3 π 0 π¬ 0 π 0And although models still make lots of mistakes on compositionality, that alone also isn't enough because humans do too. So, if we want to make claims about models being human-like or not, what we really need are finer-grained characterizations of what human-like compositionality is.
31.08.2025 22:54 β π 0 π 0 π¬ 1 π 0Agreed with these points broadly! But though being less βbad at compositionalityβ isnβt the same as compositional like humans, it does mean that we can no longer say "models completely fail at compositionality and are thus non human like" (because they no longer completely fail).
31.08.2025 22:53 β π 1 π 0 π¬ 1 π 0I agree that garden paths & agreement attraction could be explained with fairly superficial statistics. For priming, what I had in mind was syntactic priming, which I do think requires some sort of structural abstraction.
31.08.2025 22:44 β π 2 π 0 π¬ 1 π 0What would you view as evidence for true productivity?
31.08.2025 22:42 β π 1 π 0 π¬ 1 π 0Definitely true that LLM-style models can't go gather new data (they're restricted to focusing on a subset of their input), but it doesn't feel outside the spirit of ML to allow the system to seek new data which it then applies statistical learning over, if seeking is also statistically-driven
30.08.2025 21:01 β π 2 π 0 π¬ 1 π 0E.g., in ML, datapoint importance is determined by some inscrutable statistics, while in more nativist approaches it's determined by a desire to build a high-level causal model of the world?
30.08.2025 20:58 β π 1 π 0 π¬ 1 π 0It feels like a false dichotomy to me? In ML models, some training examples are more influential than others, so you could say an ML model can "decide" to ignore some data. In that sense both model types decide which data to learn from, but they differ in what criteria they use to do so.
30.08.2025 20:55 β π 1 π 0 π¬ 1 π 0Yes, this is a great point! I do think language (which is the domain I mainly study) gets around these concerns a bit: for language, human children primarily have to rely on being fed data, and that data is symbolic in nature. But I agree these properties don't hold for all cognitive domains!
30.08.2025 20:26 β π 1 π 0 π¬ 1 π 0In other words, our argument is very much based on the available evidence. New, stricter evidence could very well push the needle back toward needing symbols at the algorithmic level - and that would be exciting if so!
30.08.2025 19:57 β π 1 π 0 π¬ 1 π 0One key next step, then, is stricter diagnostics of symbolic behavior that go beyond βcan humans/models be compositionalβ into βin what specific ways are we compositionalβ, βwhat types of errors are madeβ, etc., and then comparing humans & models head-to-head
(cont.)
A broader comment: LLMs are definitely far from perfect. But there has been important progress. For a while, we could say βneural nets are so bad at compositionality that theyβre obviously different from humans.β Iβm no LLM fanboy, but I do think such sweeping arguments no longer apply
(cont.)
FWIW, the types of productivity that we look at go beyond n-grams; thereβs also novelty in syntactic tree structures, and in things like βusing a word as the subject of a sentence when the LLM has only ever seen it as the direct objectβ
30.08.2025 19:51 β π 1 π 0 π¬ 2 π 0I completely agree that the differences in training/learning between LLMs and humans are a major shortcoming of LLMs as cognitive models - probably the biggest edge that symbolic models have over neural networks. Meta-learning seems promising here, but is still in early stages.
30.08.2025 19:50 β π 2 π 0 π¬ 1 π 0Maybe no one talks about AlphaGo as a cognitive model because thereβs no history of research on βhere are the particularly informative behavioral quirks that humans show when playing Goβ, such that itβs not clear what evidence we would look for to argue a model is playing Go in a human-like way?
30.08.2025 19:49 β π 1 π 0 π¬ 1 π 0E.g., for psycholinguistics, LLMs show garden path effects, agreement attraction, & priming. I would also put compositionality, productivity, & rapid learning in this category of βparticularly informative cognitive phenomena.β
30.08.2025 19:47 β π 2 π 0 π¬ 2 π 0I completely agree that matching human performance is not all that matters - models should be human-like, not just human-level. That said, certain behaviors are viewed in CogSci as particularly illuminating about the mind, and one exciting thing about LLMs is that they display many such properties
30.08.2025 19:47 β π 1 π 0 π¬ 1 π 0Thank you for all these thoughts!! Here are some scattered responses:
30.08.2025 19:40 β π 4 π 0 π¬ 2 π 0But that personal dative type analysis might not work, since I don't think personal datives can get reflexive morphology (ygdp.yale.edu/phenomena/pe...). So causative might fit better!
24.08.2025 22:58 β π 2 π 0 π¬ 1 π 0I was viewing "myself" in "I got myself born" as something like a personal dative (like "I ate me some lunch"), so that "I got myself born" is basically still "I got born", with "myself" added for emphasis rather than for structural reasons. (So, this focuses on "myself" and "I" being coreferential)
24.08.2025 22:57 β π 0 π 0 π¬ 1 π 0Ooh good point! Yes, a causative analysis seems plausible here!
24.08.2025 22:56 β π 0 π 0 π¬ 1 π 0