Tom McCoy's Avatar

Tom McCoy

@rtommccoy.bsky.social

Assistant professor at Yale Linguistics. Studying computational linguistics, cognitive science, and AI. He/him.

1,912 Followers  |  325 Following  |  180 Posts  |  Joined: 10.12.2023  |  2.1304

Latest posts by rtommccoy.bsky.social on Bluesky

Still waiting for Google to release Ge (the large version of Gemini)

18.11.2025 20:24 β€” πŸ‘ 7    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Apply - Interfolio {{$ctrl.$state.data.pageTitle}} - Apply - Interfolio

Come be my colleague!

The Yale Dept of Linguistics is hiring a 3-year Lecturer in Historical Linguistics. There's a great group here working on language change, and you could become part of it!

Application review begins Dec 14. For more info, see apply.interfolio.com/177395

18.11.2025 15:52 β€” πŸ‘ 7    πŸ” 1    πŸ’¬ 0    πŸ“Œ 0
Image of a Laffy Taffy wrapper. The joke asks "What do you deserve and is also a type of bagel?" And the answer is "Everything"

Image of a Laffy Taffy wrapper. The joke asks "What do you deserve and is also a type of bagel?" And the answer is "Everything"

I am partial to Laffy Taffy mainly because of this one (via www.reddit.com/r/wholesomem...)

18.11.2025 06:39 β€” πŸ‘ 3    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

I need to know what joke prompted this!!

18.11.2025 06:38 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

Ooh jackpot!! Thanks Adina!!

18.11.2025 04:18 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

Nice, thank you!

18.11.2025 00:54 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

Question motivated by a talk by Squid Tamar-Mattis!

18.11.2025 00:20 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

Do any languages have different forms of "and" for different phrase types?

English uses "and" no matter what's being joined: "bread and butter" (nouns), "divide and conquer" (verbs), "short and sweet" (adjs). Are there languages that vary the conjunction across these contexts?

18.11.2025 00:19 β€” πŸ‘ 10    πŸ” 0    πŸ’¬ 3    πŸ“Œ 0

I mean, that's basically what doing a postdoc is!

14.11.2025 17:08 β€” πŸ‘ 2    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Research

More info about my research: rtmccoy.com/research/

14.11.2025 16:40 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Top: A syntax tree for the sentence "the doctor by the lawyer saw the artist".

Bottom: A continuous vector.

Top: A syntax tree for the sentence "the doctor by the lawyer saw the artist". Bottom: A continuous vector.

πŸ€–πŸ§ I'll be considering applications for PhD students & postdocs to start at Yale in Fall 2026!

If you are interested in the intersection of linguistics, cognitive science, & AI, I encourage you to apply!

PhD link: rtmccoy.com/prospective_...
Postdoc link: rtmccoy.com/prospective_...

14.11.2025 16:40 β€” πŸ‘ 36    πŸ” 13    πŸ’¬ 2    πŸ“Œ 2

Thanks for the very clear summary of our work, Ben!!

13.11.2025 15:17 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

Who's up for a short thread demonstrating just how wrong this claim is by Mistral's CEO, using research by @rtommccoy.bsky.social (an co-authors) regarding crossword puzzles? Just me? Well whatever, let's dive into a mini-exploration of the Embers of Autoregression and why they singe...

12.11.2025 13:16 β€” πŸ‘ 49    πŸ” 16    πŸ’¬ 3    πŸ“Œ 0
Post image

I’m excited to share our Findings of EMNLP paper w/ @cocoscilab.bsky.social , @rtommccoy.bsky.social, and @rdhawkins.bsky.social !

Language models, unlike humans, require large amounts of data, which suggests the need for an inductive bias.
But what kind of inductive biases do we need?

07.11.2025 09:17 β€” πŸ‘ 7    πŸ” 5    πŸ’¬ 1    πŸ“Œ 1

An exciting postdoc opportunity for researchers in CogSci!

The deadline is coming up soon: Nov 10

02.11.2025 00:24 β€” πŸ‘ 2    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

Come be my colleague! The Wu Tsai Institute at Yale has two faculty searches ongoing:
- One in computational cognitive science (due Dec 1)
- One in neurodevelopment (rolling)

πŸ§ πŸ€–

02.11.2025 00:17 β€” πŸ‘ 14    πŸ” 7    πŸ’¬ 0    πŸ“Œ 0

Congratulations - it's very well deserved!!

23.10.2025 15:41 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

Much as "cupboard" is pronounced "cubberd", I think that "clipboard" should be pronounced "clibberd"

11.10.2025 00:24 β€” πŸ‘ 10    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

Beginning a Grand Tour of California!
- Oct 6: Colloquium at Berkeley Linguistics
- Oct 9: Workshop at Google Mountain View
- Oct 14: Talk at UC Irvine Center for Lg, Intelligence & Computation
- Oct 16: NLP / Text-as-Data talk at NYU

Say hi if you'll be around!

05.10.2025 21:51 β€” πŸ‘ 4    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Post image

Exciting talk in the linguistics department at UC Berkeley tomorrow!
@rtommccoy.bsky.social

05.10.2025 21:13 β€” πŸ‘ 10    πŸ” 1    πŸ’¬ 0    πŸ“Œ 0

Yes!! An excellent point!!

30.09.2025 15:41 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Illustration of the blog post's main argument, summarized as: "Theory of Mind as a Central Skill for Researchers: Research involves many skills.If each skill is viewed separately, each one takes a long time to learn. These skills can instead be connected via theory of mind – the ability to reason about the mental states of others. This allows you to transfer your abilities across areas, making it easier to gain new skills."

Illustration of the blog post's main argument, summarized as: "Theory of Mind as a Central Skill for Researchers: Research involves many skills.If each skill is viewed separately, each one takes a long time to learn. These skills can instead be connected via theory of mind – the ability to reason about the mental states of others. This allows you to transfer your abilities across areas, making it easier to gain new skills."

πŸ€– 🧠 NEW BLOG POST 🧠 πŸ€–

What skills do you need to be a successful researcher?

The list seems long: collaborating, writing, presenting, reviewing, etc

But I argue that many of these skills can be unified under a single overarching ability: theory of mind

rtmccoy.com/posts/theory...

30.09.2025 15:14 β€” πŸ‘ 20    πŸ” 2    πŸ’¬ 2    πŸ“Œ 2

Totally. I think one key question is whether you want to model the whole developmental process or just the end state. If just the end state, LLMs have a lot to offer; but if the whole development (which is what we ultimately should aim for!) there are many issues in how LLMs get there

01.09.2025 00:48 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

The conversation that frequently plays out is:

A: "LLMs do lots of compositional things!"
B: "But they also make lots of mistakes!"
A: "But so do humans!"

I don't find that very productive, so would love to see the field move toward more detailed/contentful comparisons.

01.09.2025 00:46 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

They're definitely not fully systematic, so currently it kinda comes down to personal opinion about how systematic is systematic enough. And one thing I would love to see is more systematic head-to-head comparisons of humans and neural networks so that we don't need to rely on intuitions.

01.09.2025 00:45 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

Yeah, I think that's a good definition! I also believe that some LLM behaviors qualify as this - they routinely generate sentences with a syntactic structure that never appeared in the training set.

01.09.2025 00:44 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

"Hello world!" sounds like a word followed by a crossword clue for that word: "Hell = Low world"

31.08.2025 22:55 β€” πŸ‘ 3    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

And although models still make lots of mistakes on compositionality, that alone also isn't enough because humans do too. So, if we want to make claims about models being human-like or not, what we really need are finer-grained characterizations of what human-like compositionality is.

31.08.2025 22:54 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

Agreed with these points broadly! But though being less β€œbad at compositionality” isn’t the same as compositional like humans, it does mean that we can no longer say "models completely fail at compositionality and are thus non human like" (because they no longer completely fail).

31.08.2025 22:53 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

I agree that garden paths & agreement attraction could be explained with fairly superficial statistics. For priming, what I had in mind was syntactic priming, which I do think requires some sort of structural abstraction.

31.08.2025 22:44 β€” πŸ‘ 2    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

@rtommccoy is following 20 prominent accounts