Still waiting for Google to release Ge (the large version of Gemini)
18.11.2025 20:24 β π 7 π 0 π¬ 0 π 0@rtommccoy.bsky.social
Assistant professor at Yale Linguistics. Studying computational linguistics, cognitive science, and AI. He/him.
Still waiting for Google to release Ge (the large version of Gemini)
18.11.2025 20:24 β π 7 π 0 π¬ 0 π 0Come be my colleague!
The Yale Dept of Linguistics is hiring a 3-year Lecturer in Historical Linguistics. There's a great group here working on language change, and you could become part of it!
Application review begins Dec 14. For more info, see apply.interfolio.com/177395
Image of a Laffy Taffy wrapper. The joke asks "What do you deserve and is also a type of bagel?" And the answer is "Everything"
I am partial to Laffy Taffy mainly because of this one (via www.reddit.com/r/wholesomem...)
18.11.2025 06:39 β π 3 π 0 π¬ 1 π 0I need to know what joke prompted this!!
18.11.2025 06:38 β π 0 π 0 π¬ 1 π 0Ooh jackpot!! Thanks Adina!!
18.11.2025 04:18 β π 1 π 0 π¬ 0 π 0Nice, thank you!
18.11.2025 00:54 β π 1 π 0 π¬ 1 π 0Question motivated by a talk by Squid Tamar-Mattis!
18.11.2025 00:20 β π 0 π 0 π¬ 0 π 0Do any languages have different forms of "and" for different phrase types?
English uses "and" no matter what's being joined: "bread and butter" (nouns), "divide and conquer" (verbs), "short and sweet" (adjs). Are there languages that vary the conjunction across these contexts?
I mean, that's basically what doing a postdoc is!
14.11.2025 17:08 β π 2 π 0 π¬ 0 π 0More info about my research: rtmccoy.com/research/
14.11.2025 16:40 β π 0 π 0 π¬ 0 π 0Top: A syntax tree for the sentence "the doctor by the lawyer saw the artist". Bottom: A continuous vector.
π€π§ I'll be considering applications for PhD students & postdocs to start at Yale in Fall 2026!
If you are interested in the intersection of linguistics, cognitive science, & AI, I encourage you to apply!
PhD link: rtmccoy.com/prospective_...
Postdoc link: rtmccoy.com/prospective_...
Thanks for the very clear summary of our work, Ben!!
13.11.2025 15:17 β π 1 π 0 π¬ 0 π 0Who's up for a short thread demonstrating just how wrong this claim is by Mistral's CEO, using research by @rtommccoy.bsky.social (an co-authors) regarding crossword puzzles? Just me? Well whatever, let's dive into a mini-exploration of the Embers of Autoregression and why they singe...
12.11.2025 13:16 β π 49 π 16 π¬ 3 π 0Iβm excited to share our Findings of EMNLP paper w/ @cocoscilab.bsky.social , @rtommccoy.bsky.social, and @rdhawkins.bsky.social !
Language models, unlike humans, require large amounts of data, which suggests the need for an inductive bias.
But what kind of inductive biases do we need?
An exciting postdoc opportunity for researchers in CogSci!
The deadline is coming up soon: Nov 10
Come be my colleague! The Wu Tsai Institute at Yale has two faculty searches ongoing:
- One in computational cognitive science (due Dec 1)
- One in neurodevelopment (rolling)
π§ π€
Congratulations - it's very well deserved!!
23.10.2025 15:41 β π 1 π 0 π¬ 0 π 0Much as "cupboard" is pronounced "cubberd", I think that "clipboard" should be pronounced "clibberd"
11.10.2025 00:24 β π 10 π 0 π¬ 1 π 0Beginning a Grand Tour of California!
- Oct 6: Colloquium at Berkeley Linguistics
- Oct 9: Workshop at Google Mountain View
- Oct 14: Talk at UC Irvine Center for Lg, Intelligence & Computation
- Oct 16: NLP / Text-as-Data talk at NYU
Say hi if you'll be around!
Exciting talk in the linguistics department at UC Berkeley tomorrow!
@rtommccoy.bsky.social
Yes!! An excellent point!!
30.09.2025 15:41 β π 0 π 0 π¬ 0 π 0Illustration of the blog post's main argument, summarized as: "Theory of Mind as a Central Skill for Researchers: Research involves many skills.If each skill is viewed separately, each one takes a long time to learn. These skills can instead be connected via theory of mind β the ability to reason about the mental states of others. This allows you to transfer your abilities across areas, making it easier to gain new skills."
π€ π§ NEW BLOG POST π§ π€
What skills do you need to be a successful researcher?
The list seems long: collaborating, writing, presenting, reviewing, etc
But I argue that many of these skills can be unified under a single overarching ability: theory of mind
rtmccoy.com/posts/theory...
Totally. I think one key question is whether you want to model the whole developmental process or just the end state. If just the end state, LLMs have a lot to offer; but if the whole development (which is what we ultimately should aim for!) there are many issues in how LLMs get there
01.09.2025 00:48 β π 1 π 0 π¬ 1 π 0The conversation that frequently plays out is:
A: "LLMs do lots of compositional things!"
B: "But they also make lots of mistakes!"
A: "But so do humans!"
I don't find that very productive, so would love to see the field move toward more detailed/contentful comparisons.
They're definitely not fully systematic, so currently it kinda comes down to personal opinion about how systematic is systematic enough. And one thing I would love to see is more systematic head-to-head comparisons of humans and neural networks so that we don't need to rely on intuitions.
01.09.2025 00:45 β π 0 π 0 π¬ 1 π 0Yeah, I think that's a good definition! I also believe that some LLM behaviors qualify as this - they routinely generate sentences with a syntactic structure that never appeared in the training set.
01.09.2025 00:44 β π 1 π 0 π¬ 1 π 0"Hello world!" sounds like a word followed by a crossword clue for that word: "Hell = Low world"
31.08.2025 22:55 β π 3 π 0 π¬ 0 π 0And although models still make lots of mistakes on compositionality, that alone also isn't enough because humans do too. So, if we want to make claims about models being human-like or not, what we really need are finer-grained characterizations of what human-like compositionality is.
31.08.2025 22:54 β π 0 π 0 π¬ 1 π 0Agreed with these points broadly! But though being less βbad at compositionalityβ isnβt the same as compositional like humans, it does mean that we can no longer say "models completely fail at compositionality and are thus non human like" (because they no longer completely fail).
31.08.2025 22:53 β π 1 π 0 π¬ 1 π 0I agree that garden paths & agreement attraction could be explained with fairly superficial statistics. For priming, what I had in mind was syntactic priming, which I do think requires some sort of structural abstraction.
31.08.2025 22:44 β π 2 π 0 π¬ 1 π 0