Ethan Gotlieb Wilcox's Avatar

Ethan Gotlieb Wilcox

@wegotlieb.bsky.social

Assistant Professor of Computational Linguistics @ Georgetown; formerly postdoc @ ETH Zurich; PhD @ Harvard Linguistics, affiliated with MIT Brain & Cog Sci. Language, Computers, Cognition.

916 Followers  |  194 Following  |  24 Posts  |  Joined: 13.11.2023  |  2.1952

Latest posts by wegotlieb.bsky.social on Bluesky

Post image

Virtual information session for Georgetown’s 2-year Master’s in Computational Linguistics! Learn about our courses in NLP, psycholinguistics, low-resource languages, digital humanities, and LLMs, plus phonology, syntax, & semantics. DM for registration link. Friday Nov. 21 | 10–11 AM #linguistics

14.11.2025 17:52 β€” πŸ‘ 3    πŸ” 2    πŸ’¬ 1    πŸ“Œ 0
Screenshot of a figure with two panels, labeled (a) and (b). The caption reads: "Figure 1: (a) Illustration of messages (left) and strings (right) in toy domain. Blue = grammatical strings. Red = ungrammatical strings. (b) Surprisal (negative log probability) assigned to toy strings by GPT-2."

Screenshot of a figure with two panels, labeled (a) and (b). The caption reads: "Figure 1: (a) Illustration of messages (left) and strings (right) in toy domain. Blue = grammatical strings. Red = ungrammatical strings. (b) Surprisal (negative log probability) assigned to toy strings by GPT-2."

New work to appear @ TACL!

Language models (LMs) are remarkably good at generating novel well-formed sentences, leading to claims that they have mastered grammar.

Yet they often assign higher probability to ungrammatical strings than to grammatical strings.

How can both things be true? πŸ§΅πŸ‘‡

10.11.2025 22:11 β€” πŸ‘ 84    πŸ” 19    πŸ’¬ 2    πŸ“Œ 3

I did not! Yikes! Another reason to include "pickle" and/or pickle-related emoji in any lab communication!

23.10.2025 01:04 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
GUCL: Computation and Language @ Georgetown

Georgetown Linguistics has a dedicated Computational Linguistics PhD track, and a lively CL community on campus (gucl.georgetown.edu), including my faculty colleagues @complingy.bsky.social and Amir Zeldes.

21.10.2025 21:52 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

PICoL stands for β€œPsycholinguistics, Information, and Computational Linguistics,” and I encourage applications from anyone whose research interests connect with these topics!

21.10.2025 21:52 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Post image

I will be recruiting PhD students via Georgetown Linguistics this application cycle! Come join us in the PICoL (pronounced β€œpickle”) lab. We focus on psycholinguistics and cognitive modeling using LLMs. See the linked flyer for more details: bit.ly/3L3vcyA

21.10.2025 21:52 β€” πŸ‘ 28    πŸ” 14    πŸ’¬ 2    πŸ“Œ 0

🌟🌟This paper will appear at ACL 2025 (@aclmeeting.bsky.social)! New updated version is on arXiv: arxiv.org/pdf/2505.07659 🌟🌟

03.06.2025 13:45 β€” πŸ‘ 9    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Post image

A key hypothesis in the history of linguistics is that different constructions share underlying structure. We take advantage of recent advances in mechanistic interpretability to test this hypothesis in Language Models.

New work with @kmahowald.bsky.social and @cgpotts.bsky.social!

πŸ§΅πŸ‘‡!

27.05.2025 14:32 β€” πŸ‘ 30    πŸ” 6    πŸ’¬ 1    πŸ“Œ 3
Preview
Measuring Morphological Fusion Using Partial Information Decomposition Michaela Socolof, Jacob Louis Hoover, Richard Futrell, Alessandro Sordoni, Timothy J. O’Donnell. Proceedings of the 29th International Conference on Computational Linguistics. 2022.

We see this project as in line with some other recent papers seeking to cast typological variation in information-theoretic terms, with shout-outs to Michaela Socolof, @postylem.bsky.social @futrell.bsky.social (aclanthology.org/2022.coling-...) and Julius Steuer (aclanthology.org/2023.sigtyp-...)

13.05.2025 13:21 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Preview
Quantifying the redundancy between prosody and text Lukas Wolf, Tiago Pimentel, Evelina Fedorenko, Ryan Cotterell, Alex Warstadt, Ethan Wilcox, Tamar Regev. Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing. 2023.

⭐ ⭐This paper also makes several technical contributions to the mixed-pair mutual information estimation pipeline of Wolf et al., (aclanthology.org/2023.emnlp-m...). Shout out to @cuiding.bsky.social for all her hard work on this aspect of the paper! ⭐⭐

13.05.2025 13:21 β€” πŸ‘ 2    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Post image

βœ…In line with our prediction, we find that mutual information is higher in tonal languages than in non-tonal languages. BUT, the way one represents context is important. When full sentential context is taken into account (mBERT and mGPT), the distinction collapses.

13.05.2025 13:21 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

🌏🌍We test this prediction by estimating mutual information in an audio dataset of 10 different languages across 6 language families. 🌏🌍

13.05.2025 13:21 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

We propose a way to do so using β€¦πŸ“‘information theory.πŸ“‘ In tonal languages, pitch reduces uncertainty about lexical identity, therefore, the mutual information between pitch and words should be higher.

13.05.2025 13:21 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

🌐But there are intermediate languages, which have lexically contrastive tone, but only sporadically, making some linguists doubt the tonal/non-tonal dichotomy. So, how can we measure how β€œtonal” a language is? 🧐🧐

13.05.2025 13:21 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

🌏 Different languages use pitch in different ways. 🌏 β€œTonal” languages, like Cantonese, use it to make lexical distinctions. πŸ“– While others, like English, use it for other functions, like marking whether or not a sentence is a question. ❓

13.05.2025 13:21 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Preview
Using Information Theory to Characterize Prosodic Typology: The Case of Tone, Pitch-Accent and Stress-Accent This paper argues that the relationship between lexical identity and prosody -- one well-studied parameter of linguistic variation -- can be characterized using information theory. We predict that lan...

β­πŸ—£οΈNew preprint out: πŸ—£οΈβ­ β€œUsing Information Theory to Characterize Prosodic Typology: The Case of Tone, Pitch-Accent and Stress-Accent” with @cuiding.bsky.social , Giovanni Acampa, @tpimentel.bsky.social , @alexwarstadt.bsky.social ,Tamar Regev: arxiv.org/abs/2505.07659

13.05.2025 13:21 β€” πŸ‘ 12    πŸ” 5    πŸ’¬ 1    πŸ“Œ 2
Preview
GitHub - babylm/evaluation-pipeline-2025 Contribute to babylm/evaluation-pipeline-2025 development by creating an account on GitHub.

I’ll also use this as a way to plug human-scale language modeling in the wild: This year’s BabyLM eval pipeline was just released last week at github.com/babylm/evalu.... For more info on BabyLM head to babylm.github.io

12.05.2025 15:48 β€” πŸ‘ 3    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

Couldn’t be happier to have co-authored this will a stellar team, including: Michael Hu, @amuuueller.bsky.social, @alexwarstadt.bsky.social, @lchoshen.bsky.social, Chengxu Zhuang, @adinawilliams.bsky.social, Ryan Cotterell, @tallinzen.bsky.social

12.05.2025 15:48 β€” πŸ‘ 3    πŸ” 1    πŸ’¬ 1    πŸ“Œ 0

This version includes 😱New analyses 😱new arguments 😱 and a whole new β€œLooking Forward” section! If you’re interested in what a team of (psycho) computational linguists thinks the future will hold, check out our brand new Section 8!

12.05.2025 15:48 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
OSF

πŸ“£Paper Update πŸ“£It’s bigger! It’s better! Even if the language models aren’t. πŸ€–New version of β€œBigger is not always Better: The importance of human-scale language modeling for psycholinguistics” osf.io/preprints/ps...

12.05.2025 15:48 β€” πŸ‘ 18    πŸ” 3    πŸ’¬ 1    πŸ“Œ 2
OSF

Excited to share our preprint "Using MoTR to probe agreement errors in Russian"! w/ Metehan Oğuz, @wegotlieb.bsky.social, Zuzanna Fuchs Link: osf.io/preprints/ps...
1- We provide moderate evidence that processing of agreement errors is modulated by agreement type (internal vs external agr.)

07.03.2025 22:21 β€” πŸ‘ 3    πŸ” 1    πŸ’¬ 1    πŸ“Œ 0
Preview
Looking forward: Linguistic theory and methods This chapter examines current developments in linguistic theory and methods, focusing on the increasing integration of computational, cognitive, and evolutionary perspectives. We highlight four major ...

Me and @wegotlieb.bsky.social were recently invited to write a wide-ranging reflection on the current state of linguistic theory and methodology.
A draft is up here. For anyone interested in thinking big about linguistics, we'd be happy to hear your thoughts!
arxiv.org/abs/2502.18313
#linguistics

27.02.2025 14:47 β€” πŸ‘ 14    πŸ” 2    πŸ’¬ 0    πŸ“Œ 0

βš–οΈπŸ“£This paper was a big departure from my typical cognitive science fare, and so much fun to write! πŸ“£βš–οΈ Thank you to @bwal.bsky.social and especially to @kevintobia.bsky.social for their legal expertise on this project!

19.02.2025 14:25 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

On the positive side, we suggest that LLMs can serve a role as β€œdialectic” partners πŸ—£οΈβ”πŸ—£οΈ helping judges and clerks strengthen their arguments, as long as judicial sovereignty is maintained πŸ‘©β€βš–οΈπŸ‘‘πŸ‘©β€βš–οΈ

19.02.2025 14:25 β€” πŸ‘ 2    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

βš–οΈ We also show, through demonstration, that it’s very easy to engineer prompts that steer models toward one’s desired interpretation of a word or phrase. πŸ“–Prompting is the new β€œdictionary shopping” 😬 πŸ“– 😬

19.02.2025 14:25 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

πŸ›οΈWe identify five β€œmyths” about LLMs which, when dispelled, reveal their limitations as legal tools for textual interpretation. To take one example, during instruction tuning, LLMs are trained on highly structured, non-natural inputs.

19.02.2025 14:25 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

We argue no! πŸ™…β€β™‚οΈ While LLMs appear to possess excellent language capabilities, they should not be used as references for β€œordinary language use,” at least in the legal setting. βš–οΈ The reasons are manifold.

19.02.2025 14:25 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

πŸ›οΈLast year a U.S. judge queried Chat GPT to help with their interpretation of β€œordinary meaning,” in the same way one might use a dictionary to look up the ordinary definition of a word πŸ“– … But is it the same?

19.02.2025 14:25 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Large Language Models for Legal Interpretation? Don't Take Their Word for It <p><span>Recent breakthroughs in statistical language modeling have impacted countless domains, including the law. Chatbot applications such as ChatGPT, Claude,

πŸ“£ New Paper βš–οΈπŸ§‘β€βš–οΈπŸ›οΈ Large Language Models for Legal Interpretation? Don't Take Their Word for It πŸ‘©β€βš–οΈπŸ›οΈβš–οΈ with @bwal.bsky.social , @complingy.bsky.social Amir Zeldes, and @kevintobia.bsky.social papers.ssrn.com/sol3/papers....

19.02.2025 14:25 β€” πŸ‘ 13    πŸ” 3    πŸ’¬ 1    πŸ“Œ 0

@wegotlieb is following 20 prominent accounts