Anzi Wang's Avatar

Anzi Wang

@anziw.bsky.social

linguistics phd student

68 Followers  |  85 Following  |  10 Posts  |  Joined: 06.11.2024  |  2.0417

Latest posts by anziw.bsky.social on Bluesky


Preview
TunePad TunePad is a free online platform for creating music with the Python programming language. Our step-by-step tutorials are perfect for beginners, and our advanced production tools power music making fo...

TunePad (tunepad.com)! Developed by amazing researchers at Northwestern, extremely friendly and intuitive UI. I'm teaching fifth-graders at local schools TunePad and most of them are loving it

22.02.2026 03:23 β€” πŸ‘ 2    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

one year ago i thought p-side was about predicates and s-side was about subjects :3

18.01.2026 20:05 β€” πŸ‘ 4    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

Wow!! Congratulations!!

07.01.2026 23:28 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

Reposting because the link has expired:

PDF: drive.google.com/file/d/1t2EF... (if this doesn't work, lmk)

Publisher link: www.sciencedirect.com/science/arti...

01.12.2025 19:03 β€” πŸ‘ 10    πŸ” 2    πŸ’¬ 0    πŸ“Œ 1

so cool!!

14.11.2025 21:18 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Video thumbnail

New Preprint: osf.io/eq2ra

Reading feels effortless, but it's actually quite complex under the hood. Most words are easy to process, but some words make us reread or linger. It turns out that LLMs can tell us about why, but only in certain cases... (1/n)

14.11.2025 19:18 β€” πŸ‘ 12    πŸ” 5    πŸ’¬ 2    πŸ“Œ 1
Screenshot of a figure with two panels, labeled (a) and (b). The caption reads: "Figure 1: (a) Illustration of messages (left) and strings (right) in toy domain. Blue = grammatical strings. Red = ungrammatical strings. (b) Surprisal (negative log probability) assigned to toy strings by GPT-2."

Screenshot of a figure with two panels, labeled (a) and (b). The caption reads: "Figure 1: (a) Illustration of messages (left) and strings (right) in toy domain. Blue = grammatical strings. Red = ungrammatical strings. (b) Surprisal (negative log probability) assigned to toy strings by GPT-2."

New work to appear @ TACL!

Language models (LMs) are remarkably good at generating novel well-formed sentences, leading to claims that they have mastered grammar.

Yet they often assign higher probability to ungrammatical strings than to grammatical strings.

How can both things be true? πŸ§΅πŸ‘‡

10.11.2025 22:11 β€” πŸ‘ 91    πŸ” 20    πŸ’¬ 2    πŸ“Œ 3
Preview
Training an NLP Scholar at a Small Liberal Arts College: A Backwards Designed Course Proposal The rapid growth in natural language processing (NLP) over the last couple years has generated student interest and excitement in learning more about the field. In this paper, we present two types of ...

I took Grusha and Forrest's version of NLP at Colgate (arxiv.org/abs/2408.05664) and as a current linguistics phd student still doing NLP research, I can say that this is THE undergrad course that has benefited me the most

10.11.2025 03:12 β€” πŸ‘ 2    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Preview
To model human linguistic prediction, make LLMs less superhuman When people listen to or read a sentence, they actively make predictions about upcoming words: words that are less predictable are generally read more slowly than predictable ones. The success of larg...

arxiv.org/abs/2510.05141

13.10.2025 23:59 β€” πŸ‘ 13    πŸ” 4    πŸ’¬ 0    πŸ“Œ 0
Preview
GitHub - aaronstevenwhite/glazing: Unified data models and interfaces for syntactic and semantic frame ontologies. Unified data models and interfaces for syntactic and semantic frame ontologies. - aaronstevenwhite/glazing

I've found it kind of a pain to work with resources like VerbNet, FrameNet, PropBank (frame files), and WordNet using existing tools. Maybe you have too. Here's a little package that handles data management, loading, and cross-referencing via either a CLI or a python API.

27.09.2025 13:51 β€” πŸ‘ 27    πŸ” 7    πŸ’¬ 3    πŸ“Œ 1

Brand new version of this paper (now a short book!) available at lingbuzz.net/lingbuzz/008...!

26.09.2025 18:12 β€” πŸ‘ 7    πŸ” 2    πŸ’¬ 1    πŸ“Œ 2

favorite garden path sentence of the year: "It's better to be hurt by someone you know accidentally, than by a stranger on purpose" by Dwight Schrute

23.09.2025 17:14 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Preview
Language Models Identify Ambiguities and Exploit Loopholes Studying the responses of large language models (LLMs) to loopholes presents a two-fold opportunity. First, it affords us a lens through which to examine ambiguity and pragmatics in LLMs, since exploi...

arxiv.org/abs/2508.19546

02.09.2025 09:29 β€” πŸ‘ 15    πŸ” 6    πŸ’¬ 0    πŸ“Œ 0

probably? maybe you can have different policies each semester and test it out lol

02.09.2025 00:14 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

might depend more on participation policy

01.09.2025 19:13 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

A paper with Vic Ferreira and Norvin Richards is now out

(1) Speakers syntactically encode zero complementizers as cognitively active mental object.

(2) No evidence LLMs capture cross constructional generalizations about null complementizers.

nam10.safelinks.protection.outlook.com?url=https%3A...

03.08.2025 22:21 β€” πŸ‘ 15    πŸ” 7    πŸ’¬ 1    πŸ“Œ 1
Preview
NKC Resource Library - The National Kitten Coalition’s Guide to Help You Save More Kittensβ„’ National Kitten Coalition

Work with kittens? Check out the National Kitten Coalition's new Kitten Resource Library! They're an org I like a lot!! library.kittencoalition.org

25.07.2025 23:50 β€” πŸ‘ 214    πŸ” 72    πŸ’¬ 2    πŸ“Œ 1
Preview
Collaborative Rational Speech Act: Pragmatic Reasoning for Multi-Turn Dialog As AI systems take on collaborative roles, they must reason about shared goals and beliefs-not just generate fluent language. The Rational Speech Act (RSA) framework offers a principled approach to pr...

"We introduce Collaborative Rational Speech Act (CRSA), an information-theoretic (IT) extension of RSA that models multi-turn dialog by optimizing a gain function adapted from rate-distortion theory."

arxiv.org/abs/2507.14063

21.07.2025 13:29 β€” πŸ‘ 13    πŸ” 5    πŸ’¬ 0    πŸ“Œ 0

Someone asked me today how to get better at scientific writing. I'm not the best person to ask because I find my own writing very inadequate! But the tips I thought of were:

1. Practice, and practice with co-authors who are better writers than you. Observe how they make edits and copy them.

(1/n)

04.07.2025 10:46 β€” πŸ‘ 56    πŸ” 11    πŸ’¬ 1    πŸ“Œ 1

early November is also the best season for crabs!!

19.05.2025 01:58 β€” πŸ‘ 3    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

is computational psycholinguistics a poly-sci? #puns #linguistics

24.01.2025 14:54 β€” πŸ‘ 5    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

@anziw is following 20 prominent accounts