Ben Prystawski's Avatar

Ben Prystawski

@benpry.bsky.social

Cognitive science PhD student at Stanford, studying iterated learning and reasoning.

830 Followers  |  361 Following  |  5 Posts  |  Joined: 15.09.2023  |  1.9025

Latest posts by benpry.bsky.social on Bluesky

Now out in Cognition, work with the great @gershbrain.bsky.social @tobigerstenberg.bsky.social on formalizing self-handicapping as rational signaling!
πŸ“ƒ authors.elsevier.com/a/1lo8f2Hx2-...

19.09.2025 03:46 β€” πŸ‘ 33    πŸ” 13    πŸ’¬ 1    πŸ“Œ 1
Post image

How do we predict what others will do next? πŸ€”
We look for patterns. But what are the limits of this ability?
In our new paper at CCN 2025 (@cogcompneuro.bsky.social), we explore the computational constraints of human pattern recognition using the classic game of Rock, Paper, Scissors πŸ—ΏπŸ“„βœ‚οΈ

12.08.2025 22:55 β€” πŸ‘ 13    πŸ” 3    πŸ’¬ 1    πŸ“Œ 0
APA PsycNet

My final project from grad school is out now in Dev Psych! Mombasa County preschoolers were more accurate on object-based than picture-based vocabulary assessments, whereas Bay Area preschoolers were equally accurate on object-based and picture-based assessments.

psycnet.apa.org/doiLanding?d...

06.08.2025 23:54 β€” πŸ‘ 10    πŸ” 2    πŸ’¬ 1    πŸ“Œ 0
What do representations tell us about a system? Image of a mouse with a scope showing a vector of activity patterns, and a neural network with a vector of unit activity patterns
Common analyses of neural representations: Encoding models (relating activity to task features) drawing of an arrow from a trace saying [on_____on____] to a neuron and spike train. Comparing models via neural predictivity: comparing two neural networks by their R^2 to mouse brain activity. RSA: assessing brain-brain or model-brain correspondence using representational dissimilarity matrices

What do representations tell us about a system? Image of a mouse with a scope showing a vector of activity patterns, and a neural network with a vector of unit activity patterns Common analyses of neural representations: Encoding models (relating activity to task features) drawing of an arrow from a trace saying [on_____on____] to a neuron and spike train. Comparing models via neural predictivity: comparing two neural networks by their R^2 to mouse brain activity. RSA: assessing brain-brain or model-brain correspondence using representational dissimilarity matrices

In neuroscience, we often try to understand systems by analyzing their representations β€” using tools like regression or RSA. But are these analyses biased towards discovering a subset of what a system represents? If you're interested in this question, check out our new commentary! Thread:

05.08.2025 14:36 β€” πŸ‘ 163    πŸ” 53    πŸ’¬ 5    πŸ“Œ 0
Preview
Idiosyncratic but not opaque: Linguistic conventions formed in reference games are interpretable by naïve humans and visionΓ’β‚¬β€œlanguage models Author(s): Boyce, Veronica; Prystawski, Ben; Tan, Alvin Wei Ming; Frank, Michael C. | Abstract: When are in-group linguistic conventions opaque to non-group members (teen slang like "rizz") or general...

When people form conventions in reference games, how easy are they for outsiders to interpret? (for values of "outsider" that include naΓ―ve humans and vision-language models) Check out @vboyce.bsky.social's poster today at #CogSci2025 to find out.
paper: escholarship.org/uc/item/16c4...

01.08.2025 16:00 β€” πŸ‘ 5    πŸ” 3    πŸ’¬ 0    πŸ“Œ 0
Preview
Scaling up the think-aloud method The think-aloud method, where participants voice their thoughts as they solve a task, is a valuable source of rich data about human reasoning processes. Yet, it has declined in popularity in contempor...

How can we use modern NLP methods to get lots of granular data from think-aloud experiments? Watch @danielwurgaft.bsky.social explain how in the Reasoning session at 4pm this afternoon at #CogSci2025
paper: arxiv.org/abs/2505.23931

01.08.2025 15:57 β€” πŸ‘ 2    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Preview
Thinking fast, slow, and everywhere in between in humans and language models Author(s): Prystawski, Ben; Goodman, Noah | Abstract: How do humans adapt how they reason to varying circumstances? Prior research has argued that reasoning comes in two types: a fast, intuitive type ...

How do people trade off between speed and accuracy in reasoning tasks without easy heuristics? Come to my talk, "Thinking fast, slow, and everywhere in between in humans and language models," in the Reasoning session this afternoon #CogSci2025 to find out!
paper: escholarship.org/uc/item/5td9...

01.08.2025 15:49 β€” πŸ‘ 4    πŸ” 1    πŸ’¬ 0    πŸ“Œ 0
Video thumbnail

🚨New paper! We know models learn distinct in-context learning strategies, but *why*? Why generalize instead of memorize to lower loss? And why is generalization transient?

Our work explains this & *predicts Transformer behavior throughout training* without its weights! 🧡

1/

28.06.2025 02:35 β€” πŸ‘ 47    πŸ” 7    πŸ’¬ 2    πŸ“Œ 2

How can we combine the process-level insight that think-aloud studies give us with the large scale that modern online experiments permit? In our new CogSci paper, we show that speech-to-text models and LLMs enable us to scale up the think-aloud method to large experiments!

25.06.2025 05:32 β€” πŸ‘ 22    πŸ” 5    πŸ’¬ 0    πŸ“Œ 0

Delighted to announce our CogSci '25 workshop at the interface between cognitive science and design πŸ§ πŸ–ŒοΈ!

We're calling it: 🏺Minds in the Making🏺
πŸ”— minds-making.github.io

June – July 2024, free & open to the public
(all career stages, all disciplines)

06.06.2025 00:30 β€” πŸ‘ 57    πŸ” 21    πŸ’¬ 2    πŸ“Œ 2
figure 2 from our preprint, reporting the results from two experiments 

we measure moral judgments about dividing money between two parties and manipulate the degree of asymmetry in the outside options each party has

we find that moral judgments track predictions from rational bargaining models like the nash bargaining solution and the kalai-smorodinsky solution in a negotiation context

by contrast, in a donation context, moral intuitions completely reverse, instead tracking redistributive and egalitarian principles

preprint link: https://osf.io/preprints/psyarxiv/3uqks_v1

figure 2 from our preprint, reporting the results from two experiments we measure moral judgments about dividing money between two parties and manipulate the degree of asymmetry in the outside options each party has we find that moral judgments track predictions from rational bargaining models like the nash bargaining solution and the kalai-smorodinsky solution in a negotiation context by contrast, in a donation context, moral intuitions completely reverse, instead tracking redistributive and egalitarian principles preprint link: https://osf.io/preprints/psyarxiv/3uqks_v1

the functional form of moral judgment is (sometimes) the nash bargaining solution

new preprintπŸ‘‡

20.05.2025 15:08 β€” πŸ‘ 23    πŸ” 7    πŸ’¬ 1    πŸ“Œ 2

Despite the world being on fire, I can't help but be thrilled to announce that I'll be starting as an Assistant Professor in the Cognitive Science Program at Dartmouth in Fall '26. I'll be recruiting grad students this upcoming cycleβ€”get in touch if you're interested!

07.05.2025 22:08 β€” πŸ‘ 142    πŸ” 24    πŸ’¬ 17    πŸ“Œ 4
title of paper (in text) plus author list

title of paper (in text) plus author list

Time course of word recognition for kids at different ages.

Time course of word recognition for kids at different ages.

Super excited to submit a big sabbatical project this year: "Continuous developmental changes in word
recognition support language learning across early
childhood": osf.io/preprints/ps...

14.04.2025 21:58 β€” πŸ‘ 68    πŸ” 27    πŸ’¬ 1    πŸ“Œ 1
OSF

Hello bluesky world :) excited to share a new paper on data visualization literacy πŸ“ˆ 🧠 w/ @judithfan.bsky.social, @arnavverma.bsky.social, Holly Huey, Hannah Lloyd, @lacepadilla.bsky.social!

πŸ“ preprint: osf.io/preprints/ps...
πŸ’» code: github.com/cogtoolslab/...

07.03.2025 17:05 β€” πŸ‘ 26    πŸ” 7    πŸ’¬ 4    πŸ“Œ 1
Figure 1. A schematic depiction of a model-mechanism mapping between a human learning system (left side) and a cognitive model (right side). Candidate model mechanism mappings are pictured as mapping between representations but also can be in terms of input data, architecture, or learning objective.

Figure 1. A schematic depiction of a model-mechanism mapping between a human learning system (left side) and a cognitive model (right side). Candidate model mechanism mappings are pictured as mapping between representations but also can be in terms of input data, architecture, or learning objective.

Figure 2. Data efficiency in human learning. (left) Order of magnitude of LLM vs. human training data, plotted by human age. Ranges are approximated from Frank (2023a). (right) A schematic depiction of evaluation scaling curves for human learners vs. models plotted by training data
quantity.

Figure 2. Data efficiency in human learning. (left) Order of magnitude of LLM vs. human training data, plotted by human age. Ranges are approximated from Frank (2023a). (right) A schematic depiction of evaluation scaling curves for human learners vs. models plotted by training data quantity.

Paper abstract

Paper abstract

AI models are fascinating, impressive, and sometimes problematic. But what can they tell us about the human mind?

In a new review paper, @noahdgoodman.bsky.social and I discuss how modern AI can be used for cognitive modeling: osf.io/preprints/ps...

06.03.2025 17:39 β€” πŸ‘ 63    πŸ” 25    πŸ’¬ 2    πŸ“Œ 0
Post image

1/13 New Paper!! We try to understand why some LMs self-improve their reasoning while others hit a wall. The key? Cognitive behaviors! Read our paper on how the right cognitive behaviors can make all the difference in a model's ability to improve with RL! 🧡

04.03.2025 18:15 β€” πŸ‘ 56    πŸ” 17    πŸ’¬ 2    πŸ“Œ 3
Post image

New paper in Psychological Review!

In "Causation, Meaning, and Communication" Ari Beller (cicl.stanford.edu/member/ari_b...) develops a computational model of how people use & understand expressions like "caused", "enabled", and "affected".

πŸ“ƒ osf.io/preprints/ps...
πŸ“Ž github.com/cicl-stanfor...
🧡

12.02.2025 18:25 β€” πŸ‘ 57    πŸ” 17    πŸ’¬ 1    πŸ“Œ 0
Preview
The broader spectrum of in-context learning The ability of language models to learn a task from a few examples in context has generated substantial interest. Here, we provide a perspective that situates this type of supervised few-shot learning...

What counts as in-context learning (ICL)? Typically, you might think of it as learning a task from a few examples. However, we’ve just written a perspective (arxiv.org/abs/2412.03782) suggesting interpreting a much broader spectrum of behaviors as ICL! Quick summary thread: 1/7

10.12.2024 18:17 β€” πŸ‘ 123    πŸ” 31    πŸ’¬ 2    πŸ“Œ 1

Hey! Could you add me?

23.11.2024 23:16 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Aerial picture of the UBC campus, with an arrow pointing to a building and text asking "Your PhD lab?"

Aerial picture of the UBC campus, with an arrow pointing to a building and text asking "Your PhD lab?"

Do you want to understand how language models work, and how they can change language science? I'm recruiting PhD students at UBC Linguistics! The research will be fun, and Vancouver is lovely. So much cool NLP happening at UBC across both Ling and CS! linguistics.ubc.ca/graduate/adm...

18.11.2024 19:43 β€” πŸ‘ 23    πŸ” 8    πŸ’¬ 1    πŸ“Œ 2
OSF

If you try to replicate a finding so you can build on it, but your study fails, what should you do? Should you follow up and try to "rescue" the failed rep, or should you move on? Boyce et al. tried to answer this question; in our sample, 5 of 17 rescue projects succeeded.

osf.io/preprints/ps...

18.10.2024 15:51 β€” πŸ‘ 36    πŸ” 20    πŸ’¬ 1    πŸ“Œ 3
A promotional image of One Hour One Life, showing a character growing up from a baby, to a child, to an adult, to an old woman, to a pile of bones. This work is not affiliated with One Hour One Life; we are grateful to Jason Rohrer, the game's developer, for making the game open data and open source.

A promotional image of One Hour One Life, showing a character growing up from a baby, to a child, to an adult, to an old woman, to a pile of bones. This work is not affiliated with One Hour One Life; we are grateful to Jason Rohrer, the game's developer, for making the game open data and open source.

Preprint alert! After 4 years, I’m super excited to share work with @thecharleywu.bsky.social @gershbrain.bsky.social and Eric Schulz on the rise and fall of technological development in virtual communities in #OneHourOneLife #ohol
doi.org/10.31234/osf...

13.09.2024 19:29 β€” πŸ‘ 39    πŸ” 11    πŸ’¬ 1    πŸ“Œ 1
Clear clusters in model representations driven by some features (plot colors) but neglecting other more complex ones (plotted as shapes) which are mixed within the color clusters.

Clear clusters in model representations driven by some features (plot colors) but neglecting other more complex ones (plotted as shapes) which are mixed within the color clusters.

How well can we understand an LLM by interpreting its representations? What can we learn by comparing brain and model representations? Our new paper highlights intriguing biases in learned feature representations that make interpreting them more challenging! 1/

23.05.2024 18:58 β€” πŸ‘ 19    πŸ” 10    πŸ’¬ 2    πŸ“Œ 3
Post image Post image

When a replication fails, researchers have to decide whether to make another attempt or move on. How should we think about this decision? Here's a new paper trying to answer this question, led by Veronica Boyce and featuring student authors from my class!

osf.io/preprints/ps...

06.05.2024 19:23 β€” πŸ‘ 15    πŸ” 10    πŸ’¬ 0    πŸ“Œ 0

@benpry is following 20 prominent accounts