If any friends are at Cog Sci, Iโll be in SF tomorrow! Let me know if youโd like to meet!
01.08.2025 03:24 โ ๐ 0 ๐ 0 ๐ฌ 0 ๐ 0@sathvik.bsky.social
computational psycholinguistics @ umd he/him
If any friends are at Cog Sci, Iโll be in SF tomorrow! Let me know if youโd like to meet!
01.08.2025 03:24 โ ๐ 0 ๐ 0 ๐ฌ 0 ๐ 0DEFINITELY not talking about predictive coding the next time i go back to the bay area
14.07.2025 01:42 โ ๐ 1 ๐ 0 ๐ฌ 0 ๐ 0The sycophantic tone of ChatGPT always sounded familiar, and then I recognized where I'd heard it before: author response letters to reviewer comments.
"You're exactly right, that's a great point!"
"Thank you so much for this insight!"
Also how it always agrees even when it contradicts itself.
When it comes to text prediction, where does one LM outperform another? If you've ever worked on LM evals, you know this question is a lot more complex than it seems. In our new #acl2025 paper, we developed a method to find fine-grained differences between LMs:
๐งต1/9
Table 1 Typology of traps, how they can be avoided, and what goes wrong if not avoided. Note that all traps in a sense constitute category errors (Ryle & Tanney, 2009) and the success-to-truth inference (Guest & Martin, 2023) is an important driver in most, if not all, of the traps.
NEW paper! ๐ญ๐ฅ๏ธ
โCombining Psychology with Artificial Intelligence: What could possibly go wrong?โ
โ Brief review paper by @olivia.science & myself, highlighting traps to avoid when combining Psych with AI, and why this is so important. Check out our proposed way forward! ๐๐ก
osf.io/preprints/ps...
A bit late but since I really like this paper, a bit of self-advertising! I am presenting at CMCL today work showing that metrics measuring how a Minimalist Grammar parser modulates memory usage can help us model Self-paced reading data for SRC/ORC contrasts: aclanthology.org/2025.cmcl-1.5/
03.05.2025 16:43 โ ๐ 28 ๐ 6 ๐ฌ 4 ๐ 0New preprint on controlled generation from LMs!
I'll be presenting at NENLP tomorrow 12:50-2:00pm
Longer thread coming soon :)
thanks a lot Kanishka! happy i made a little contribution & that the probabilities are corrected :)
08.04.2025 15:10 โ ๐ 2 ๐ 0 ๐ฌ 0 ๐ 0from minicons import scorer from nltk.tokenize import TweetTokenizer lm = scorer.IncrementalLMScorer("gpt2") # your own tokenizer function that returns a list of words # given some sentence input word_tokenizer = TweetTokenizer().tokenize # word scoring lm.word_score_tokenized( ["I was a matron in France", "I was a mat in France"], bos_token=True, # needed for GPT-2/Pythia and NOT needed for others tokenize_function=word_tokenizer, bow_correction=True, # Oh and Schuler correction surprisal=True, base_two=True ) ''' First word = -log_2 P(word | <beginning of text>) [[('I', 6.1522440910339355), ('was', 4.033324718475342), ('a', 4.879510402679443), ('matron', 17.611848831176758), ('in', 2.5804288387298584), ('France', 9.036953926086426)], [('I', 6.1522440910339355), ('was', 4.033324718475342), ('a', 4.879510402679443), ('mat', 19.385351181030273), ('in', 6.76780366897583), ('France', 10.574726104736328)]] '''
another day another minicons update (potentially a significant one for psycholinguists?)
"Word" scoring is now a thing! You just have to supply your own splitting function!
pip install -U minicons for merriment
Iโll also be presenting a talk based on this work Friday afternoon at HSP. Very excited to share it with a psycholinguistics-focused audience!
25.03.2025 23:21 โ ๐ 6 ๐ 1 ๐ฌ 0 ๐ 0Iโll be at #HSP2025! Iโm presenting a poster in session 4 on how semantic factors might affect timing data from a speeded cloze task (w @virmalised.us, Philip Resnik, and @colinphillips.bsky.social)
hsp2025.github.io/abstracts/19...
Iโll be presenting a poster at HSP 2025 in about a week. Itโs on memory for pronominal clitic placement in Spanish, come stop by and say hi if you can!
22.03.2025 22:33 โ ๐ 3 ๐ 2 ๐ฌ 0 ๐ 0๐ฌ๐ฅ๐ฟ Video of my keynote at MathPsych2024 now available online www.youtube.com/watch?v=WrwN...
#CogSci #CriticalAI #AIhype #AGI #PsychSci #PhilSci ๐งช
What do YOU mean by "intelligence", and does ChatGPT fit your definition?
We collected the major criteria used in CogSci and other fields, and designed a survey to find out!
Access link: www.survey-xact.dk/collect
Code: 4S7V-SN4M-S536
Time: 5-10 mins
starter pack for the Computational Linguistics and Information Processing group at the University of Maryland - get all your NLP and data science here!
go.bsky.app/V9qWjEi
@kanishka.bsky.social and I have made a starter pack for researchers working broadly on linguistic interpretability and LLMs!
go.bsky.app/F9qzAUn
Please message me or comment on this post if you've noticed someone who we forgot or would like to be added yourself!
"Hey everyone! ๐ Iโve created a starter pack of South Asian artists, authors, academics, activists, and orgs. Iโll keep it updatedโDM me or reply if you or someone you know should be added! โจ" go.bsky.app/GGd6dxU
23.11.2024 13:50 โ ๐ 152 ๐ 63 ๐ฌ 44 ๐ 4I'll be there as well - excited to chat!
11.11.2024 20:30 โ ๐ 1 ๐ 0 ๐ฌ 0 ๐ 0RNN LMs can learn many syntactic relations but fail to capture a shared generalization across constructions. Augmenting the training data with more examples helps, but not how we'd expect!
(with Katherine Howitt, @allidods.bsky.social , and Robert Hopkins)
aclanthology.org/2024.conll-1...
Language models can identify who did what to whom, but they may not be using human-like mechanisms, based on materials from different psycholinguistic studies finding systematic patterns in human processing.
(with Rosa Lee & Naomi Feldman)
aclanthology.org/2024.finding...
Iโll be presenting two posters on (psycho)linguistically motivated perspectives on LM generalization at #EMNLP2024!
1. Sensitivity to Argument Roles - Session 2 & #BlackBoxNLP
2. Learning & Filler-Gap Dependencies - #CoNLL
Excited to chat with other folks interested in compling x cogsci!
papersโฌ๏ธ
5-gram of the day: "language models from computational linguistics"
06.06.2024 15:17 โ ๐ 0 ๐ 0 ๐ฌ 0 ๐ 0Iโm so sorry but cool to see another psycholinguist who came from the CS route! I interned on a database team and hated it. The other reviews were good. Itโs just a very random process
18.04.2024 01:17 โ ๐ 1 ๐ 0 ๐ฌ 0 ๐ 0GRFP reviews sigh
16.04.2024 22:05 โ ๐ 1 ๐ 0 ๐ฌ 1 ๐ 0Today I learned that I may not have a successful psycholinguistics career because I got a B in databases.
16.04.2024 18:13 โ ๐ 2 ๐ 0 ๐ฌ 2 ๐ 0Panicked after seeing AGI on my tax form
30.03.2024 04:31 โ ๐ 2 ๐ 0 ๐ฌ 0 ๐ 0โThere is no ethical way to use the major AI image generators. All of them are trained on stolen images, and all of them are built for the purpose of deskilling, disempowering and replacing real human artists.โ
This sums it up perfectly. Itโs not a conversation.
04.03.2024 02:20 โ ๐ 17377 ๐ 7487 ๐ฌ 162 ๐ 163yelled about lexicalism in my NLU seminar do i get a prize
22.02.2024 19:38 โ ๐ 1 ๐ 0 ๐ฌ 0 ๐ 0LLMs are so weird because one side is people with five PhDs who have been studying neuron activations for the past three decades and on the other side is someone called leetm5n with an anime avatar just casually releasing increasingly better performing fine tunes of mistral
31.12.2023 01:32 โ ๐ 38 ๐ 5 ๐ฌ 2 ๐ 0Weโre excited about our first paper looking at speech encoding in single neurons across the depth of human cortex. Out today in @nature! www.nature.com/articles/s41... [1/6]
13.12.2023 16:48 โ ๐ 96 ๐ 49 ๐ฌ 9 ๐ 1