Kanishka Misra 🌊's Avatar

Kanishka Misra 🌊

@kanishka.bsky.social

Assistant Professor of Linguistics, and Harrington Fellow at UT Austin. Works on computational understanding of language, concepts, and generalization. πŸ•ΈοΈπŸ‘οΈ: https://kanishka.website

2,469 Followers  |  261 Following  |  193 Posts  |  Joined: 06.07.2023  |  2.4161

Latest posts by kanishka.bsky.social on Bluesky

I will be giving a short talk on this work at the COLM Interplay workshop on Friday (also to appear at EMNLP)!

Will be in Montreal all week and excited to chat about LM interpretability + its interaction with human cognition and ling theory.

06.10.2025 12:05 β€” πŸ‘ 8    πŸ” 5    πŸ’¬ 0    πŸ“Œ 0
Post image

On my way to #COLM2025 🍁

Check out jessyli.com/colm2025

QUDsim: Discourse templates in LLM stories arxiv.org/abs/2504.09373

EvalAgent: retrieval-based eval targeting implicit criteria arxiv.org/abs/2504.15219

RoboInstruct: code generation for robotics with simulators arxiv.org/abs/2405.20179

06.10.2025 15:50 β€” πŸ‘ 12    πŸ” 4    πŸ’¬ 0    πŸ“Œ 0
Preview
Language Models Fail to Introspect About Their Knowledge of Language There has been recent interest in whether large language models (LLMs) can introspect about their own internal states. Such abilities would make LLMs more interpretable, and also validate the use of s...

I’m at #COLM2025 from Wed with:

@siyuansong.bsky.social Tue am introspection arxiv.org/abs/2503.07513

@qyao.bsky.social Wed am controlled rearing: arxiv.org/abs/2503.20850

@sashaboguraev.bsky.social INTERPLAY ling interp: arxiv.org/abs/2505.16002

I’ll talk at INTERPLAY too. Come say hi!

06.10.2025 15:57 β€” πŸ‘ 20    πŸ” 6    πŸ’¬ 1    πŸ“Œ 0

Heading to #COLM2025 to present my first paper w/ @jennhu.bsky.social @kmahowald.bsky.social !

When: Tuesday, 11 AM – 1 PM
Where: Poster #75

Happy to chat about my work and topics in computational linguistics & cogsci!

Also, I'm on the PhD application journey this cycle!

Paper info πŸ‘‡:

06.10.2025 16:05 β€” πŸ‘ 7    πŸ” 3    πŸ’¬ 0    πŸ“Œ 0

Also, I’m on the look out for my first PhD student! If you’d like to be the one, please reach out to me (dms/email open) and we can chat!!

@jessyjli.bsky.social and @kmahowald.bsky.social are also hiring students, and we’re all eager to co-advise!

06.10.2025 15:22 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

I’ll also be moderating a roundtable at the INTERPLAY workshop on Oct 10 β€” excited to discuss behavior, representations, and a third secret thing with folks!

06.10.2025 15:22 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Preview
On Language Models' Sensitivity to Suspicious Coincidences Humans are sensitive to suspicious coincidences when generalizing inductively over data, as they make assumptions as to how the data was sampled. This results in smaller, more specific hypotheses bein...

2. @sriramp05.bsky.social on LMs and Suspicious Coincidences at the PragLM workshop poster session on Friday, (work w/ me, @kmahowald.bsky.social, and @eunsol.bsky.social)

Paper: arxiv.org/abs/2504.09387

06.10.2025 15:22 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Preview
Both Direct and Indirect Evidence Contribute to Dative Alternation Preferences in Language Models Language models (LMs) tend to show human-like preferences on a number of syntactic phenomena, but the extent to which these are attributable to direct exposure to the phenomena or more general propert...

Traveling to my first @colmweb.org🍁

Not presenting anything but here are two posters you should visit:

1. @qyao.bsky.social on Controlled rearing for direct and indirect evidence for datives (w/ me, @weissweiler.bsky.social and @kmahowald.bsky.social), W morning

Paper: arxiv.org/abs/2503.20850

06.10.2025 15:22 β€” πŸ‘ 13    πŸ” 5    πŸ’¬ 1    πŸ“Œ 0

All of us (@kanishka.bsky.social @kmahowald.bsky.social and me) are looking for PhD students this cycle! If computational linguistics/NLP is your passion, join us at UT Austin!

For my areas see jessyli.com

30.09.2025 19:30 β€” πŸ‘ 4    πŸ” 5    πŸ’¬ 0    πŸ“Œ 0

We'll all be attending #COLM2025 -- come say hi if you are interested in working with us!!

Separate tweet incoming for COLM papers!

30.09.2025 16:17 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Preview
Characterizing the Role of Similarity in the Property Inferences of Language Models Property inheritance -- a phenomenon where novel properties are projected from higher level categories (e.g., birds) to lower level ones (e.g., sparrows) -- provides a unique window into how humans or...

Understanding how internal representations drive conceptual behavior in LMs

E.g., arxiv.org/abs/2410.22590

30.09.2025 16:17 β€” πŸ‘ 2    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Preview
Vision-and-Language Training Helps Deploy Taxonomic Knowledge but Does Not Fundamentally Alter It Does vision-and-language (VL) training change the linguistic representations of language models in meaningful ways? Most results in the literature have shown inconsistent or marginal differences, both...

Role of language (vs. other modalities) in learning meaning-sensitivities

E.g., arxiv.org/abs/2507.13328

30.09.2025 16:17 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Preview
Generating novel experimental hypotheses from language models: A case study on cross-dative generalization Neural network language models (LMs) have been shown to successfully capture complex linguistic knowledge. However, their utility for understanding language acquisition is still debated. We contribute...

Using neural networks to generate experimental hypotheses about language acquisition in scenarios where hypothesis-spaces are intractable

E.g., arxiv.org/abs/2408.05086

30.09.2025 16:17 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Preview
Language Models Learn Rare Phenomena from Less Rare Phenomena: The Case of the Missing AANNs Language models learn rare syntactic phenomena, but the extent to which this is attributable to generalization vs. memorization is a major open question. To that end, we iteratively trained transforme...

"Controlled Rearing" of LMs to understand the role of input in acquiring linguistic generalization

E.g., arxiv.org/abs/2403.19827, arxiv.org/abs/2503.20850

30.09.2025 16:17 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Picture of the UT Tower with "UT Austin Computational Linguistics" written in bigger font, and "Humans processing computers processing human processing language" in smaller font

Picture of the UT Tower with "UT Austin Computational Linguistics" written in bigger font, and "Humans processing computers processing human processing language" in smaller font

The compling group at UT Austin (sites.utexas.edu/compling/) is looking for PhD students!

Come join me, @kmahowald.bsky.social, and @jessyjli.bsky.social as we tackle interesting research questions at the intersection of ling, cogsci, and ai!

Some topics I am particularly interested in:

30.09.2025 16:17 β€” πŸ‘ 18    πŸ” 10    πŸ’¬ 3    πŸ“Œ 2
Sigmoid function. Non-linearities in neural network allow it to behave in distributed and near-symbolic fashions.

Sigmoid function. Non-linearities in neural network allow it to behave in distributed and near-symbolic fashions.

New paper! 🚨 I argue that LLMs represent a synthesis between distributed and symbolic approaches to language, because, when exposed to language, they develop highly symbolic representations and processing mechanisms in addition to distributed ones.
arxiv.org/abs/2502.11856

30.09.2025 13:15 β€” πŸ‘ 26    πŸ” 11    πŸ’¬ 1    πŸ“Œ 0

Hehe but really β€” unifying all of them + easy access = yes plssss

27.09.2025 15:01 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

Friendship ended with minicons, glazing is my new fav package!

27.09.2025 14:47 β€” πŸ‘ 5    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

God’s work πŸ™

27.09.2025 14:46 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Preview
GitHub - aaronstevenwhite/glazing: Unified data models and interfaces for syntactic and semantic frame ontologies. Unified data models and interfaces for syntactic and semantic frame ontologies. - aaronstevenwhite/glazing

I've found it kind of a pain to work with resources like VerbNet, FrameNet, PropBank (frame files), and WordNet using existing tools. Maybe you have too. Here's a little package that handles data management, loading, and cross-referencing via either a CLI or a python API.

27.09.2025 13:51 β€” πŸ‘ 27    πŸ” 7    πŸ’¬ 3    πŸ“Œ 1

Would you say it’s a dead area right now? (Ignoring the podcasts)

26.09.2025 03:14 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

Love to start the day by mistakenly stumbling onto hate speech against south-asians ty internet

21.09.2025 15:09 β€” πŸ‘ 4    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

Accepted at #NeurIPS2025! So proud of Yulu and Dheeraj for leading this! Be on the lookout for more "nuanced yes/no" work from them in the future πŸ‘€

18.09.2025 16:12 β€” πŸ‘ 6    πŸ” 1    πŸ’¬ 0    πŸ“Œ 0

Abstract deadline changed to *December 1, 2025*

07.09.2025 21:48 β€” πŸ‘ 13    πŸ” 5    πŸ’¬ 0    πŸ“Œ 0
How Linguistics Learned to Stop Worrying and Love the Language Models How Linguistics Learned to Stop Worrying and Love the Language Models

πŸ“£@futrell.bsky.social and I have a BBS target article with an optimistic take on LLMs + linguistics. Commentary proposals (just need a few hundred words) are OPEN until Oct 8. If we are too optimistic for you (or not optimistic enough!) or you have anything to say: www.cambridge.org/core/journal...

15.09.2025 15:46 β€” πŸ‘ 50    πŸ” 10    πŸ’¬ 4    πŸ“Œ 3

Happy and proud to see @rjantonello.bsky.social’s work awarded by SNL!

13.09.2025 21:47 β€” πŸ‘ 28    πŸ” 4    πŸ’¬ 1    πŸ“Œ 0
Density plot with X axis being probability of text being synthetic from an AI detector model. Plots show that GPT4.1 outputs are assigned high probability of being AI text, but GPT5 outputs are assigned low probability of being AI text.

Density plot with X axis being probability of text being synthetic from an AI detector model. Plots show that GPT4.1 outputs are assigned high probability of being AI text, but GPT5 outputs are assigned low probability of being AI text.

Exhibit N on how synthetic text/AI detectors just don't work reliably. Generating some (long) sentences from GPT4.1 and GPT5 with the same prompt, the top open-source model on the RAID benchmark classifies most GPT4.1 outputs as synthetic and most GPT5 as not synthetic.

10.09.2025 20:05 β€” πŸ‘ 5    πŸ” 1    πŸ’¬ 0    πŸ“Œ 0
Title: The Cats of Cogsci. Two cats, Coco and Loki, with Northwestern Cognitive Science logo in the background. Coco is sitting on some books and Loki is holding an apple. Both are wearing glasses b/c they are academics.

Title: The Cats of Cogsci. Two cats, Coco and Loki, with Northwestern Cognitive Science logo in the background. Coco is sitting on some books and Loki is holding an apple. Both are wearing glasses b/c they are academics.

Loki the cat has his paw on a laptop; text "Remember to add Cog Sci 110 to you shopping cart now in Caesar so that you're ready to enroll come September 12th!"

Loki the cat has his paw on a laptop; text "Remember to add Cog Sci 110 to you shopping cart now in Caesar so that you're ready to enroll come September 12th!"

Super happy with Cogsci program assistant Chris Kent's work for our college Instagram feed. Glad I could get our Loki featured to advertise my class.

06.09.2025 20:15 β€” πŸ‘ 13    πŸ” 2    πŸ’¬ 0    πŸ“Œ 0

Excited to speak alongside such an illustrious set of speakers!

24.08.2025 14:49 β€” πŸ‘ 6    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Preview
Composition with Grid A few months into the pandemic, my wife and I adopted a new pastime: we would complete the New York Times crossword puzzle every day.

Lovely write up by @ksetiya.bsky.social on @rkubala.bsky.social’s
piece on the art of crosswords! Come for Robbie, stay for Sondheim crossword quotes. ksetiya.substack.com/p/compositio...

23.08.2025 18:45 β€” πŸ‘ 3    πŸ” 1    πŸ’¬ 0    πŸ“Œ 0

@kanishka is following 20 prominent accounts