Linyang He's Avatar

Linyang He

@linyanghe.bsky.social

PhD Student @ Mesgarani Lab, Columbia University Neuroscience+ML+Language https://linyanghe.github.io/

45 Followers  |  117 Following  |  9 Posts  |  Joined: 28.02.2025  |  1.4301

Latest posts by linyanghe.bsky.social on Bluesky

Many thanks to my amazing co-authors:
@tianjunzhong.bsky.social, @rjantonello.bsky.social, Gavin Mischler, Prof. Micah Goldblum and my advisor Prof. Nima Mesgarani!

#NeuroAI #LLM #NeurIPS2025 #NeurIPS

30.10.2025 22:25 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

5️⃣ Takeaway:
- Raw LLM embeddings = biased toward shallow linguistic features.
- Residual disentanglement exposes the deeper, reasoning-specific representations shared by brains and models.

30.10.2025 22:25 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Post image

4️⃣Spatial pattern: reasoning even recruits visual cortex beyond classical language areas.

30.10.2025 22:25 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Post image

3️⃣ Temporal dynamics: reasoning peaks later (~350–400 ms) than shallow features.

30.10.2025 22:25 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Post image

2️⃣ We introduce the first "reasoning embedding", a disentangled representation that isolates reasoning from lexicon, syntax, and meaning.
- The disentangled representations are orthogonal to each other.

30.10.2025 22:25 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

1️⃣ Why "Far from the Shallow"?
- Traditional LLM embeddings are entangled, they mix shallow linguistic features (lexicon/syntax) with deeper signals.
- This makes brain encoding studies misleading: success often comes from shallow correlations, not true semantics/reasoning alignment.

30.10.2025 22:25 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Post image

🧠 New at #NeurIPS2025!
🎡 We're far from the shallow now🎡
TL;DR: We introduce the first "reasoning embedding" and uncover its unique spatio-temporal pattern in the brain.

πŸ”— arxiv.org/abs/2510.228...

30.10.2025 22:25 β€” πŸ‘ 8    πŸ” 4    πŸ’¬ 1    πŸ“Œ 0

3️⃣ Unique spatial-temporal pattern of reasoning:
- Temporal dynamics: reasoning peaks later (~350–400 ms).
- Spatially: it even recruits visual cortex beyond classical language areas (IFG/STG), suggesting reasoning involves multimodal integration.
(4/6)

30.10.2025 21:39 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

2️⃣ Our contribution:
- We introduce the first β€œreasoning embedding”, a disentangled representation that isolates reasoning from lexicon, syntax, and meaning.
- It captures variance in brain activity that shallow features can't explain, revealing a distinct neural signature for reasoning.
(3/6)

30.10.2025 21:39 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Post image

🌍Introducing BabyBabelLM: A Multilingual Benchmark of Developmentally Plausible Training Data!

LLMs learn from vastly more data than humans ever experience. BabyLM challenges this paradigm by focusing on developmentally plausible data

We extend this effort to 45 new languages!

15.10.2025 10:53 β€” πŸ‘ 43    πŸ” 16    πŸ’¬ 1    πŸ“Œ 3
Preview
Temporal integration in human auditory cortex is predominantly yoked to absolute time - Nature Neuroscience Temporal integration throughout the human auditory cortex is predominantly locked to absolute time and does not vary with the duration of speech structures such as phonemes or words.

What happens when you listen to speech a different speeds? Does your brain change its processing speed too? It turns out, no
@samnorman-haignere.bsky.social & researchers at
@zuckermanbrain.bsky.social found the auditory part of the brain keeps clocking in at a fixed time
via @natneuro.nature.com

18.09.2025 18:15 β€” πŸ‘ 6    πŸ” 3    πŸ’¬ 1    πŸ“Œ 0
Video thumbnail

In our new paper, we explore how we can build encoding models that are both powerful and understandable. Our model uses an LLM to answer 35 questions about a sentence's content. The answers linearly contribute to our prediction of how the brain will respond to that sentence. 1/6

18.08.2025 09:44 β€” πŸ‘ 25    πŸ” 9    πŸ’¬ 1    πŸ“Œ 1

@linyanghe is following 20 prominent accounts