6/
We also wondered: if neuroscientists use functional localizers to map networks in the brain, could we do the same for MiCRoβs experts?
The answer: yes! The very same localizers successfully recovered the corresponding expert modules in our models!
20.10.2025 12:05 β π 1 π 1 π¬ 1 π 0
π Excited to share a major update to our βMixture of Cognitive Reasonersβ (MiCRo) paper!
We ask: What benefits can we unlock by designing language models whose inner structure mirrors the brainβs functional specialization?
More below π§ π
cognitive-reasoners.epfl.ch
20.10.2025 12:05 β π 28 π 9 π¬ 2 π 1
Check out @mryskina.bsky.social's talk and poster at COLM on Tuesdayβwe present a method to identify 'semantically consistent' brain regions (responding to concepts across modalities) and show that more semantically consistent brain regions are better predicted by LLMs.
04.10.2025 12:43 β π 15 π 4 π¬ 0 π 0
I'm recruiting PhD students to join my new lab in Fall 2026! The Shared Minds Lab at @usc.edu will combine deep learning and ecological human neuroscience to better understand how we communicate our thoughts from one brain to another.
01.10.2025 22:39 β π 117 π 72 π¬ 8 π 3
Do you want to use AI models to understand human language?
Are you fascinated by whether linguistic representations are lurking in LLMs?
Are you in need of a richer model of spatial words across languages?
Consider UT Austin for all your Computational Linguistics Ph.D. needs!
mahowak.github.io
30.09.2025 17:26 β π 6 π 1 π¬ 0 π 0
Elizabeth Lee smiles at the camera.
Elizabeth Lee, a first-year Ph.D. student in Neural Computation, has been awarded CMUβs 2025 Sutherland-Merlino Fellowship. Her work bridges neuroscience and machine learning, and sheβs passionate about advancing STEM access for underrepresented groups.
www.cmu.edu/mcs/news-eve...
30.09.2025 20:58 β π 7 π 3 π¬ 0 π 0
π¨ Paper alert:
To appear in the DBM Neurips Workshop
LITcoder: A General-Purpose Library for Building and Comparing Encoding Models
π arxiv: arxiv.org/abs/2509.091...
π project: litcoder-brain.github.io
29.09.2025 14:28 β π 18 π 4 π¬ 1 π 3
Now that the ICLR deadline is behind us, happy to share that From Language to Cognition has been accepted as an Oral at #EMNLP2025! π
Looking forward to seeing many of you in Suzhou π¨π³
25.09.2025 14:56 β π 20 π 3 π¬ 1 π 0
Excited to share new work with @hleemasson.bsky.social , Ericka Wodka, Stewart Mostofsky and @lisik.bsky.social! We investigated how simultaneous vision and language signals are combined in the brain using naturalistic+controlled fMRI. Read the paper here: osf.io/b5p4n
1/n
24.09.2025 19:46 β π 48 π 11 π¬ 1 π 2
Are there conceptual directions in VLMs that transcend modality? Check out our COLM oral spotlight π¦ paper! We use SAEs to analyze the multimodality of linear concepts in VLMs
with @chloesu07.bsky.social, @thomasfel.bsky.social, @shamkakade.bsky.social and Stephanie Gil
arxiv.org/abs/2504.11695
17.09.2025 19:12 β π 25 π 6 π¬ 1 π 1
Here is our best thinking about how to make world models. I would apologize for it being a massive 40-page behemoth, but it's worth reading. arxiv.org/pdf/2509.09737
15.09.2025 23:47 β π 71 π 18 π¬ 2 π 2
I thought I wouldnβt be one of those academics super into outreach talks, but I just put together something about understanding LLMs for laypeople and I get to talk about results that I donβt really focus on in any of my technical talks! Itβs actually really cool. I made this lil takeaway slide
10.09.2025 13:21 β π 78 π 8 π¬ 8 π 0
Interspeech paper title: What do self-supervised speech models know about Dutch? Analyzing advantages of language-specific pre-training
Authors: Marianne de Heer Kloots, Hosein Mohebbi, Charlotte Pouw, Gaofei Shen, Willem Zuidema, Martijn Bentum
β¨ Do self-supervised speech models learn to encode language-specific linguistic features from their training data, or only more language-general acoustic correlates?
At #Interspeech2025 we presented our new Wav2Vec2-NL model and SSL-NL evaluation dataset to test this!
π arxiv.org/abs/2506.00981
β¬οΈ
27.08.2025 14:31 β π 19 π 6 π¬ 1 π 0
So, what is #EurIPS anyway? π€
EurIPS is a community-driven conference taking place in Copenhagen Denmark endorsed by @neuripsconf.bsky.social and @nordicair.bsky.social and co-developed with @ellis.eu, where you can additionally present your NeurIPS papers.
27.08.2025 06:41 β π 17 π 5 π¬ 1 π 0
Had such a great time presenting our tutorial on Interpretability Techniques for Speech Models at #Interspeech2025! π
For anyone looking for an introduction to the topic, we've now uploaded all materials to the website: interpretingdl.github.io/speech-inter...
19.08.2025 21:23 β π 40 π 14 π¬ 2 π 1
Connectivity structure and dynamics of nonlinear recurrent neural networks
Studies of the dynamics of nonlinear recurrent neural networks often assume independent and identically distributed couplings, but large-scale connectomics data indicate that biological neural circuit...
Wanted to share a new version (much cleaner!) of a preprint on how connectivity structure shapes collective dynamics in nonlinear RNNs. Neural circuits have highly non-iid connectivity (e.g., rapidly decaying singular values, structured singular-vector overlaps), unlike classical random RNN models.
19.08.2025 15:42 β π 40 π 9 π¬ 1 π 0
Iβm recruiting committee members for the Technical Program Committee at #CCN2026.
Please apply if you want to help make submission, review & selection of contributed work (Extended Abstracts & Proceedings) more useful for everyone! π
Helps to have: programming/communications/editorial experience.
19.08.2025 14:12 β π 19 π 14 π¬ 3 π 1
We hope that AuriStream will serve as a task-performant model system for studying how language structure is learned from speech.
The Interspeech paper sets the stageβmore work building on this idea coming soon! And as always, please feel free to get in touch with comments etc.!
19.08.2025 01:12 β π 1 π 0 π¬ 0 π 0
3οΈβ£ Temporally fine-grained β 5ms tokens preserve acoustic detail (e.g. speaker identity).
4οΈβ£ Unified β AuriStream learns strong speech representations and generates plausible continuationsβbridging representation learning and sequence modeling in the audio domain.
19.08.2025 01:12 β π 2 π 0 π¬ 1 π 0
4 key advantages of AuriStream:
1οΈβ£ Causal β allows the study of speech/language processing as it unfolds in real time.
2οΈβ£ Inspectable β predictions can naturally be decoded into the cochleagram/audio, enabling visualization and interpretation.
19.08.2025 01:12 β π 2 π 0 π¬ 1 π 0
Examples: audio before red line = ground-truth prompt; after = AuriStreamβs prediction, visualized in the time-frequency cochleagram space.
AuriStream shows that causal prediction over short audio chunks (cochlear tokens) is enough to generate meaningful sentence continuations!
19.08.2025 01:12 β π 3 π 1 π¬ 1 π 0
Complementing AuriStreamβs strong representational capabilities, AuriStream learns short- and long-range speech statisticsβcompleting phonemes and common words at short scales, and generating diverse continuations at longer scales, as evident by the qualitative examples below.
19.08.2025 01:12 β π 1 π 0 π¬ 1 π 0
We demonstrate that:
πΉ AuriStream embeddings capture information about phoneme identity, word identity, and lexical semantics.
πΉ AuriStream embeddings serve as a strong backbone on downstream audio tasks (SUPERB benchmark, such as ASR and intent classification).
19.08.2025 01:12 β π 1 π 0 π¬ 1 π 0
We present a two-stage framework, loosely inspired by the human auditory hierarchy:
1οΈβ£ WavCoch: a small model that transforms raw audio into a cochlea-like time-frequency representation, from which we extract discrete βcochlear tokensβ.
2οΈβ£ AuriStream: an autoregressive model over the cochlear tokens.
19.08.2025 01:12 β π 2 π 0 π¬ 1 π 0
Many prior speech-based models rely on heuristics such as:
πΉ Global clustering of the embedding space
πΉ Non-causal objectives
πΉ Fixed-duration βlanguageβ units
...
We believe that no high-performing, open-source audio model exists without such constraintsβAuriStream is built to fill that gap.
19.08.2025 01:12 β π 3 π 0 π¬ 1 π 0
ISCA Archive - Representing Speech Through Autoregressive Prediction of Cochlear Tokens
Joint with @klemenkotar.bsky.social, and with @evfedorenko.bsky.social @dyamins.bsky.social
Paper: www.isca-archive.org/interspeech_...
Website: tukoresearch.github.io/auristream-s... (with audio examples)
HuggingFace: huggingface.co/TuKoResearch...
19.08.2025 01:12 β π 2 π 1 π¬ 1 π 0
Humans largely learn language through speech. In contrast, most LLMs learn from pre-tokenized text.
In our #Interspeech2025 paper, we introduce AuriStream: a simple, causal model that learns phoneme, word & semantic information from speech.
Poster P6, tomorrow (Aug 19) at 1:30 pm, Foyer 2.2!
19.08.2025 01:12 β π 52 π 10 π¬ 1 π 1
In our new paper, we explore how we can build encoding models that are both powerful and understandable. Our model uses an LLM to answer 35 questions about a sentence's content. The answers linearly contribute to our prediction of how the brain will respond to that sentence. 1/6
18.08.2025 09:44 β π 25 π 9 π¬ 1 π 1
Computer Science + Cognitive Science @harvard.edu, class of '26. Interested in language β© thought, language acquisition.
Visiting Student @MITCoCoSci @csail.mit.edu
PhD student @csail.mit.edu π€ & π§
Postdoctoral researcher, studying belief-based exploration and (un)predictability in biological and artificial agents
https://thirzadado.com/
neuroscience and behavior in parrots and songbirds
Simons junior fellow and post-doc at NYU Langone studying vocal communication, PhD MIT brain and cognitive sciences
Interpretable Deep Networks. http://baulab.info/ @davidbau
PhD student in neuroscience at McGill, working on origins of music πΆ π§
Studying neural computation β’ Assistant Professor of Neuroscience at Baylor College of Medicine β’ lipshutzlab.com
Neuroscience postdoc at Stanford and Enigma (enigmaproject.ai), previously philosophy of neuroscience PhD. Building large-scale brain models using deep learning.
Senior Director of AI/ML Research Engineering, Kempner Institute @Harvard , views are my own
Research Fellow @ox.ac.uk | Multimodal ML PhD University of Amsterdam | Previously @msftresearch.bsky.social, Google AI Gemini, Bloomberg AI, Amazon Science, ETH, KU Leuven | MS AI, BS Computational Linguistics
mariyahendriksen.github.io
PhD student in cognitive neuroscience at UT Dallas
post-bac @kempnerinstitute.bsky.social⬠at Harvard, studying intuitive physics in CCNLab under @gershbrain.bsky.social and @tomerullman.bsky.social
https://beheryd.github.io/
theory of neural networks for natural and artificial intelligence
https://pehlevan.seas.harvard.edu/
π Please now follow @interspeech.bsky.social instead!
I work on speech and language technologies at Google. I like languages, history, maps, traveling, cycling, and buying way too many books.
Postdoc @ Neural dynamics of visual cognition (Cichy lab) @FU_Berlin
interested in languageπ£οΈπ and its representations in the brainπ§
https://jiawei-liiiii.github.io/
neuroscience, AI, methods/workflows/data; research assistant prof. @ Northwestern, affiliate @ MNI & MPI CBS| he/him |
https://peerherholz.github.io/
Assistant Professor @UvA_Amsterdam | Cognitive neuroscience, Scene perception, Computational vision | Chair of CCN2025 | www.irisgroen.com
Postdoctoral researcher @NeuroSpin | AI π€ & neuroscienceΒ π§ Β enthusiast
https://linktr.ee/alirezakr
Computational Cognitive Neuroscientist at INS Marseille. I'm interested in speech, information theory and network modelling.