Greta Tuckute's Avatar

Greta Tuckute

@gretatuckute.bsky.social

Studying language in biological brains and artificial ones at the Kempner Institute at Harvard University. www.tuckute.com

873 Followers  |  270 Following  |  76 Posts  |  Joined: 13.12.2023  |  1.7867

Latest posts by gretatuckute.bsky.social on Bluesky

Post image

6/
We also wondered: if neuroscientists use functional localizers to map networks in the brain, could we do the same for MiCRo’s experts?

The answer: yes! The very same localizers successfully recovered the corresponding expert modules in our models!

20.10.2025 12:05 β€” πŸ‘ 1    πŸ” 1    πŸ’¬ 1    πŸ“Œ 0
Post image

πŸš€ Excited to share a major update to our β€œMixture of Cognitive Reasoners” (MiCRo) paper!

We ask: What benefits can we unlock by designing language models whose inner structure mirrors the brain’s functional specialization?

More below πŸ§ πŸ‘‡
cognitive-reasoners.epfl.ch

20.10.2025 12:05 β€” πŸ‘ 28    πŸ” 9    πŸ’¬ 2    πŸ“Œ 1

Check out @mryskina.bsky.social's talk and poster at COLM on Tuesdayβ€”we present a method to identify 'semantically consistent' brain regions (responding to concepts across modalities) and show that more semantically consistent brain regions are better predicted by LLMs.

04.10.2025 12:43 β€” πŸ‘ 15    πŸ” 4    πŸ’¬ 0    πŸ“Œ 0
Post image

I'm recruiting PhD students to join my new lab in Fall 2026! The Shared Minds Lab at @usc.edu will combine deep learning and ecological human neuroscience to better understand how we communicate our thoughts from one brain to another.

01.10.2025 22:39 β€” πŸ‘ 117    πŸ” 72    πŸ’¬ 8    πŸ“Œ 3

Do you want to use AI models to understand human language?

Are you fascinated by whether linguistic representations are lurking in LLMs?

Are you in need of a richer model of spatial words across languages?

Consider UT Austin for all your Computational Linguistics Ph.D. needs!

mahowak.github.io

30.09.2025 17:26 β€” πŸ‘ 6    πŸ” 1    πŸ’¬ 0    πŸ“Œ 0
Elizabeth Lee smiles at the camera.

Elizabeth Lee smiles at the camera.

Elizabeth Lee, a first-year Ph.D. student in Neural Computation, has been awarded CMU’s 2025 Sutherland-Merlino Fellowship. Her work bridges neuroscience and machine learning, and she’s passionate about advancing STEM access for underrepresented groups.
www.cmu.edu/mcs/news-eve...

30.09.2025 20:58 β€” πŸ‘ 7    πŸ” 3    πŸ’¬ 0    πŸ“Œ 0
Post image

🚨 Paper alert:
To appear in the DBM Neurips Workshop

LITcoder: A General-Purpose Library for Building and Comparing Encoding Models

πŸ“„ arxiv: arxiv.org/abs/2509.091...
πŸ”— project: litcoder-brain.github.io

29.09.2025 14:28 β€” πŸ‘ 18    πŸ” 4    πŸ’¬ 1    πŸ“Œ 3
Preview
Faculty Position in Neuroscience The School of Life Sciences at EPFL invites applications for a Tenure Track Assistant Professor position in Neuroscience. At EPFL researchers develop and apply innovative technologies to understand br...

Come be our colleague at EPFL! Several open calls for positions πŸ§ͺπŸ§ πŸ€–

* Neuroscience www.epfl.ch/about/workin... (deadline Oct 1)

* Life Science Engineering www.epfl.ch/about/workin...

* CS general call www.epfl.ch/about/workin...

* Learning Sciences www.epfl.ch/about/workin...

29.09.2025 11:46 β€” πŸ‘ 31    πŸ” 10    πŸ’¬ 1    πŸ“Œ 0

Now that the ICLR deadline is behind us, happy to share that From Language to Cognition has been accepted as an Oral at #EMNLP2025! πŸŽ‰

Looking forward to seeing many of you in Suzhou πŸ‡¨πŸ‡³

25.09.2025 14:56 β€” πŸ‘ 20    πŸ” 3    πŸ’¬ 1    πŸ“Œ 0
Post image

Excited to share new work with @hleemasson.bsky.social , Ericka Wodka, Stewart Mostofsky and @lisik.bsky.social! We investigated how simultaneous vision and language signals are combined in the brain using naturalistic+controlled fMRI. Read the paper here: osf.io/b5p4n
1/n

24.09.2025 19:46 β€” πŸ‘ 48    πŸ” 11    πŸ’¬ 1    πŸ“Œ 2
Post image

Are there conceptual directions in VLMs that transcend modality? Check out our COLM oral spotlight πŸ”¦ paper! We use SAEs to analyze the multimodality of linear concepts in VLMs

with @chloesu07.bsky.social, @thomasfel.bsky.social, @shamkakade.bsky.social and Stephanie Gil
arxiv.org/abs/2504.11695

17.09.2025 19:12 β€” πŸ‘ 25    πŸ” 6    πŸ’¬ 1    πŸ“Œ 1

Here is our best thinking about how to make world models. I would apologize for it being a massive 40-page behemoth, but it's worth reading. arxiv.org/pdf/2509.09737

15.09.2025 23:47 β€” πŸ‘ 71    πŸ” 18    πŸ’¬ 2    πŸ“Œ 2
Post image

I thought I wouldnβ€˜t be one of those academics super into outreach talks, but I just put together something about understanding LLMs for laypeople and I get to talk about results that I don’t really focus on in any of my technical talks! It’s actually really cool. I made this lil takeaway slide

10.09.2025 13:21 β€” πŸ‘ 78    πŸ” 8    πŸ’¬ 8    πŸ“Œ 0
Interspeech paper title: What do self-supervised speech models know about Dutch? Analyzing advantages of language-specific pre-training

Authors: Marianne de Heer Kloots, Hosein Mohebbi, Charlotte Pouw, Gaofei Shen, Willem Zuidema, Martijn Bentum

Interspeech paper title: What do self-supervised speech models know about Dutch? Analyzing advantages of language-specific pre-training Authors: Marianne de Heer Kloots, Hosein Mohebbi, Charlotte Pouw, Gaofei Shen, Willem Zuidema, Martijn Bentum

✨ Do self-supervised speech models learn to encode language-specific linguistic features from their training data, or only more language-general acoustic correlates?

At #Interspeech2025 we presented our new Wav2Vec2-NL model and SSL-NL evaluation dataset to test this!

πŸ“„ arxiv.org/abs/2506.00981

⬇️

27.08.2025 14:31 β€” πŸ‘ 19    πŸ” 6    πŸ’¬ 1    πŸ“Œ 0
Post image

So, what is #EurIPS anyway? πŸ€”

EurIPS is a community-driven conference taking place in Copenhagen Denmark endorsed by @neuripsconf.bsky.social and @nordicair.bsky.social and co-developed with @ellis.eu, where you can additionally present your NeurIPS papers.

27.08.2025 06:41 β€” πŸ‘ 17    πŸ” 5    πŸ’¬ 1    πŸ“Œ 0

Had such a great time presenting our tutorial on Interpretability Techniques for Speech Models at #Interspeech2025! πŸ”

For anyone looking for an introduction to the topic, we've now uploaded all materials to the website: interpretingdl.github.io/speech-inter...

19.08.2025 21:23 β€” πŸ‘ 40    πŸ” 14    πŸ’¬ 2    πŸ“Œ 1
Preview
Connectivity structure and dynamics of nonlinear recurrent neural networks Studies of the dynamics of nonlinear recurrent neural networks often assume independent and identically distributed couplings, but large-scale connectomics data indicate that biological neural circuit...

Wanted to share a new version (much cleaner!) of a preprint on how connectivity structure shapes collective dynamics in nonlinear RNNs. Neural circuits have highly non-iid connectivity (e.g., rapidly decaying singular values, structured singular-vector overlaps), unlike classical random RNN models.

19.08.2025 15:42 β€” πŸ‘ 40    πŸ” 9    πŸ’¬ 1    πŸ“Œ 0

I’m recruiting committee members for the Technical Program Committee at #CCN2026.

Please apply if you want to help make submission, review & selection of contributed work (Extended Abstracts & Proceedings) more useful for everyone! 🌐

Helps to have: programming/communications/editorial experience.

19.08.2025 14:12 β€” πŸ‘ 19    πŸ” 14    πŸ’¬ 3    πŸ“Œ 1

We hope that AuriStream will serve as a task-performant model system for studying how language structure is learned from speech.

The Interspeech paper sets the stageβ€”more work building on this idea coming soon! And as always, please feel free to get in touch with comments etc.!

19.08.2025 01:12 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

3️⃣ Temporally fine-grained β†’ 5ms tokens preserve acoustic detail (e.g. speaker identity).
4️⃣ Unified β†’ AuriStream learns strong speech representations and generates plausible continuationsβ€”bridging representation learning and sequence modeling in the audio domain.

19.08.2025 01:12 β€” πŸ‘ 2    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

4 key advantages of AuriStream:

1️⃣ Causal β†’ allows the study of speech/language processing as it unfolds in real time.
2️⃣ Inspectable β†’ predictions can naturally be decoded into the cochleagram/audio, enabling visualization and interpretation.

19.08.2025 01:12 β€” πŸ‘ 2    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Video thumbnail

Examples: audio before red line = ground-truth prompt; after = AuriStream’s prediction, visualized in the time-frequency cochleagram space.

AuriStream shows that causal prediction over short audio chunks (cochlear tokens) is enough to generate meaningful sentence continuations!

19.08.2025 01:12 β€” πŸ‘ 3    πŸ” 1    πŸ’¬ 1    πŸ“Œ 0
Post image

Complementing AuriStream’s strong representational capabilities, AuriStream learns short- and long-range speech statisticsβ€”completing phonemes and common words at short scales, and generating diverse continuations at longer scales, as evident by the qualitative examples below.

19.08.2025 01:12 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

We demonstrate that:

πŸ”Ή AuriStream embeddings capture information about phoneme identity, word identity, and lexical semantics.
πŸ”Ή AuriStream embeddings serve as a strong backbone on downstream audio tasks (SUPERB benchmark, such as ASR and intent classification).

19.08.2025 01:12 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Post image

We present a two-stage framework, loosely inspired by the human auditory hierarchy:

1️⃣ WavCoch: a small model that transforms raw audio into a cochlea-like time-frequency representation, from which we extract discrete β€œcochlear tokens”.
2️⃣ AuriStream: an autoregressive model over the cochlear tokens.

19.08.2025 01:12 β€” πŸ‘ 2    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

Many prior speech-based models rely on heuristics such as:
πŸ”Ή Global clustering of the embedding space
πŸ”Ή Non-causal objectives
πŸ”Ή Fixed-duration β€œlanguage” units
...

We believe that no high-performing, open-source audio model exists without such constraintsβ€”AuriStream is built to fill that gap.

19.08.2025 01:12 β€” πŸ‘ 3    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
ISCA Archive - Representing Speech Through Autoregressive Prediction of Cochlear Tokens

Joint with @klemenkotar.bsky.social, and with @evfedorenko.bsky.social @dyamins.bsky.social

Paper: www.isca-archive.org/interspeech_...

Website: tukoresearch.github.io/auristream-s... (with audio examples)

HuggingFace: huggingface.co/TuKoResearch...

19.08.2025 01:12 β€” πŸ‘ 2    πŸ” 1    πŸ’¬ 1    πŸ“Œ 0
Video thumbnail

Humans largely learn language through speech. In contrast, most LLMs learn from pre-tokenized text.

In our #Interspeech2025 paper, we introduce AuriStream: a simple, causal model that learns phoneme, word & semantic information from speech.

Poster P6, tomorrow (Aug 19) at 1:30 pm, Foyer 2.2!

19.08.2025 01:12 β€” πŸ‘ 52    πŸ” 10    πŸ’¬ 1    πŸ“Œ 1
Preview
Kempner Research Fellowship - Kempner Institute The Kempner brings leading, early-stage postdoctoral scientists to Harvard to work on projects that advance the fundamental understanding of intelligence.

If you work on artificial or natural intelligence and are finishing your PhD, consider applying for a Kempner research fellowship at Harvard:
kempnerinstitute.harvard.edu/kempner-inst...

18.08.2025 17:27 β€” πŸ‘ 48    πŸ” 31    πŸ’¬ 0    πŸ“Œ 0
Video thumbnail

In our new paper, we explore how we can build encoding models that are both powerful and understandable. Our model uses an LLM to answer 35 questions about a sentence's content. The answers linearly contribute to our prediction of how the brain will respond to that sentence. 1/6

18.08.2025 09:44 β€” πŸ‘ 25    πŸ” 9    πŸ’¬ 1    πŸ“Œ 1

@gretatuckute is following 20 prominent accounts