Title page of SCIL extended abstract titled: semantic-features: A User-Friendly Tool for Studying Contextual Word Embeddings in Interpretable Semantic Spaces
I will unfortunately have to skip SCiL this year, but I am thrilled to share that Jwalanthi will be presenting this work by her, @rjha.bsky.social, me, and @kmahowald.bsky.social on a tool that allows you to project contextualized embeddings from LMs to interpretable semantic spaces!
17.07.2025 17:47 — 👍 15 🔁 4 💬 1 📌 1
Finally, we developed a set of interactive tutorials for preprocessing and running encoding models to get you started. Happy to hear any feedback or field any questions about the dataset! hassonlab.github.io/podcast-ecog...
07.07.2025 21:00 — 👍 7 🔁 2 💬 0 📌 0
We validated both the data and stimulus features using encoding models, replicating previous findings showing an advantage for LLM embeddings.
07.07.2025 21:00 — 👍 1 🔁 0 💬 1 📌 0
We also provide word-level transcripts and stimulus features ranging from low-level acoustic features to large language model embeddings.
07.07.2025 21:00 — 👍 0 🔁 0 💬 1 📌 0
We recorded ECoG data in nine subjects while they listened to a 30-minute story. We provide a minimally preprocessed derivative of the raw data, ready to be used.
07.07.2025 21:00 — 👍 1 🔁 0 💬 1 📌 0
Check out Zaid's open "Podcast" ECoG dataset for natural language comprehension (w/ Hasson Lab). The paper is now out at Scientific Data (nature.com/articles/s41...) and the data are available on OpenNeuro (openneuro.org/datasets/ds0...).
07.07.2025 21:00 — 👍 35 🔁 17 💬 1 📌 0
These findings suggest that, despite the diversity of languages, shared meaning emerges from our interactions with one another and our shared world.
30.06.2025 20:56 — 👍 3 🔁 2 💬 1 📌 0
Our results suggest that neural representations of meaning underlying different languages are shared across speakers of various languages, and that LMs trained on different languages converge on this shared meaning.
30.06.2025 20:56 — 👍 1 🔁 0 💬 1 📌 0
We then tested the extent to which each of these 58 languages can predict the brain activity of our participants. We found that languages that are more similar to the listener’s native language, the better the prediction:
30.06.2025 20:56 — 👍 1 🔁 0 💬 1 📌 0
What about multilingual models? We translated the story from English to 57 other languages spanning 14 families, and extracted embeddings for each from multilingual BERT. We visualized the dissimilarity matrix using MDS and found clusters corresponding to language family types.
30.06.2025 20:56 — 👍 1 🔁 0 💬 1 📌 0
We found that models trained to predict neural activity for one language generalize to different subjects listening to the same content in a different language, across high-level language and default-mode regions.
30.06.2025 20:56 — 👍 1 🔁 0 💬 1 📌 0
We then used the encoding models trained on one language to predict the neural activity in listeners of other languages.
30.06.2025 20:56 — 👍 0 🔁 0 💬 1 📌 0
We then aimed to find if a similar shared space exists in the brains of native speakers of the three different languages. We used voxelwise encoding models that align the LM embeddings with brain activity from one group of subjects listening to the story in their native language.
30.06.2025 20:56 — 👍 0 🔁 0 💬 1 📌 0
We extracted embeddings from three unilingual BERT models—trained on entirely different languages)—and found that (with a rotation) they converge onto similar embeddings, especially in the middle layers:
30.06.2025 20:56 — 👍 0 🔁 0 💬 1 📌 0
We used naturalistic fMRI and language models (LMs) to identify neural representations of the shared conceptual meaning of the same story as heard by native speakers of three languages: English, Chinese, and French.
30.06.2025 20:56 — 👍 0 🔁 0 💬 1 📌 0
We hypothesized that the brains of native speakers of different languages would converge on the same supra-linguistic conceptual structures when listening to the same story in their respective languages:
30.06.2025 20:56 — 👍 0 🔁 0 💬 1 📌 0
Previous research has found that language models trained on different languages learn embedding spaces with similar geometry. This suggests that internal geometry of different languages may converge on similar conceptual structures:
30.06.2025 20:56 — 👍 0 🔁 0 💬 1 📌 0
There are over 7,000 human languages in the world and they're remarkably diverse in their forms and patterns. Nonetheless, we often use different languages to convey similar ideas, and we can learn to translate from one language to another.
30.06.2025 20:56 — 👍 0 🔁 0 💬 1 📌 0
How do different languages converge on a shared neural substrate for conceptual meaning? Happy to share a new preprint led by Zaid Zada that specifically addresses this question:
30.06.2025 20:56 — 👍 34 🔁 5 💬 1 📌 0
We're very excited to share this work and happy to hear your feedback! If you're attending #OHBM2025, @ahmadsamara.bsky.social will be presenting a poster on this project (poster #804, June 25 and 26)—be sure to stop by and chat with him about it!
www.biorxiv.org/content/10.1...
24.06.2025 23:25 — 👍 1 🔁 1 💬 0 📌 0
Our findings suggest that different language areas are coupled via a mixture of linguistic features—this yields what we refer to as a "soft hierarchy" from lower-order to higher-order language areas, and may facilitate efficient, context-sensitive language processing.
24.06.2025 23:25 — 👍 0 🔁 0 💬 1 📌 0
Taking a slightly different approach, we assess how well specific model features capture larger-scale patterns of connectivity. We find that feature-specific model connectivity partly recapitulates stimulus-driven cortical network configuration.
24.06.2025 23:25 — 👍 0 🔁 0 💬 1 📌 0
We observe a clear progression of feature-specific connectivity from early auditory to lateral temporal areas, advancing from acoustic-driven connectivity to speech- and finally language-driven connectivity.
24.06.2025 23:25 — 👍 0 🔁 0 💬 1 📌 0
We show that early auditory areas are coupled to intermediate language areas via lower-level acoustic and speech features. In contrast, higher-order language and default-mode regions are predominantly coupled through more abstract language features.
24.06.2025 23:25 — 👍 0 🔁 0 💬 1 📌 0
We developed a model-based framework for quantifying stimulus-driven, feature-specific connectivity between regions. We used parcel-wise encoding models to align feature-specific embeddings to brain activity and then evaluated how well these models generalize to other parcels.
24.06.2025 23:25 — 👍 0 🔁 0 💬 1 📌 0
Put differently, ISFC can tell us *where* and *how much* connectivity is driven by the stimulus, but not *what* stimulus features are driving the connectivity. How can we begin to unravel what linguistic features are shared across different language regions?
24.06.2025 23:25 — 👍 0 🔁 0 💬 1 📌 0
Following the logic of intersubject correlation (ISC) analysis, intersubject functional connectivity (ISFC) isolates stimulus-driven connectivity between regions (e.g., in response to naturalistic stimuli)—but is agnostic to the content of the stimulus shared between regions.
24.06.2025 23:25 — 👍 0 🔁 0 💬 1 📌 0
Traditional within-subject functional connectivity metrics cannot distinguish between intrinsic and extrinsic (i.e., stimulus-driven) co-fluctuations between brain regions.
24.06.2025 23:25 — 👍 0 🔁 0 💬 1 📌 0
(Sometimes Tucović)
Comics:
The 7 Deadly Sins/First Knife/Slavic Nihilism
Games:
Cantata/Dread Delusion
bio.site/ohotnig
OR
patreon.com/ohotnig
ohotnig.itch.io / ohotnig.gumroad.com
inprnt.com/gallery/ohotnig
buymeacoffee.com/ohotnig
Associate professor of cognitive psychology at Penn State University. Director of the Cognitive Neuroscience of Creativity Lab (beatylab.la.psu.edu). President-elect of the Society for the Neuroscience of Creativity (SfNC; tsfnc.org; @tsfnc.bsky.social).
she/her
Cognitive Neuroscientist and Assistant Prof @ UCRiverside (studying memory and the human brain)
Dog mom
Eating enthusiast
Interested in understanding the neural mechanisms underlying music cognition via neuroimaging and computational modeling
Neuroscience PhD from Princeton
Neuroscience Postdoc at MIT
https://scholar.google.co.uk/citations?user=RjmK2NgAAAAJ&hl=en
just here to have fun online
Curious about memory, spontaneous thought, brains, and stories ⦿ Prof at York University, Glendon Campus ⦿ PI of the Memory & Meaning Lab (www.bellanalab.com)
Kempner Institute research fellow @Harvard interested in scaling up (deep) reinforcement learning theories of human cognition
prev: deepmind, umich, msr
https://cogscikid.com/
Malicious animal magnetism
https://www.twitch.tv/shaun_vids
https://www.youtube.com/@Shaun_vids
https://www.patreon.com/shaunfromyoutube
Cognitive scientist at Barnard College; visual categorization; EEG; eye movements; machine learning; childless cat fae; will ask to see a picture of your pet. Opinions my own. They/she. 🏳️🌈
Assemblymember. Democratic Nominee for Mayor of NYC. Running to freeze the rent, make buses fast + free, and deliver universal childcare. Democratic Socialist. zohranfornyc.com
Postdoc at Penn, computational models of perception and decision making, illusions, cognitive biases, he/his/him
Only L I ever took is my name.
PhD student at NYU working with @s-michelmann.bsky.social | Interested in episodic memory, schemas and imagination.
Computational neuroscientist/biophysicist interested in memory, dynamics, and spikes.
https://rkp.science
A student of systems, computational and cognitive neuroscience.
Cognitive neuroscientist studying memory, imagery, and cognitive control; PI @ Institute of Neuroscience, Chinese Academy of Sciences; was @ UW-Madison, Dartmouth, Peking University.
https://yuqingpsy.github.io/