David Amadeus  Vogelsang's Avatar

David Amadeus Vogelsang

@davogelsang.bsky.social

Lecturer in Brain & Cognition at the University of Amsterdam

49 Followers  |  142 Following  |  12 Posts  |  Joined: 14.09.2025  |  1.6935

Latest posts by davogelsang.bsky.social on Bluesky

Preview
Language models cannot reliably distinguish belief from knowledge and fact - Nature Machine Intelligence Suzgun et al. find that current large language models cannot reliably distinguish between belief, knowledge and fact, raising concerns for their use in healthcare, law and journalism, where such disti...

Do we need a Nature paper for that?

Language models cannot reliably distinguish belief from knowledge and fact
www.nature.com/articles/s42...

06.11.2025 14:42 β€” πŸ‘ 84    πŸ” 20    πŸ’¬ 11    πŸ“Œ 2
Post image

Introducing CorText: a framework that fuses brain data directly into a large language model, allowing for interactive neural readout using natural language.

tl;dr: you can now chat with a brain scan πŸ§ πŸ’¬

1/n

03.11.2025 15:17 β€” πŸ‘ 129    πŸ” 52    πŸ’¬ 4    πŸ“Œ 8
Preview
BBC Radio 4 - All in the Mind, Are there multiple subtypes of autism, and how vivid are your memories? A new study suggests that autism’s genetic profile differs with age at diagnosis.

The brilliant @kasiamojescik.bsky.social and Martha McGill join @claudiahammond.bsky.social and @catherineloveday.bsky.social on BBC Radio 4 All in the Mind this morning to launch our new public survey of vivid memories. You can take part here: cambridge.eu.qualtrics.com/jfe/form/SV_...

21.10.2025 08:24 β€” πŸ‘ 14    πŸ” 11    πŸ’¬ 0    πŸ“Œ 1
Preview
Dense Phenotyping of Human Brain Network Organization Using Precision fMRI The advent of noninvasive imaging methods like functional magnetic resonance imaging (fMRI) transformed cognitive neuroscience, providing insights into large-scale brain networks and their link to cog...

Why do brain networks vary? Do these differences shape behavior? If every 🧠 is unique, how can we detect common features of brain organization?
@rodbraga.bsky.social and I dig in, in @annualreviews.bsky.social (ahead of print):
go.illinois.edu/Gratton2025-...

#neuroskyence #psychscisky #MedSky
πŸ§΅πŸ‘‡

16.10.2025 15:00 β€” πŸ‘ 83    πŸ” 46    πŸ’¬ 1    πŸ“Œ 3
Universiteit Leiden - Atopia |

Vacancy alert! Prof. Ineke van der Ham and I are looking for a post-doc (from early 2026 on). For this project we aim to investigate why some individuals often get lost: a condition recently coined atopia. The project involves behavioral, eye tracking and potentially fMRI. See dewegkwijt.com

16.10.2025 09:38 β€” πŸ‘ 1    πŸ” 1    πŸ’¬ 0    πŸ“Œ 0
Post image

Elke schooldag staat leren centraal. Gek eigenlijk dat er geen vakken zijn die je leren hoe je dit het beste aanpakt. In een nieuwe Klokhuis-aflevering geven Erik Scherder (vanaf een sportveld) en ik (vanuit onze Sylvius VR labs) tips om je te leren leren: hetklokhuis.nl/tv-uitzendin...

16.10.2025 07:10 β€” πŸ‘ 0    πŸ” 1    πŸ’¬ 0    πŸ“Œ 0
Preview
Principles for proper peer review

For all the knucklehead reviewers out there.
Principles for proper peer review - Earl K. Miller
jocnf.pubpub.org/pub/qag76ip8...
#neuroscience

06.10.2025 19:59 β€” πŸ‘ 69    πŸ” 24    πŸ’¬ 1    πŸ“Œ 4
Preview
Confronting the connectivity crisis in human M/EEG research The cognitive neuroscience community using M/EEG has not converged on measures of task-related inter-regional brain connectivity that generalize acros…

In our Trends in Cogn Sci paper we point to the connectivity crisis in task-based human EEG/MEG research: many connectivity metrics, too little replication. Time for community-wide benchmarking to build robust, generalisable measures across labs & tasks. www.sciencedirect.com/science/arti...

18.09.2025 15:23 β€” πŸ‘ 87    πŸ” 28    πŸ’¬ 2    πŸ“Œ 0

Thank you; and that is an interesting question. My prediction is that it may not work so well (would be fun to test)

18.09.2025 15:56 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

Thank you for your reply. Unfortunately, we did not examine within-category effects, but that would certainly be interesting to do

18.09.2025 15:51 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

Our takeaway:
Memory has a geometry.
The magnitude of representations predicts memorability across vision and language, providing a new lens for understanding why some stimuli are memorable.

18.09.2025 10:00 β€” πŸ‘ 2    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

Think of memory as geometry:
An item’s vector length in representational space predicts how likely it is to stick in your mind β€” at least for images and words.

18.09.2025 10:00 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

So what did we learn?
βœ… Robust effect for images
βœ… Robust effect for words
❌ No effect for voices
β†’ Memorability seems tied to how strongly items project onto meaningful representational dimensions, not all sensory domains.

18.09.2025 09:59 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Post image

Then we asked: does this principle also apply to voices?
Using a recent dataset with >600 voice clips, we tested whether wav2vec embeddings showed the same effect.
πŸ‘‰ They didn’t. No consistent link between L2 norm and voice memorability.

18.09.2025 09:59 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Post image

And crucially:
This effect held even after controlling for word frequency, valence, and size.
So representational magnitude is not just a proxy for familiar or emotionally loaded words.

18.09.2025 09:58 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Post image

Then we asked: is this just a visual trick, or is it present in other domains as well?
When we turned to words, the result was striking:
Across 3 big datasets, words with higher vector magnitude in embeddings were consistently more memorable, revealing the same L2 norm principle

18.09.2025 09:58 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

In CNNs, the effect is strongest in later layers, where abstract, conceptual features are represented.
πŸ“Š Larger representational magnitude β†’ higher memorability.

18.09.2025 09:56 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Post image

We first wanted to examine whether we could replicate this L2 norm effect as reported by Jaegle et al. (2019).
Using the massive THINGS dataset (>26k images, 13k participants), we replicated that the L2 norm of CNN representations predicts image memorability.

18.09.2025 09:56 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

Why do we remember some things better than others?
Memory varies across people, but some items are intrinsically more memorable.
Jeagle et al. (2019) showed that a simple geometric property of representations β€” the L2 norm (vector magnitude) β€” positively correlates with image memorability

18.09.2025 09:54 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Preview
Representational magnitude as a geometric signature of image and word memorability What makes some stimuli more memorable than others? While memory varies across individuals, research shows that some items are intrinsically more memorable, a property quantifiable as β€œmemorability”. ...

New preprint out together with @mheilbron.bsky.social

We find that a stimulus' representational magnitudeβ€”the L2 norm of its DNN representationβ€”predicts intrinsic memorability not just for images, but for words too.
www.biorxiv.org/content/10.1...

18.09.2025 09:53 β€” πŸ‘ 23    πŸ” 6    πŸ’¬ 4    πŸ“Œ 1
Video thumbnail

Interested in hippocampal dynamics and their interactions with cortical rhythms?

Our physically constrained model of cortico-hippocampal interactions - complete with fast geometrically informed numerical simulation (available at embedded github repo)

www.biorxiv.org/content/10.1...

14.09.2025 11:35 β€” πŸ‘ 55    πŸ” 22    πŸ’¬ 0    πŸ“Œ 1

@davogelsang is following 20 prominent accounts