Nassau Hall. Photo credit to Debbie and John O'Boyle
I'm joining Princeton University as an Associate Professor of Computer Science and Psychology this fall! Princeton is ambitiously investing in AI and Natural & Artificial Minds, and I'm excited for my lab to contribute. Recruiting postdocs and Ph.D. students in CS and Psychology — join us!
12.06.2025 14:29 — 👍 46 🔁 2 💬 4 📌 0
Are transformers smarter than you? Hypernetworks might explain why.
Come checkout our Oral at #ICLR tomorrow (Apr 26th, poster at 10:00, Oral session 6C in the afternoon).
openreview.net/forum?id=V4K...
25.04.2025 04:50 — 👍 8 🔁 0 💬 1 📌 0
Pre-print 🧠🧪
Is mechanism modeling dead in the AI era?
ML models trained to predict neural activity fail to generalize to unseen opto perturbations. But mechanism modeling can solve that.
We say "perturbation testing" is the right way to evaluate mechanisms in data-constrained models
1/8
08.01.2025 16:33 — 👍 115 🔁 46 💬 4 📌 2
2024: A Review of the Year in Neuroscience
Feeling a bit wired
Cutting it a bit fine, but here’s my review of the year in neuroscience for 2024
The eighth of these, would you believe? We’ve got dark neurons, tiny monkeys, the most complete brain wiring diagram ever constructed, and much more…
Published on The Spike
Enjoy!
medium.com/the-spike/20...
30.12.2024 16:00 — 👍 191 🔁 74 💬 7 📌 18
An introduction to reinforcement learning for neuroscience | Published in Neurons, Behavior, Data analysis, and Theory
By Kristopher T. Jensen. Reinforcement learning for neuroscientists
I wrote an introduction to RL for neuroscience last year that was just published in NBDT: tinyurl.com/5f58zdy3
This review aims to provide some intuition for and derivations of RL methods commonly used in systems neuroscience, ranging from TD learning through the SR to deep and distributional RL!
21.12.2024 17:59 — 👍 130 🔁 32 💬 6 📌 0
Monster Models
Systems-level biology is hard because systems-level engineering is hard.
Stitching component models into system models has proven difficult in biology. But how much easier has it been in engineering? www.argmin.net/p/monster-mo...
20.12.2024 15:29 — 👍 12 🔁 2 💬 3 📌 1
🚨 New Paper!
Can neuroscience localizers uncover brain-like functional specializations in LLMs? 🧠🤖
Yes! We analyzed 18 LLMs and found units mirroring the brain's language, theory of mind, and multiple demand networks!
w/ @gretatuckute.bsky.social, @abosselut.bsky.social, @mschrimpf.bsky.social
🧵👇
19.12.2024 15:06 — 👍 105 🔁 27 💬 2 📌 5
1/ Okay, one thing that has been revealed to me from the replies to this is that many people don't know (or refuse to recognize) the following fact:
The unts in ANN are actually not a terrible approximation of how real neurons work!
A tiny 🧵.
🧠📈 #NeuroAI #MLSky
16.12.2024 20:03 — 👍 153 🔁 39 💬 21 📌 17
For my first post on Bluesky .. I'll start by announcing our 2025 edition of EEML which will be in Sarajevo :) ! I'm really excited about it and hope to see many of you there. Please follow the website (and Bluesky account) for more details which are coming soon ..
15.12.2024 18:39 — 👍 31 🔁 7 💬 1 📌 0
Have you had private doubts whether we'll ever understand the brain? Whether we'll be able explain psychological phenomena in an exhaustive way that ranges from molecules to membranes to synapses to cells to cell types to circuits to computation to perception and behavior?
14.11.2024 05:18 — 👍 39 🔁 12 💬 1 📌 1
The broader spectrum of in-context learning
The ability of language models to learn a task from a few examples in context has generated substantial interest. Here, we provide a perspective that situates this type of supervised few-shot learning...
What counts as in-context learning (ICL)? Typically, you might think of it as learning a task from a few examples. However, we’ve just written a perspective (arxiv.org/abs/2412.03782) suggesting interpreting a much broader spectrum of behaviors as ICL! Quick summary thread: 1/7
10.12.2024 18:17 — 👍 123 🔁 31 💬 2 📌 1
Thrilled to share our NeurIPS Spotlight paper with Jan Bauer*, @aproca.bsky.social*, @saxelab.bsky.social, @summerfieldlab.bsky.social, Ali Hummos*! openreview.net/pdf?id=AbTpJ...
We study how task abstractions emerge in gated linear networks and how they support cognitive flexibility.
03.12.2024 16:04 — 👍 64 🔁 15 💬 2 📌 1
Would love to be added as well :)
20.11.2024 20:50 — 👍 0 🔁 0 💬 0 📌 0
Great thread from @michaelhendricks.bsky.social!
Reminds me of something Larry Abbott once said to me at a summer school:
Many physicists come into neuroscience assuming that the failure to find laws of the brain was just because biologists aren't clever enough. In fact, there are no laws.
🧠📈 🧪
13.11.2024 18:49 — 👍 69 🔁 9 💬 4 📌 1
(1/5) Very excited to announce the publication of Bayesian Models of Cognition: Reverse Engineering the Mind. More than a decade in the making, it's a big (600+ pages) beautiful book covering both the basics and recent work: mitpress.mit.edu/978026204941...
18.11.2024 16:25 — 👍 522 🔁 120 💬 15 📌 16
🙋♂️
16.11.2024 15:30 — 👍 1 🔁 0 💬 0 📌 0
To help find people at the intersection of neuroscience and AI. Of course let me know if I missed someone or you’d like to be added 🧪 🧠
#neuroskyence
go.bsky.app/CAfmKQs
13.11.2024 15:26 — 👍 50 🔁 18 💬 33 📌 0
I think you are already part of it - just double checked :)
13.11.2024 15:20 — 👍 1 🔁 0 💬 0 📌 0
Indeed in line with the hypothesis that the hypernetwork mechanism supports compositionality, this modification (hyla) improves performance on unseen tasks.
28.10.2024 15:26 — 👍 0 🔁 0 💬 1 📌 0
So what happens if we strengthen the hypernetwork mechanism?
Could we maybe further improve compositionality?
We can for instance make the value network nonlinear - without introducing additional parameters.
28.10.2024 15:26 — 👍 0 🔁 0 💬 1 📌 0
Training a simple decoder on the latent codes of training tasks allows us to predict the operations performed by the network on unseen tasks - especially for later layers.
28.10.2024 15:25 — 👍 0 🔁 0 💬 1 📌 0
To test this hypothesis we train small transformer models to solve abstract reasoning tasks. When we look at the latent code of tasks they have never seen before, we find a highly structured space.
28.10.2024 15:25 — 👍 0 🔁 0 💬 1 📌 0
From the hypernetwork perspective, a compact latent code specifies key-query specific operations.
Importantly, these operations are reusable: the same hypernetwork is used across all key-query pairs.
Could their reuse allow transformers to compositionally generalize?
28.10.2024 15:24 — 👍 0 🔁 0 💬 1 📌 0
For a given query, multi-head attention can be rewritten as a sum over the outputs of key-query specific value networks configured by a hypernetwork.
These hypernetworks are comparably simple: Both the hypernetwork and its value network are linear.
So why could this matter?
28.10.2024 15:23 — 👍 0 🔁 0 💬 1 📌 0
We know that hypernetworks - neural networks that generate the weights of another neural network - can compositionally generalize. So, should we build more hypernetworks into our transformers?
It turns out that attention with multiple heads already has them built-in!
28.10.2024 15:22 — 👍 0 🔁 0 💬 1 📌 0
PhD in AI @mila-quebec.bsky.social RLHF and language grounding, whatever that means. Whitespace aficianado. mnoukhov.github.io
Studying cognition in humans and machines https://scholar.google.com/citations?user=WCmrJoQAAAAJ&hl=en
Neuroscientist studying social behavior and decision making @ETH Zürich @Grewe lab
Outside of the lab, I love doing nature-inspired artistic projects! My most recent project: https://hellodino.ch 🦕🦕
Cognitive Scientist at Max Planck, Professor of Psychology
https://www.falkhuettig.com/
Author of 'Looking Ahead: The New Science of the Predictive Mind' published by Cambridge University Press on 6 March 2025.
prev: @BrownUniversity, @uwcse/@uw_wail phd, ex-@cruise, RS @waymo. 0.1x engineer, 10x friend.
spondyloarthritis, cars ruin cities, open source
European Research Council, set up by the EU, funds top researchers of any nationality, helping them pursue great ideas at the frontiers of knowledge. #HorizonEU
Academic nomad | Wandering neuroscientist | Climate worrier |
Assistant Professor @ Leiden University, views very obviously mine | https://anneurai.net/
Sakana AI is an AI R&D company based in Tokyo, Japan. 🗼🧠
https://sakana.ai/careers
Computational neuroscientist @princetonneuro.bsky.social deciphering natural and advancing artificial intelligence.
Neuroscientist at The Sainsbury Wellcome Centre, UCL, in London. Leading the "Learning, Inference & Memory" laboratory. Accidental advocate of #longcovid
https://www.lim.bio/
Google Chief Scientist, Gemini Lead. Opinions stated here are my own, not those of Google. Gemini, TensorFlow, MapReduce, Bigtable, Spanner, ML things, ...
Postdoctoral researcher in Computational Neuroscience @Stanford
Building Astral: Ruff, uv, and other high-performance Python tools, written in Rust.
Natural and artificial general intelligence.
https://marcelbinz.github.io/
@Cohere.com's non-profit research lab and open science initiative that seeks to solve complex machine learning problems. Join us in exploring the unknown, together. https://cohere.com/research
Professor at JLU | cognitive computational neuroscientist | mom of 3 | she/her
hessian.AI conducts cutting-edge AI research, provides computing infrastructure & services, supports start-up projects, ensures the transfer to business and society and thus strengthens the AI ecosystem in Hesse & beyond. https://hessian.ai/legal-notice