Thomas Fel's Avatar

Thomas Fel

@thomasfel.bsky.social

Explainability, Computer Vision, Neuro-AI.πŸͺ΄ Kempner Fellow @Harvard. Prev. PhD @Brown, @Google, @GoPro. CrΓͺpe lover. πŸ“ Boston | πŸ”— thomasfel.me

1,321 Followers  |  332 Following  |  12 Posts  |  Joined: 16.11.2024  |  1.6004

Latest posts by thomasfel.bsky.social on Bluesky

Home First Workshop on Interpreting Cognition in Deep Learning Models (NeurIPS 2025)

Excited to announce the first workshop on CogInterp: Interpreting Cognition in Deep Learning Models @ NeurIPS 2025! πŸ“£

How can we interpret the algorithms and representations underlying complex behavior in deep learning models?

🌐 coginterp.github.io/neurips2025/

1/4

16.07.2025 13:08 β€” πŸ‘ 52    πŸ” 18    πŸ’¬ 1    πŸ“Œ 1

How do language models generalize from information they learn in-context vs. via finetuning? In arxiv.org/abs/2505.00661 we show that in-context learning can generalize more flexibly, illustrating key differences in the inductive biases of these modes of learning β€” and ways to improve finetuning. 1/

02.05.2025 17:02 β€” πŸ‘ 78    πŸ” 21    πŸ’¬ 4    πŸ“Œ 4

Our work finding universal concepts in vision models is accepted at #ICML2025!!!

My first major conference paper with my wonderful collaborators and friends @matthewkowal.bsky.social @thomasfel.bsky.social
@Julian_Forsyth
@csprofkgd.bsky.social

Working with y'all is the best πŸ₯Ή

Preprint ⬇️!!

01.05.2025 22:57 β€” πŸ‘ 15    πŸ” 4    πŸ’¬ 0    πŸ“Œ 1

Accepted at #ICML2025! Check out the preprint.

HUGE shoutout to Harry (1st PhD paper, in 1st year), Julian (1st ever, done as an undergrad), Thomas and Matt!

@hthasarathan.bsky.social @thomasfel.bsky.social @matthewkowal.bsky.social

01.05.2025 15:03 β€” πŸ‘ 35    πŸ” 7    πŸ’¬ 2    πŸ“Œ 0
Video thumbnail

<proud advisor>
Hot off the arXiv! 🦬 "Appa: Bending Weather Dynamics with Latent Diffusion Models for Global Data Assimilation" 🌍 Appa is our novel 1.5B-parameter probabilistic weather model that unifies reanalysis, filtering, and forecasting in a single framework. A thread 🧡

29.04.2025 04:48 β€” πŸ‘ 49    πŸ” 15    πŸ’¬ 2    πŸ“Œ 3
Post image

Have you thought that in computer memory model weights are given in terms of discrete values in any case. Thus, why not do probabilistic inference on the discrete (quantized) parameters. @trappmartin.bsky.social is presenting our work at #AABI2025 today. [1/3]

29.04.2025 06:58 β€” πŸ‘ 46    πŸ” 11    πŸ’¬ 3    πŸ“Œ 1
Preview
Interpreting the Linear Structure of Vision-Language Model Embedding Spaces - Kempner Institute Using sparse autoencoders, the authors show that vision-language embeddings boil down to a small, stable dictionary of single-modality concepts that snap together into cross-modal bridges. This resear...

New in the Deeper Learning blog: Kempner researchers show how VLMs speak the same semantic language across images and text.

bit.ly/KempnerVLM

by @isabelpapad.bsky.social ,Chloe Huangyuan Su, @thomasfel.bsky.social, Stephanie Gil, and @shamkakade.bsky.social

#AI #ML #VLMs #SAEs

28.04.2025 16:57 β€” πŸ‘ 9    πŸ” 3    πŸ’¬ 0    πŸ“Œ 0
Preview
Firing rates in visual cortex show representational drift, while temporal spike sequences remain stable Neural firing-rate responses to sensory stimuli show progressive changes both within and across sessions, raising the question of how the brain mainta…

Firing rates in visual cortex show representational drift, while temporal spike sequences remain stable

www.sciencedirect.com/science/arti...

Great work by Boris Sotomayor and with @battaglialab.bsky.social

10.04.2025 10:44 β€” πŸ‘ 70    πŸ” 20    πŸ’¬ 0    πŸ“Œ 1
APA PsycNet

PINEAPPLE, LIGHT, HAPPY, AVALANCHE, BURDEN

Some of these words are consistently remembered better than others. Why is that?
In our paper, just published in J. Exp. Psychol., we provide a simple Bayesian account and show that it explains >80% of variance in word memorability: tinyurl.com/yf3md5aj

10.04.2025 14:38 β€” πŸ‘ 40    πŸ” 15    πŸ’¬ 1    πŸ“Œ 0

πŸ“½οΈRecordings from our
@cosynemeeting.bsky.social
#COSYNE2025 workshop on β€œAgent-Based Models in Neuroscience: Complex Planning, Embodiment, and Beyond" are now online: neuro-agent-models.github.io
πŸ§ πŸ€–

07.04.2025 20:57 β€” πŸ‘ 36    πŸ” 11    πŸ’¬ 1    πŸ“Œ 0

[...] overall, we argue an SAE does not just reveal conceptsβ€”it determines what can be seen at all."

We propose to examine how constraints on SAE impose dual assumptions on the data, led by the amazing
@sumedh-hindupur.bsky.social 😎

07.03.2025 03:27 β€” πŸ‘ 6    πŸ” 1    πŸ’¬ 0    πŸ“Œ 0
Post image

New paper–accepted as *spotlight* at #ICLR2025! πŸ§΅πŸ‘‡

We show a competition dynamic between several algorithms splits a toy model’s ICL abilities into four broad phases of train/test settings! This means ICL is akin to a mixture of different algorithms, not a monolithic ability.

16.02.2025 18:57 β€” πŸ‘ 32    πŸ” 5    πŸ’¬ 2    πŸ“Œ 1
Post image

Want strong SSL, but not the complexity of DINOv2?

CAPI: Cluster and Predict Latents Patches for Improved Masked Image Modeling.

14.02.2025 18:04 β€” πŸ‘ 49    πŸ” 10    πŸ’¬ 1    πŸ“Œ 1
Post image

🚨 New Paper!

Can neuroscience localizers uncover brain-like functional specializations in LLMs? πŸ§ πŸ€–

Yes! We analyzed 18 LLMs and found units mirroring the brain's language, theory of mind, and multiple demand networks!

w/ @gretatuckute.bsky.social, @abosselut.bsky.social, @mschrimpf.bsky.social
πŸ§΅πŸ‘‡

19.12.2024 15:06 β€” πŸ‘ 105    πŸ” 27    πŸ’¬ 2    πŸ“Œ 5
Post image Post image Post image Post image

Scaling Laws in Patchification: An Image Is Worth 50,176 Tokens And More

Feng Wang, Yaodong Yu, Guoyizhe Wei, Wei Shao, Yuyin Zhou, Alan Yuille, Cihang Xie

tl;dr: we trained 1px patch size ViT so you don't have to. It improves results, but costly.

arxiv.org/abs/2502.03738

08.02.2025 17:10 β€” πŸ‘ 19    πŸ” 1    πŸ’¬ 1    πŸ“Œ 0

I'm delighted to share this latest research, led by the talented @hthasarathan.bsky.social
and Julian. Their work uncovered both universal conceptual across models but also unique concepts specific to DINOv2 and SigLip! πŸ”₯

07.02.2025 23:39 β€” πŸ‘ 4    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Post image

πŸŒŒπŸ›°οΈπŸ”­Wanna know which features are universal vs unique in your models and how to find them? Excited to share our preprint: "Universal Sparse Autoencoders: Interpretable Cross-Model Concept Alignment"!

arxiv.org/abs/2502.03714

(1/9)

07.02.2025 15:15 β€” πŸ‘ 56    πŸ” 17    πŸ’¬ 1    πŸ“Œ 5
Post image

Using mechanistic #interpretability πŸ’» to advance scientific #discovery πŸ§ͺ & capture striking biology? 🧬
Come see @jhartford.bsky.social's oral presentation πŸ‘¨β€πŸ« @ #NeurIPS2024 Interpretable AI workshop 🦾 to learn more about extracting features from large πŸ”¬ MAEs! Paper πŸ“„ ➑️: openreview.net/forum?id=jYl...

15.12.2024 14:30 β€” πŸ‘ 22    πŸ” 2    πŸ’¬ 1    πŸ“Œ 0
Post image

Did you know that @PyTorch implements the Bessel's correction to standard deviation, but not numpy or jax.

A possible source of disagreements when porting models to pytorch! @numpy_team

03.02.2025 12:52 β€” πŸ‘ 22    πŸ” 3    πŸ’¬ 1    πŸ“Œ 0
Picture of neural manifolds for the two choices in a decision-making task, depicted in 3D and in 2D

Picture of neural manifolds for the two choices in a decision-making task, depicted in 3D and in 2D

New preprint: "The geometry of the neural state space of decisions", work by Mauro Monsalve-Mercado, buff.ly/42wVHD5. Surprising results & predictions! (Thread) We analyze neuropixel population recordings in macaque area LIP during a reaction time, random-dot motion 1/

31.01.2025 14:32 β€” πŸ‘ 175    πŸ” 67    πŸ’¬ 3    πŸ“Œ 2
Video thumbnail

How do tokens evolve as they are processed by a deep Transformer?

With JosΓ© A. Carrillo, @gabrielpeyre.bsky.social and @pierreablin.bsky.social, we tackle this in our new preprint: A Unified Perspective on the Dynamics of Deep Transformers arxiv.org/abs/2501.18322

ML and PDE lovers, check it out!

31.01.2025 16:56 β€” πŸ‘ 96    πŸ” 16    πŸ’¬ 2    πŸ“Œ 0
Preview
AxBench: Steering LLMs? Even Simple Baselines Outperform Sparse Autoencoders ArXiv link for AxBench: Steering LLMs? Even Simple Baselines Outperform Sparse Autoencoders

The research introduces AXBENCH, demonstrating that prompting outperforms complex representation methods like sparse autoencoders. It features a new weakly-supervised method, ReFT-r1, which combines interpretability with competitive performance. https://arxiv.org/abs/2501.17148

29.01.2025 11:40 β€” πŸ‘ 2    πŸ” 1    πŸ’¬ 0    πŸ“Œ 0
Preview
Graduate Fellowship - Kempner Institute Kempner Graduate Fellowships support a dynamic and diverse community of PhD students across a number of graduate programs at Harvard, seeding new and innovative scientific discoveries in labs across t...

The application for our #KempnerInstitute graduate fellowship for Harvard Ph.D. students is now open! Register for an upcoming virtual open house (2/6 & 2/18) and apply by 3/1. Details: bit.ly/40RY6qS

30.01.2025 16:04 β€” πŸ‘ 5    πŸ” 2    πŸ’¬ 0    πŸ“Œ 0

True haha !!

30.01.2025 15:21 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

Haha could be, but i have another hypothesis, i'll show you next week πŸ˜‰

30.01.2025 15:21 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Post image

DinoV2, C:5232... πŸ˜Άβ€πŸŒ«οΈ

30.01.2025 02:32 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 2    πŸ“Œ 0
Post image

Does the culture you grow up in shape the way you see the world? In a new Psych Review paper, @chazfirestone.bsky.social & I tackle this centuries-old question using the Müller-Lyer illusion as a case study. Come think through one of history's mysteries with us🧡(1/13):

25.01.2025 22:05 β€” πŸ‘ 1095    πŸ” 419    πŸ’¬ 33    πŸ“Œ 79
Preview
ICLR 2025 Workshop XAI4Science Welcome to the OpenReview homepage for ICLR 2025 Workshop XAI4Science

Join us at our upcoming workshop at ICLR, XAI4Science: From Understanding Model Behavior to Discovering New Scientific Knwlg, Apr 27-28

submissions on a-priori (ante-hoc) & a-posteriori (post-hoc) interpretability & self-explainable models for understanding model’s behvr welcm tinyurl.com/3w8sddpm

20.01.2025 17:04 β€” πŸ‘ 33    πŸ” 15    πŸ’¬ 1    πŸ“Œ 0
Post image

The Mathematics of Artificial Intelligence: In this introductory and highly subjective survey, aimed at a general mathematical audience, I showcase some key theoretical concepts underlying recent advancements in machine learning. arxiv.org/abs/2501.10465

22.01.2025 09:11 β€” πŸ‘ 140    πŸ” 40    πŸ’¬ 2    πŸ“Œ 1

@thomasfel is following 20 prominent accounts