Bonnie Berger Lab's Avatar

Bonnie Berger Lab

@bergerlab.bsky.social

The Berger lab at @csail.mit.edu works on a diverse set of problems in computational biology and biomedicine. Account run by lab members. https://people.csail.mit.edu/bab/

71 Followers  |  6 Following  |  10 Posts  |  Joined: 11.03.2025  |  1.526

Latest posts by bergerlab.bsky.social on Bluesky

Preview
Sparse autoencoders uncover biologically interpretable features in protein language model representations | PNAS Foundation models in biology—particularly protein language models (PLMs)—have enabled ground-breaking predictions in protein structure, function, a...

8/ Paper: pnas.org/doi/10.1073/...
MIT News article: news.mit.edu/2025/researc...
Github:

03.09.2025 16:51 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

7/ We show that SAE & transcoder features are much more interpretable than ESM neurons, for both protein-level & amino acid-level representations. This has the potential to improve safety, trust & explainability of PLMs. As PLMs improve, SAEs could help us learn new biology.

03.09.2025 16:51 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

6/ We also use Claude to autointerpret SAE features based on protein names, families, gene names & GO terms. Many features correspond to families (like NAD Kinase, IUNH, PTH) & functions (like methyltransferase activity, olfactory/gustatory perception).

03.09.2025 16:51 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

5/ We interpret these SAE features using Gene Ontology (GO) enrichment. Many protein-level SAE features align tightly with GO terms across all levels of the GO hierarchy.

03.09.2025 16:51 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

4/ SAEs have a very wide latent dimension with a sparsity constraint. This forces PLM representations to disentangle into biologically interpretable, sparsely activating features without any supervision.

03.09.2025 16:50 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

3/ We train sparse autoencoders (SAEs) on protein-level and amino acid-level representations from layers 6-10 of ESM2_t12_35M_UR50D. We also train transcoders (an SAE variant) on protein-level representations.

03.09.2025 16:50 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

2/ Protein-level representations from PLMs are used in many downstream tasks. Disentangling their features can enhance interpretability, helping us trust and explain downstream applications.

03.09.2025 16:50 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

1/ PLMs like ESM have made big strides in predicting protein structure & function. But they feel like a β€œblack-box.” What biological information do PLM representations contain? Can we disentangle them systematically?

03.09.2025 16:50 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Post image

Excited to share our recent work: Sparse autoencoders uncover biologically interpretable features in protein language model representations now in PNAS. Thread below 🧡

03.09.2025 16:50 β€” πŸ‘ 7    πŸ” 2    πŸ’¬ 1    πŸ“Œ 0

Hello world! See the thread below for our recent work 🌿 MINT!

12.03.2025 18:34 β€” πŸ‘ 4    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

@bergerlab is following 6 prominent accounts