[8/n] More details in the paper: arxiv.org/abs/2411.06090 and our blog post ncfrey.substack.com/p/building-t....
You can also find the code on GitHub: github.com/prescient-de... and model weights on ๐ค.
[8/n] More details in the paper: arxiv.org/abs/2411.06090 and our blog post ncfrey.substack.com/p/building-t....
You can also find the code on GitHub: github.com/prescient-de... and model weights on ๐ค.
[7/n] Our architecture allows us to know what concepts the model learned and which concepts the model uses during inference by inspecting the weights of the final linear layer; this offers a way to debug and asses the model's quality.
12.12.2024 22:50 โ ๐ 2 ๐ 0 ๐ฌ 1 ๐ 0[6/n] Interpretability: The concept bottleneck can be used to understand which concept the model uses to predict a certain amino acid. Reliably controlling model behavior: The concepts can be used as knobs to control the model's output.
12.12.2024 22:50 โ ๐ 1 ๐ 0 ๐ฌ 1 ๐ 0[5/n] We train a mask language model with up to 3 Billion parameters with a layer that directly encodes biophysical and biochemical concepts that biologists care about. These models match the performance of unconstrained masked language model.
12.12.2024 22:50 โ ๐ 3 ๐ 0 ๐ฌ 1 ๐ 0[4/n] Joint work with @ncfrey.bsky.social Tuomas Oikarinen @amywang1.bsky.social @juliusad.bsky.social @activelearner.bsky.social Taylor Joren @allen.bsky.social Hector Corrada Bravo @kyunghyuncho.bsky.social and Joseph Kleinhenz in @prescientdesign.bsky.social
12.12.2024 22:50 โ ๐ 4 ๐ 0 ๐ฌ 1 ๐ 0[3/n] In our concept bottleneck protein language model paper, we show that we can train the model with billions of parameters, with interpretability constraints, without performance degradation.
12.12.2024 22:50 โ ๐ 3 ๐ 1 ๐ฌ 1 ๐ 0[2/n] But the thing is, more often than not, we know beforehand what we want/expect our model to learn, especially in very well-studied domains like Biology. So, instead of playing the guessing game, we trained a model that explicitly learns different concepts that biologists care about.
12.12.2024 22:50 โ ๐ 2 ๐ 0 ๐ฌ 1 ๐ 0
๐งต
[1/n] Does AlphaFold3 "know" biophysics and the physics of protein folding? Are protein language models (pLMs) learning coevolutionary patterns? You can try to guess the answer to these questions using mechanistic interpretability.