Right. We have done something similar in our previous work (cinema-ot) where we validated casual inferences using synthetic data where we know the ground truth.
19.04.2025 14:00 β π 1 π 0 π¬ 0 π 0@vandijklab.bsky.social
Learning the rules of life. Assistant Professor of Medicine and Computer Science @ Yale
Right. We have done something similar in our previous work (cinema-ot) where we validated casual inferences using synthetic data where we know the ground truth.
19.04.2025 14:00 β π 1 π 0 π¬ 0 π 0Right, and I do believe this is possible based on other experiments we have done where we translate between biological language and natural language. Your proposed experiment may be more specific and Iβm interested in trying it.
19.04.2025 13:58 β π 1 π 0 π¬ 0 π 0Zero shot is possible but obviously much harder and also very much depends on the specific system.
19.04.2025 13:54 β π 1 π 0 π¬ 0 π 0We have focussed on fine tuning on one immune cell cytokine stim dataset and on (bulk) L1000. In both cases we show generalization by leaving out conditions (eg cytokine combos).
19.04.2025 13:53 β π 1 π 0 π¬ 0 π 0And the reasoning here is that if they improve then that shows that our model generates meaningful data? Thatβs interesting. Itβs a convenient way of validating without doing experiments I guess
19.04.2025 13:50 β π 1 π 0 π¬ 1 π 0I see. We havenβt done this specific experiment where we compare well studied vs poorly studied genes. Itβs an interesting idea. We will look into it. I would expect that genes/cell types/tissues that have a lot of training data, both expression and meta data, generalize better.
19.04.2025 13:40 β π 1 π 0 π¬ 0 π 0Yes. We showed that natural language pretraining vs training on cell sentences from scratch, significantly boosts performance.
In addition, in the spatial reasoning task (fig.6) we did an ablation where we trained with and without metadata. With metadata performed significantly better.
Finally, asking a model to generate a βcell sentenceβ (e.g. for perturbation response prediction) is novel by design, since no LLM has encountered that representation in its training data.
18.04.2025 17:32 β π 1 π 0 π¬ 0 π 0Second, several test setsβsuch as Dataset Interpretation on held-out studiesβuse scRNA-seq datasets published after each modelβs pretraining cutoff, giving us strong assurance that those examples werenβt seen during training.
18.04.2025 17:32 β π 1 π 0 π¬ 1 π 0We took several steps to ensure robust evaluation. First, we tested both open- and closed-source LLMs (GPT-4o, Gemini, LLaMA-3) on our benchmarks and found they perform poorly out of the box, indicating minimal overlap with pretraining corpora.
18.04.2025 17:32 β π 1 π 0 π¬ 2 π 0For this paper, we chose a prompt structure that helps the model learn perturbations effectively, but initial tests suggest the model handles prompt variations well as long as the data formatting is consistentβso we don't expect prompt engineering to be a major issue.
18.04.2025 17:19 β π 2 π 0 π¬ 0 π 0We'll formally test prompt robustness in future work, but from experience with earlier Cell2Sentence models, we've found minimal performance loss when using new or varied prompts. In general, we always train on a wide variety of prompts to avoid overfitting.
18.04.2025 17:19 β π 1 π 0 π¬ 1 π 0Thank you!
18.04.2025 17:13 β π 1 π 0 π¬ 0 π 0- For dataset interpretation, we evaluate on scRNA-seq studies published after the model was pretrained.
Performance drops in these settings let us estimate generalization gaps, but we're also interested in developing confidence measures in future work.
This is still an open challenge - we don't yet have confidence estimation built into the model, but we do evaluate C2S-Scale in out-of-distribution regimes. For example:
- In perturbation prediction, we test on unseen cell typeβdrug combinations and combinatorial perturbations.
So performance likely reflects both mechanistic pattern recognition and domain transfer from literature and metadata. Our training corpus was intentionally multimodal to support this integration, letting the model ground textual knowledge in expression-level representations.
18.04.2025 17:10 β π 1 π 0 π¬ 1 π 0Great question, it might be a combination of both. For tasks like scQA, the model must (i) interpret gene expression patterns from cell sentences (e.g., identify marker genes or activation signatures), and (ii) relate those to biological concepts learned from the textual domain.
18.04.2025 17:10 β π 1 π 0 π¬ 1 π 0Many downstream tasks (e.g. scQA) require the model to reason jointly over cell sentences and biological text/metadata. We also explored this in our spatial reasoning ablation studies, where interleaving training with gene interaction data improved accuracy over training with expression alone.
18.04.2025 17:09 β π 1 π 0 π¬ 0 π 0C2S-Scale interleaves gene expression (as "cell sentences") with biological text during training to enable reasoning across both modalities. This multimodal integration is a key difference from expression-only models and is important for complex tasks.
18.04.2025 17:09 β π 1 π 0 π¬ 1 π 0We thank our amazing team at Yale, Google Research, and Google DeepMind
18.04.2025 14:13 β π 2 π 0 π¬ 0 π 0Dive into the details:
π Preprint: biorxiv.org/content/10.1...
π Google AI Blog: research.google/blog/teachin...
π» Code/Models: huggingface.co/collections/... github.com/vandijklab/c...
What's next for C2S-Scale?
β’ True Multimodality: Integrating proteomics, epigenomics, imaging data πΌοΈ
β’ Deeper Biology: Modeling cell interactions, dynamics, & development β³
β’ Enhanced Trust: Improving interpretability & reliability β
β’ Community Tools: Building shared benchmarks & platforms π
Let's build together! π οΈ We're open-sourcing C2S-Scale to empower the community.
Models up to 1B parameters are already available on HF, and models up to 27B parameters will be released in the next few weeks!
huggingface.co/collections/... github.com/vandijklab/c...
Beyond standard training, we used Reinforcement Learning (RL) π€ to fine-tune C2S-Scale.
Using GRPO + biological rewards, we specifically improved:
β’ Perturbation prediction accuracy π§ͺ
β’ Biological Q&A relevance β
Aligning LLMs with biological goals! β
Size matters! π We observed clear scaling laws: As model size increased from 410M β 27 Billion parameters, performance consistently improved across tasks.
This confirms that LLMs learn better biological representations at scale using the C2S approach. Even works with efficient LoRA tuning! πͺ
And it works! π C2S-Scale achieves SOTA performance, surpassing specialized single-cell models AND general LLMs:
β’ π― Cell type annotation
β’ π§ͺ Predicting perturbation responses
β’ βοΈ Generating dataset summaries from cells
β’ πΊοΈ Inferring spatial relationships
β’ β Answering complex biological questions
To truly "teach" biology to LLMs, we built a massive corpus: Over 1 BILLION tokens! π
This wasn't just cell sentences β it included:
⒠𧬠50M+ cell profiles (human/mouse)
β’ π·οΈ Annotations & Metadata
β’ π Biological Text (abstracts, etc.)
Result? One model, many tasks!
We enable LLMs to "read" biology via Cell2Sentence (C2S) π§¬β‘οΈπ: ranking genes creates text.
This lets us leverage massive pre-trained models, unifying transcriptomic data with biological text (annotations, papers) for richer understanding.
What if LLMs could βreadβ & βwriteβ biology? π€
Introducing C2SβScaleβa Yale and Google collab: we scaled LLMs (up to 27B!) to analyze & generate singleβcell data 𧬠β‘οΈ π
π Blog: research.google/blog/teachin...
π Preprint: biorxiv.org/content/10.1...
Huge thanks to the team: Zhikai Wu, Shiyang Zhang, Sizhuang He, Sifan Wang, Min Zhu, Anran Jiao, Lu Lu! Let us know what you think! #OperatorLearning #LLM #AI4Science
13.02.2025 19:23 β π 1 π 0 π¬ 0 π 0