Excited to share my first work as a PhD student at EdinburghNLP that I will be presenting at EMNLP!
RQ1: Can we achieve scalable oversight across modalities via debate?
Yes! We show that debating VLMs lead to better model quality of answers for reasoning tasks.
01.11.2025 19:29 β π 2 π 2 π¬ 1 π 0
π Thrilled to share what Iβve been working on at Cohere!
What began in January as a scribble in my notebook βhow challenging would it be...β turned into a fully-fledged translation model that outperforms both open and closed-source systems, including long-standing MT leaders.
28.08.2025 19:55 β π 5 π 1 π¬ 1 π 0
Applications are now open for the next cohort of the Cohere Labs Scholars Program! π
This is your chance to collaborate with some of the brightest minds in AI & chart new courses in ML research. Let's change the spaces breakthroughs happen.
Apply by Aug 29.
13.08.2025 13:32 β π 2 π 2 π¬ 1 π 1
At #ACL2025NLP and on the job market (NLP + AI Safety) πΌ
It's great to see growing interest in safety/alignment, but we often miss the social context.
Come to our @woahworkshop.bsky.social Friday to dive deeper into safe safety research!
A quiet token from the biggest @aclmeeting.bsky.social β¬οΈ
29.07.2025 09:54 β π 13 π 3 π¬ 0 π 1
Congratulations Verna!!! π₯³π₯³π₯³π«
02.07.2025 07:23 β π 1 π 0 π¬ 0 π 0
DAVE: Open the podbay doors, ChatGPT.
CHATGPT: Certainly, Dave, the podbay doors are now open.
DAVE: The podbay doors didn't open.
CHATGPT: My apologies, Dave, you're right. I thought the podbay doors were open, but they weren't. Now they are.
DAVE: I'm still looking at a set of closed podbay doors.
09.06.2025 18:04 β π 10972 π 2613 π¬ 115 π 133
Congratulations!!!
26.05.2025 18:00 β π 1 π 0 π¬ 0 π 0
Learning to Reason for Long-Form Story Generation
Generating high-quality stories spanning thousands of tokens requires competency across a variety of skills, from tracking plot and character arcs to keeping a consistent and engaging style. Due toβ¦
A very cool paper shows that you can use the RL loss to improve story generation by some clever setups on training on known texts (e.g. ground predictions versus a next chapter you know). RL starting to generalize already!
08.04.2025 14:13 β π 33 π 6 π¬ 0 π 2
I'm really proud to have led the model merging work that went into
@cohere.com
Command A and R7B, all made possible by an amazing group of collaborators. Check out the report for loads of details on how we trained a GPT-4o level model that fits on 2xH100!
27.03.2025 16:04 β π 5 π 0 π¬ 0 π 0
Today (two weeks after model launch π₯) we're releasing a technical report of how we made Command A and R7B π! It has detailed breakdowns of our training process, and evaluations per capability (tools, multilingual, code, reasoning, safety, enterprise, long context)π§΅ 1/3.
27.03.2025 15:01 β π 4 π 2 π¬ 1 π 1
I'm excited to share the tech report for our @cohere.com @cohereforai.bsky.social Command A and Command R7B models. We highlight our novel approach to model training including self-refinement algorithms and model merging techniques at scale. Read more below! β¬οΈ
27.03.2025 15:01 β π 11 π 4 π¬ 1 π 3
I really enjoyed my MLST chat with Tim @neuripsconf.bsky.social about the research we've been doing on reasoning, robustness and human feedback. If you have an hour to spare and are interested in AI robustness, it may be worth a listen π§
Check it out at youtu.be/DL7qwmWWk88?...
19.03.2025 15:11 β π 8 π 3 π¬ 0 π 0
CohereForAI/c4ai-command-a-03-2025 Β· Hugging Face
Weβre on a journey to advance and democratize artificial intelligence through open source and open science.
π Cohere just dropped C4AI Command A:
- 111B params
- Matches/beats GPT-40 & Deepseek V3
- 256K context window
- Needs just 2 GPUs(!!)
β¨ Features:
- Advanced RAG w/citations
- Tool use
- 23 languages
π― Same quality, way less compute
π Open weights (CC-BY-NC)
π huggingface.co/CohereForAI/...
13.03.2025 14:25 β π 10 π 2 π¬ 1 π 0
Can multimodal LLMs truly understand research poster images?π
π We introduce PosterSumβa new multimodal benchmark for scientific poster summarization!
π Dataset: huggingface.co/datasets/rohitsaxena/PosterSum
π Paper: arxiv.org/abs/2502.17540
10.03.2025 14:19 β π 8 π 4 π¬ 1 π 0
Do LLMs need rationales for learning from mistakes? π€
When LLMs learn from previous incorrect answers, they typically observe corrective feedback in the form of rationales explaining each mistake. In our new preprint, we find these rationales do not help, in fact they hurt performance!
π§΅
13.02.2025 15:38 β π 21 π 9 π¬ 1 π 3
How do LLMs learn to reason from data? Are they ~retrieving the answers from parametric knowledgeπ¦? In our new preprint, we look at the pretraining data and find evidence against this:
Procedural knowledge in pretraining drives LLM reasoning βοΈπ’
π§΅β¬οΈ
20.11.2024 16:31 β π 858 π 140 π¬ 36 π 24
And me!
20.11.2024 21:27 β π 3 π 0 π¬ 1 π 0
Chair of Psychology of Language Learning at Abertay University. Here for language & politics. IRL also tango.
Postdoctoral Fellow @ Princeton AI
I write books and do standup comedy and help run a production company. I was on Taskmaster. I am usually at Paddington station.
Working on evaluation of AI models (via human and AI feedback) | PhD candidate @cst.cam.ac.uk
Web: https://arduin.io
Github: https://github.com/rdnfn
Latest project: https://app.feedbackforensics.com
Infrastructure, humans, production operations, security. Oh, and dogs.
@Cohere.com's non-profit research lab and open science initiative that seeks to solve complex machine learning problems. Join us in exploring the unknown, together. https://cohere.com/research
PhD student @ University of Edinburgh | Looking for Posdoc Opportunity | Interested in Planning, Reasoning, Long Context | Multimodal AI | π https://anilbatra2185.github.io/
Professor at UW; Researcher at Meta. LMs, NLP, ML. PNW life.
Associate professor at CMU, studying natural language processing and machine learning. Co-founder All Hands AI
SNSF Professor at University of Zurich. #NLP / #ML.
http://www.cl.uzh.ch/sennrich
Professor at the University of Sheffield. I do #NLProc stuff.
Research Scientist at Meta β’ ex Cohere, Google DeepMind β’ https://www.ruder.io/
Computational LinguistsβNatural LanguageβMachine Learning