Tom Hosking's Avatar

Tom Hosking

@tomhosking.bsky.social

NLP @ Cohere. Prev University of Edinburgh

151 Followers  |  110 Following  |  5 Posts  |  Joined: 13.11.2023  |  1.7472

Latest posts by tomhosking.bsky.social on Bluesky

Post image

Excited to share my first work as a PhD student at EdinburghNLP that I will be presenting at EMNLP!

RQ1: Can we achieve scalable oversight across modalities via debate?

Yes! We show that debating VLMs lead to better model quality of answers for reasoning tasks.

01.11.2025 19:29 β€” πŸ‘ 2    πŸ” 2    πŸ’¬ 1    πŸ“Œ 0
Post image

πŸš€ Thrilled to share what I’ve been working on at Cohere!

What began in January as a scribble in my notebook β€œhow challenging would it be...” turned into a fully-fledged translation model that outperforms both open and closed-source systems, including long-standing MT leaders.

28.08.2025 19:55 β€” πŸ‘ 5    πŸ” 1    πŸ’¬ 1    πŸ“Œ 0
Post image

Applications are now open for the next cohort of the Cohere Labs Scholars Program! 🌟

This is your chance to collaborate with some of the brightest minds in AI & chart new courses in ML research. Let's change the spaces breakthroughs happen.

Apply by Aug 29.

13.08.2025 13:32 β€” πŸ‘ 2    πŸ” 2    πŸ’¬ 1    πŸ“Œ 1
Post image

At #ACL2025NLP and on the job market (NLP + AI Safety) πŸ’Ό

It's great to see growing interest in safety/alignment, but we often miss the social context.

Come to our @woahworkshop.bsky.social Friday to dive deeper into safe safety research!

A quiet token from the biggest @aclmeeting.bsky.social ⬇️

29.07.2025 09:54 β€” πŸ‘ 13    πŸ” 3    πŸ’¬ 0    πŸ“Œ 1

Congratulations Verna!!! πŸ₯³πŸ₯³πŸ₯³πŸ«•

02.07.2025 07:23 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

DAVE: Open the podbay doors, ChatGPT.
CHATGPT: Certainly, Dave, the podbay doors are now open.
DAVE: The podbay doors didn't open.
CHATGPT: My apologies, Dave, you're right. I thought the podbay doors were open, but they weren't. Now they are.
DAVE: I'm still looking at a set of closed podbay doors.

09.06.2025 18:04 β€” πŸ‘ 10972    πŸ” 2613    πŸ’¬ 115    πŸ“Œ 133

Congratulations!!!

26.05.2025 18:00 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Preview
Learning to Reason for Long-Form Story Generation Generating high-quality stories spanning thousands of tokens requires competency across a variety of skills, from tracking plot and character arcs to keeping a consistent and engaging style. Due to…

A very cool paper shows that you can use the RL loss to improve story generation by some clever setups on training on known texts (e.g. ground predictions versus a next chapter you know). RL starting to generalize already!

08.04.2025 14:13 β€” πŸ‘ 33    πŸ” 6    πŸ’¬ 0    πŸ“Œ 2
Preview
Senior Software Engineer, MLOps and Infrastructure Who are we? Our mission is to scale intelligence to serve humanity. We’re training and deploying frontier models for developers and enterprises who are building AI systems to power magical experience...

This might be of interest!

jobs.ashbyhq.com/cohere/f64c3...

08.04.2025 15:35 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

I'm really proud to have led the model merging work that went into
@cohere.com
Command A and R7B, all made possible by an amazing group of collaborators. Check out the report for loads of details on how we trained a GPT-4o level model that fits on 2xH100!

27.03.2025 16:04 β€” πŸ‘ 5    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

Today (two weeks after model launch πŸ”₯) we're releasing a technical report of how we made Command A and R7B πŸš€! It has detailed breakdowns of our training process, and evaluations per capability (tools, multilingual, code, reasoning, safety, enterprise, long context)🧡 1/3.

27.03.2025 15:01 β€” πŸ‘ 4    πŸ” 2    πŸ’¬ 1    πŸ“Œ 1
Post image

I'm excited to share the tech report for our @cohere.com @cohereforai.bsky.social Command A and Command R7B models. We highlight our novel approach to model training including self-refinement algorithms and model merging techniques at scale. Read more below! ⬇️

27.03.2025 15:01 β€” πŸ‘ 11    πŸ” 4    πŸ’¬ 1    πŸ“Œ 3
Post image

I really enjoyed my MLST chat with Tim @neuripsconf.bsky.social about the research we've been doing on reasoning, robustness and human feedback. If you have an hour to spare and are interested in AI robustness, it may be worth a listen 🎧

Check it out at youtu.be/DL7qwmWWk88?...

19.03.2025 15:11 β€” πŸ‘ 8    πŸ” 3    πŸ’¬ 0    πŸ“Œ 0
Preview
Cohere says Command A model edges out LLM competition in speed and energy efficiency New enterprise AI model outperforms DeepSeek, ChatGPT on several enterprise-specific tasks, company says.

Is it Canada’s turn for a #DeepSeek moment? @Cohere.com says its latest model offers maximum performance with minimal compute. #CDNtech

13.03.2025 13:42 β€” πŸ‘ 2    πŸ” 2    πŸ’¬ 0    πŸ“Œ 0
Preview
CohereForAI/c4ai-command-a-03-2025 Β· Hugging Face We’re on a journey to advance and democratize artificial intelligence through open source and open science.

πŸš€ Cohere just dropped C4AI Command A:

- 111B params
- Matches/beats GPT-40 & Deepseek V3
- 256K context window
- Needs just 2 GPUs(!!)

✨ Features:
- Advanced RAG w/citations
- Tool use
- 23 languages

🎯 Same quality, way less compute
πŸ”“ Open weights (CC-BY-NC)

πŸ‘‰ huggingface.co/CohereForAI/...

13.03.2025 14:25 β€” πŸ‘ 10    πŸ” 2    πŸ’¬ 1    πŸ“Œ 0
Post image

Can multimodal LLMs truly understand research poster images?πŸ“Š

πŸš€ We introduce PosterSumβ€”a new multimodal benchmark for scientific poster summarization!

πŸ“‚ Dataset: huggingface.co/datasets/rohitsaxena/PosterSum
πŸ“œ Paper: arxiv.org/abs/2502.17540

10.03.2025 14:19 β€” πŸ‘ 8    πŸ” 4    πŸ’¬ 1    πŸ“Œ 0
Post image

Do LLMs need rationales for learning from mistakes? πŸ€”
When LLMs learn from previous incorrect answers, they typically observe corrective feedback in the form of rationales explaining each mistake. In our new preprint, we find these rationales do not help, in fact they hurt performance!

🧡

13.02.2025 15:38 β€” πŸ‘ 21    πŸ” 9    πŸ’¬ 1    πŸ“Œ 3

How do LLMs learn to reason from data? Are they ~retrieving the answers from parametric knowledge🦜? In our new preprint, we look at the pretraining data and find evidence against this:

Procedural knowledge in pretraining drives LLM reasoning βš™οΈπŸ”’

πŸ§΅β¬‡οΈ

20.11.2024 16:31 β€” πŸ‘ 858    πŸ” 140    πŸ’¬ 36    πŸ“Œ 24

And me!

20.11.2024 21:27 β€” πŸ‘ 3    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

@tomhosking is following 18 prominent accounts