Tim Franzmeyer's Avatar

Tim Franzmeyer

@timlive.bsky.social

Machine Learning PhD student @UniofOxford interested in reinforcement learning, multi-agent systems, and LLMs. Previously @GoogleDeepMind, @MetaAI and @ETH.

25 Followers  |  54 Following  |  6 Posts  |  Joined: 21.11.2024  |  1.263

Latest posts by timlive.bsky.social on Bluesky

Preview
High Accuracy, Less Talk (HALT): Reliable LLMs through Capability-Aligned Finetuning Large Language Models (LLMs) currently respond to every prompt. However, they can produce incorrect answers when they lack knowledge or capability -- a problem known as hallucination. We instead propo...

πŸ“„ Full paper: arxiv.org/abs/2506.04051

With amazing collaborators:
Archie Sravankumar
Lijuan Liu
Yuning Mao
Rui Hou
Sinong Wang
@jfoerst.bsky.social
Madian Khabsa
@lukezettlemoyer.bsky.social

06.06.2025 08:21 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Post image

🚨 One model, high correctness:

With low-threshold tuning, we take Llama3-70B from:

➑️ 51% β†’ 87% correctness
➑️ Retaining 53% of the original completeness

06.06.2025 08:21 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Post image

βš–οΈ HALT allows you to trade off completeness and correctness

We introduce a threshold that tunes how eagerly the model should respond:

Low threshold = more reliable answers πŸ”’ (Left box)
High threshold = more detailed answers πŸ“(Right box)

06.06.2025 08:21 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Post image

πŸ› οΈ Our approach: Adjust finetuning responses to match the capabilities of the LLM

1️⃣ Break pretrained LLM responses into factual fragments
2️⃣ Use ground truth to flag incorrect fragments
3️⃣ Modify finetuning responses by removing or replacing errors with β€œUnsure from here” 🚧

06.06.2025 08:21 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

🧠 Standard LLMs always respond β€” even when unsure.

This leads to partially incorrect outputs in critical domains like Coding, Math, Medicine, and QA.

Why? Standard finetuning ignores what the pretrained model actually knows and pushes it to always complete every prompt.

06.06.2025 08:21 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Video thumbnail

What if LLMs knew when to stop? 🚧

HALT finetuning teaches LLMs to only generate content they’re confident is correct.

πŸ” Insight: Post-training must be adjusted to the model’s capabilities.
βš–οΈ Tunable trade-off: Higher correctness πŸ”’ vs. More completeness πŸ“

🧡

06.06.2025 08:21 β€” πŸ‘ 10    πŸ” 1    πŸ’¬ 1    πŸ“Œ 0

@timlive is following 19 prominent accounts