With GDM friends Adam Fisch, @jonathanberant.bsky.social, Alekh Agarwal, and special guest Anastasios Angelopoulos.
10.06.2025 15:24 โ ๐ 2 ๐ 1 ๐ฌ 0 ๐ 0@jonathanberant.bsky.social
NLP at Tel Aviv Uni and Google DeepMind
With GDM friends Adam Fisch, @jonathanberant.bsky.social, Alekh Agarwal, and special guest Anastasios Angelopoulos.
10.06.2025 15:24 โ ๐ 2 ๐ 1 ๐ฌ 0 ๐ 0We offer cost-optimal policies for selecting which rater should annotate which examples, which link the cost, the annotation noise, and the *uncertainty* of the cheaper rater.
10.06.2025 15:24 โ ๐ 1 ๐ 1 ๐ฌ 1 ๐ 0Cheap but noisy?
Or accurate but expensive?
How to split a limited annotation budget between different types of judges?๐ฉโโ๏ธ๐ค๐ฆง
www.arxiv.org/abs/2506.07949
calibration of p(answer), which is learned only when the helper identity is anonymized
An ablation reveals the importance of mechanism design: when the helper identities are known to the asker during training (CSP-DeAnon), calibrated hedging is no longer learned.
24.03.2025 15:39 โ ๐ 5 ๐ 1 ๐ฌ 1 ๐ 0task f1 and tool use
calibration curves for tool use, showing that collaborative self play teaches when to use the retrieval tools
In practice, collaborative self-play + reinforced self-training (ReST) lead to improved task performance, better calibration of confidence markers, and more efficient tool use.
24.03.2025 15:39 โ ๐ 5 ๐ 1 ๐ฌ 2 ๐ 0illustration of the equilibria of the formal model of costly information provision
A bit of game theory can help explain when this can work: we model the setup as a game of public utility provision, where the public utility is the extra information provided by the costly retrieval action. The game has a unique equilibrium when the tools are sufficiently distinct (or both bad).
24.03.2025 15:39 โ ๐ 4 ๐ 1 ๐ฌ 1 ๐ 0an example rollout, in which the asker receives contrasting advice from its helpers, and must rely on their confidence to find the accurate response
Because the identity of each helper is hidden from the asker, it is forced to rely on confidence signals when faced with incompatible answers from the helpers. Maximizing effort-penalized accuracy of the full rollout can teach the LLM to use these confidence markers correctly.
24.03.2025 15:39 โ ๐ 3 ๐ 1 ๐ฌ 1 ๐ 0schematic illustrating the collaborative self play scenario
We focus on two capabilities: knowing when to use a costly retrieval tool, and hedging non-confident answers. To teach these capabilities, we create a small multi-agent society, in which two "helpers" can use specialized retrieval tools to pass information back to an "asker"
24.03.2025 15:39 โ ๐ 5 ๐ 1 ๐ฌ 1 ๐ 0We all want LLMs to collaborate with humans to help them achieve their goals. But LLMs are not trained to collaborate, they are trained to imitate. Can we teach LM agents to help humans by first making them help each other?
arxiv.org/abs/2503.14481
Don't lie to your friends: Learning what you know from collaborative self-play Jacob Eisenstein, Reza Aghajani, Adam Fisch, Dheeru Dua, Fantine Huot, Mirella Lapata, Vicky Zayats, Jonathan Berant To be helpful assistants, AI agents must be aware of their own capabilities and limitations. This includes knowing when to answer from parametric knowledge versus using tools, when to trust tool outputs, and when to abstain or hedge. Such capabilities are hard to teach through supervised fine-tuning because they require constructing examples that reflect the agent's specific capabilities. We therefore propose a radically new approach to teaching agents what they know: \emph{collaborative self-play}. We construct multi-agent collaborations in which the group is rewarded for collectively arriving at correct answers. The desired meta-knowledge emerges from the incentives built into the structure of the interaction. We focus on small societies of agents that have access to heterogeneous tools (corpus-specific retrieval), and therefore must collaborate to maximize their success while minimizing their effort. Experiments show that group-level rewards for multi-agent communities can induce policies that \emph{transfer} to improve tool use and selective prediction in settings where individual agents are deployed in isolation.
A way to help models "be aware of their own capabilities and limitations" from @jacobeisenstein.bsky.social et al: arxiv.org/abs/2503.14481 #MLSky
22.03.2025 16:09 โ ๐ 42 ๐ 9 ๐ฌ 3 ๐ 0Fun work led by @amouyalsamuel.bsky.social and with Aya. Coming in I didn't think LLMs should have difficulties with answering questions on some of the GP sentences we used, but turns out they had! See Samuel's thread for more info...
12.03.2025 19:23 โ ๐ 0 ๐ 0 ๐ฌ 0 ๐ 0I had a lot of fun working on this with Aya Meltzer-Asscher and @jonathanberant.bsky.social .
We will soon release our materials, human results, LLM results and all the cool images the models produced on our sentences.
arxiv.org/abs/2502.09307
One intriguing follow-up: some component of the sentence understanding cognitive model fails on GP sentence. Is this component also present in LLMs? If not, then why so many LLMs are influenced by our manipulations in the same way humans are?
12.03.2025 19:12 โ ๐ 1 ๐ 1 ๐ฌ 1 ๐ 0There are many more cool insights you can find in our paper.
One takeaway from this paper for the psycholinguistics community: run your reading comprehension experiment on LLM first. You might get a general idea of the human results.
(Last image I swear)
These experiments replicated the results from the sentence comprehension one: our manipulations had the same effect on the paraphrase or drawing correctness as they had on the sentence comprehension task.
In this image: While the teacher taught the puppies looked at the board.
We also ran two additional experiments with LLMs that are challenging to perform on humans.
1. We asked the LLM to paraphrase our sentence
2. We asked text-to-image models to draw the sentences
In this image: While the horse pulled the submarine moved silently.
To answer our second question, we ran the same sentence comprehension experiment we ran on humans with over 60 LLMs.
We found that LLMs also struggle with GP sentences and that, interestingly, the manipulations we did to test our hypotheses impacted LLMs as they did with humans
In our latest paper with Aya Meltzer-Asscher and @jonathanberant.bsky.social, we try to answer both these questions.
We devise hypotheses explaining why GP sentences are harder to process and test them. Human subjects answered a reading comprehension question about a sentence they read.
The old man the boat.
You probably had to read that sentence twice. It's because it's a garden path (GP) sentence. GP sentences are read slower and often misunderstood. This begs the questions:
1. Why are these sentences harder to process?
2. How do LLMs deal with them?
Inference-time procedures (e.g. Best-of-N, CoT) have been instrumental to recent development of LLMs. Standard RLHF focuses only on improving the trained model. This creates a train/inference mismatch.
๐๐ข๐ฏ ๐ธ๐ฆ ๐ข๐ญ๐ช๐จ๐ฏ ๐ฐ๐ถ๐ณ ๐ฎ๐ฐ๐ฅ๐ฆ๐ญ ๐ต๐ฐ ๐ฃ๐ฆ๐ต๐ต๐ฆ๐ณ ๐ด๐ถ๐ช๐ต ๐ข ๐จ๐ช๐ท๐ฆ๐ฏ ๐ช๐ฏ๐ง๐ฆ๐ณ๐ฆ๐ฏ๐ค๐ฆ-๐ต๐ช๐ฎ๐ฆ ๐ฑ๐ณ๐ฐ๐ค๐ฆ๐ฅ๐ถ๐ณ๐ฆ?
Check out below.
In RMs trained on the helpfulness dataset we saw outputs tend to be a list of bullet points. In tl;dr summarization similar length and more copying (like others). Using entailment as a RM makes outputs shorter
arxiv.org/abs/2312.09244
Preference data distribution doesn't explain this btw
InfAlign: Inference-aware language model alignment Ananth Balashankar, Ziteng Sun, Jonathan Berant, Jacob Eisenstein, Michael Collins, Adrian Hutter, Jong Lee, Chirag Nagpal, Flavien Prost, Aradhana Sinha, Ananda Theertha Suresh, Ahmad Beirami
Excited to share ๐๐ง๐๐๐ฅ๐ข๐ ๐ง!
Alignment optimization objective implicitly assumes ๐ด๐ข๐ฎ๐ฑ๐ญ๐ช๐ฏ๐จ from the resulting aligned model. But we are increasingly using different and sometimes sophisticated inference-time compute algorithms.
How to resolve this discrepancy?๐งต
Sure, but the take that "we should allow people to publicly make such comments because otherwise we would never know that the problem exists" seems odd to me, and I assume that is not your claim.
14.12.2024 18:54 โ ๐ 1 ๐ 0 ๐ฌ 1 ๐ 0I haven't either don't know if such discussions exist or not.
14.12.2024 18:46 โ ๐ 0 ๐ 0 ๐ฌ 1 ๐ 0Others shouldn't speak similarly, this doesn't seem like "hiding" to me. You can be aware of problems and discuss them appropriately and that is orthogonal to expecting public speakers to not make such offhand comments (when bias is not even the topic of discussion)
14.12.2024 18:41 โ ๐ 0 ๐ 0 ๐ฌ 1 ๐ 0I think he only said the first part?
13.12.2024 23:50 โ ๐ 0 ๐ 0 ๐ฌ 0 ๐ 0Weโre really excited to release this large collaborative work for unifying web agent benchmarks under the same roof.
In this TMLR paper, we dive in-depth into #BrowserGym and #AgentLab. We also present some unexpected performances from Claude 3.5-Sonnet
Really curious to know how all this turned out keep reporting!
12.12.2024 04:49 โ ๐ 0 ๐ 0 ๐ฌ 0 ๐ 0I will also be at NeurIPS! Happy to chat about post-training, reasoning, and interesting ways you use multiple agents for things.
09.12.2024 19:34 โ ๐ 1 ๐ 0 ๐ฌ 0 ๐ 0