Instruction Following by Boosting Attention of Large Language Models
We improve instruction-following in large language models by boosting attention, a simple technique that outperforms existing steering methods.
Check out our blog and paper for more details!
πBlog: debugml.github.io/instaboost
πPaper: arxiv.org/abs/2506.13734
π€Code: github.com/BrachioLab/I...
Thank you to my awesome co-authers @viguardieiro.bsky.social, @avishree.bsky.social e.bsky.socialβ¬, and advisor @profericwong.bsky.social.
(7/7)
10.07.2025 18:21 β π 0 π 0 π¬ 0 π 0
Crucially, InstABoost achieves this control without degrading text quality. While other latent steering methods can cause generation fluency to drop sharply as you increase their strength, InstABoost maintains coherence while steering towards the instruction.
(6/7)
10.07.2025 18:21 β π 0 π 0 π¬ 1 π 0
Across 15 tasks, InstABoost either outperforms or matches the best steering method (prompt or latent-based). For tasks where prompt and latent-based steering perform equivalently, InstABoost can even combine the strengths of both and outperform both categories of methods.
(5/7)
10.07.2025 18:21 β π 0 π 0 π¬ 1 π 0
InstABoost steers an LLM in attention space, bridging the performance gap between latent and prompt-based steering. InstABoost can be implemented in ~3 lines of code which simply increases attention weight to an in-context instruction.
(3/7)
10.07.2025 18:21 β π 0 π 0 π¬ 1 π 0
Existing steering methods are either prompt or latent-based (modifying the hidden state), but which is better? We show the answer depends on the task. The steering task landscape includes those which are latent-optimal, instruction-optimal, and equivalent.
(2/7)
10.07.2025 18:21 β π 0 π 0 π¬ 1 π 0
Excited to share our new paper: "Instruction Following by Boosting Attention of Large Language Models"!
We introduce Instruction Attention Boosting (InstABoost), a simple yet powerful method to steer LLM behavior by making them pay more attention to instructions.
(π§΅1/7)
10.07.2025 18:21 β π 2 π 1 π¬ 1 π 1
As foundation models continue to scale, we argue itβs time to move beyond enforcing rigid symbolic structure in NeSy during training and tackle the exciting problem of inferring which symbols and which program are needed for each task.
(8/9)
13.06.2025 20:30 β π 0 π 0 π¬ 1 π 0
On the other hand, NeSy prompting provides two key benefits atop foundation models:
Reliability: A symbolic program enables accurate, stable, and trustworthy results.
Interpretability: Explicit symbols provide a clear, debuggable window into the model's "understanding."
(7/9)
13.06.2025 20:30 β π 0 π 0 π¬ 1 π 0
3οΈβ£ The Program Pitfall: Training neural nets in conjunction with a fixed program leads to "hallucinated" symbols, reaching the correct answer for the wrong reasons, similar to reasoning shortcuts.
(6/9)
13.06.2025 20:30 β π 0 π 0 π¬ 1 π 0
2οΈβ£ The Data Pitfall: Training on small, specialized datasets encourages overfitting.
(5/9)
13.06.2025 20:30 β π 0 π 0 π¬ 1 π 0
1οΈβ£ The Compute Pitfall: Training specialized NeSy models has diminishing returns. As foundation models scale, the gap between NeSy training and NeSy prompting disappears, making dedicated training a costly detour.
(4/9)
13.06.2025 20:30 β π 1 π 0 π¬ 1 π 0
We compare traditional NeSy systems (trained end-to-end) with what we call neuro-symbolic prompting (foundation models performing perception tasks via prompting connected to a symbolic program) and find that the NeSy training process itself introduces three key pitfalls.
(3/9)
13.06.2025 20:30 β π 0 π 0 π¬ 1 π 0
Neuro-symbolic learning combines neural nets + programs for efficient, interpretable AI. But NeSy training is challenging and brittle due to the symbolic component.
With foundation models succeeding via prompting alone, we argue itβs time to rethink NeSy system design.
(2/9)
13.06.2025 20:30 β π 0 π 0 π¬ 1 π 0
π§ Foundation models are reshaping reasoning. Do we still need specialized neuro-symbolic (NeSy) training, or can clever prompting now suffice?
Our new position paper argues the road to generalizable NeSy should be paved with foundation models.
π arxiv.org/abs/2505.24874
(π§΅1/9)
13.06.2025 20:30 β π 1 π 1 π¬ 1 π 0
NLP PhD student at UPenn | Prev USC
cylumn.com
CS PhD student @ UPenn.
Previously @ Microsoft, Google, BITS Pilani.
CS PhD @upenn.bsky.social
Computational Social Science @WorldBank
Harvard, Brown alumn
http://sehgal-neil.github.io/
Professor of Computer Science at UT Austin and Visiting Researcher at Google Deepmind, London. Automated Reasoning + Machine Learning + Formal Methods. https://www.cs.utexas.edu/~swarat
A feed of interesting AI / math / formal methods papers. Posts by @m-dodds.bsky.social
I lead Cohere For AI. Formerly Research
Google Brain. ML Efficiency, LLMs,
@trustworthy_ml.
Machine Learning Professor
https://cims.nyu.edu/~andrewgw
AI safety at Anthropic, on leave from a faculty job at NYU.
Views not employers'.
I think you should join Giving What We Can.
cims.nyu.edu/~sbowman
machine learning research
Doctor of NLP/Vision+Language from UCSB
Evals, metrics, multilinguality, multiculturality, multimodality, and (dabbling in) reasoning
https://saxon.me/
PhD student @ CMU LTI. working on text generation + long context
https://www.cs.cmu.edu/~abertsch/
The Thirty-Eighth Annual Conference on Neural Information Processing Systems will be held in Vancouver Convention Center, on Tuesday, Dec 10 through Sunday, Dec 15.
https://neurips.cc/
Research in science of {deep learning, AI security, safety}. PhD student at UPenn
davisrbrown.com
CS PhD Student @ MIT, interested in safe and reliable machine learning.