Core Francisco Parkg's Avatar

Core Francisco Parkg

@corefpark.bsky.social

https://cfpark00.github.io/

30 Followers  |  16 Following  |  29 Posts  |  Joined: 11.11.2024  |  2.548

Latest posts by corefpark.bsky.social on Bluesky

Preview
$\textit{New News}$: System-2 Fine-tuning for Robust Integration of New Knowledge Humans and intelligent animals can effortlessly internalize new information ("news") and accurately extract the implications for performing downstream tasks. While large language models (LLMs) can ach...

Thanks for following the thread and big shout out to the team:

Zechen Zhang and @hidenori8tanaka.bsky.social

Here is the preprint:
arxiv.org/abs/2505.01812

14/n

21.05.2025 00:07 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

Key Takeaways:
โœ… There's a clear FT-ICL gap
โœ… Self-QA largely mitigates it
โœ… Larger models are more data efficient learners
โœ… Contextual shadowing hurts fine-tuning

Please check out the paper (see below) for even more findings!

13/n

21.05.2025 00:07 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

For humans, this contextualizes the work and helps integration, but this might be hindering learning in LLMs.

We are working on confirming this hypothesis on real data.

12/n

21.05.2025 00:07 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

We suspect this effect significantly harms fine-tuning as we know!

Let's take the example of research papers. In a typical research paper, the abstract usually โ€œspoilsโ€ the rest of the paper.

11/n

21.05.2025 00:07 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0
Video thumbnail

โš ๏ธโš ๏ธ But here comes drama!!!

What if the news appears in the context upstream of the *same* FT data?

๐Ÿšจ Contextual Shadowing happens!

Prefixing the news during FT *catastrophically* reduces learning!

10/n

21.05.2025 00:07 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 1
Post image

Next, we analyzed Sys2-FT from a scaling law perspective. We found an empirical scaling law of Sys2-FT where the knowledge integration is a function of the compute spent.

Larger models are thus more data efficient learners!

Note that this scaling isnโ€™t evident in loss.

9/n

21.05.2025 00:07 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0
Post image

Interestingly, Sys2-FT shines most in domains where System-2 inference has seen the most success: Math and Coding.

8/n

21.05.2025 00:07 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0
Video thumbnail

Among these protocols, Self-QA especially stood out, largely mitigating the FT-ICL gap and integrating the given knowledge into the model!

Training on synthetic Q/A pairs really boost knowledge integration!

7/n

21.05.2025 00:07 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0
Post image

Inspired by cognitive science on memory consolidation, we introduce System-2 Fine-Tuning (Sys2-FT). Models actively rehearse, paraphrase, and self-play about new facts to create fine-tuning data. We explore three protocols: Paraphrase, Implication, and Self-QA.

6/n

21.05.2025 00:07 โ€” ๐Ÿ‘ 1    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0
Video thumbnail

As expected, naรฏve fine-tuning on the raw facts isnโ€™t enough to integrate knowledge across domains or model sizes up to 32B.

We call this the FT-ICL gap.

5/n

21.05.2025 00:07 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0
Post image

But how do we update the modelโ€™s weights to bake in this new rule?

To explore this, we built โ€œNew Newsโ€: 75 new hypothetical (but non-counterfactual) facts across diverse domains, paired with 375 downstream questions.

4/n

21.05.2025 00:07 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

Todayโ€™s LLMs can effectively use *truly* novel information given as a context.

Given:
- Mathematicians defined 'addiplication' as (x+y)*y

Models can answer:
Q: What is the addiplication of 3 and 4?
A: (3+4)*4=28

3/n

21.05.2025 00:07 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

TL;DR: We present โ€œNew News,โ€ a dataset for measuring belief updates, and propose Self-QA, a highly effective way to integrate new knowledge via System-2 thinking at training time. We also show that in-context learning can actually hurt fine-tuning.

2/n

21.05.2025 00:07 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0
Post image

๐Ÿšจ New Paper!

A lot happens in the world every dayโ€”how can we update LLMs with belief-changing news?

We introduce a new dataset "New News" and systematically study knowledge integration via System-2 Fine-Tuning (Sys2-FT).

1/n

21.05.2025 00:07 โ€” ๐Ÿ‘ 7    ๐Ÿ” 1    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0
Post image

New paper <3
Interested in inference-time scaling? In-context Learning? Mech Interp?
LMs can solve novel in-context tasks, with sufficient examples (longer contexts). Why? Bc they dynamically form *in-context representations*!
1/N

05.01.2025 15:49 โ€” ๐Ÿ‘ 53    ๐Ÿ” 16    ๐Ÿ’ฌ 3    ๐Ÿ“Œ 1

Special thanks to @ndif-team.bsky.social for letting us run these experiments remotely!

15/n

05.01.2025 16:02 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

This project was a true collaborative effort where everyone contributed to major parts of the project!

Big thanks to the team: @ajyl.bsky.social, @ekdeepl.bsky.social, Yongyi Yang, Maya Okawa, Kento Nishi, @wattenberg.bsky.social, @hidenori8tanaka.bsky.social

14/n

05.01.2025 16:02 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0
Post image

We further investigate how this critical context size for an in-context transition scales with graph size.

We find a power law relationship between the critical context size and the graph size.

13/n

05.01.2025 16:02 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0
Post image

We find that LLMs indeed minimize the spectral energy on the graph and the rule-following accuracy sharply rises after the energy hits a minimum!

12/n

05.01.2025 16:02 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0
Post image

How to explain these results? We hypothesize a model runs an implicit optimization process to adapt to context-specified tasks (akin to in-context GD by @oswaldjoh et al), prompting an analysis of Dirichlet energy between the ground-truth graph & model representation.

11/n

05.01.2025 16:02 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0
Post image

What happens when there is a strong semantic structure acquired during pretraining?
We set up a task where the days of the week should be navigated in an unusual way: Mon -> Thu -> Sun, etc.

Here, we find that in-context representations show up in higher PC dimensions.

10/n

05.01.2025 16:02 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0
Post image

We call these context dependent representations โ€œIn-Context Representationsโ€ and these appear robustly across graph structures and models.

9/n

05.01.2025 16:02 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0
Post image

What about a different structure?

Here, we used a ring graph and sampled random neighbors on the graph.
Again, we find that internal representations re-organizes to match the task structure.

8/n

05.01.2025 16:02 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0
Post image

Interestingly, a similar phenomenon was observed in humans! One can reconstruct the graph underlying a sequence of random images from fMRI scans of the brain during the task.

elifesciences.org/articles/17086

7/n

05.01.2025 16:02 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0
Post image

Surprisingly, when we input this sequence to Llama-3.1-8B, the modelโ€™s internal representations show an emergent grid structure matching the task in its first principal components!

6/n

05.01.2025 16:02 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0
Post image

But do LLM representations also reflect the structure of a task given purely in context?

To explore this question, we set up a synthetic task where we put words on a grid and perform a random walk. The random walk outputs the words it accessed as a sequence.

5/n

05.01.2025 16:02 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0
Post image

We know that LLM representations reflect the structure of the real worldโ€™s data generating process. For example, @JoshAEngels showed that the days of the weeks are represented as a ring in the residual stream.

x.com/JoshAEngels/...

4/n

05.01.2025 16:02 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0
Preview
ICLR: In-Context Learning of Representations Recent work has demonstrated that semantics specified by pretraining data influence how representations of different concepts are organized in a large language model (LLM). However, given the open-end...

w/ @ajyl.bsky.social, @ekdeepl.bsky.social, Yongyi Yang, Maya Okawa, Kento Nishi, @wattenberg.bsky.social, @hidenori8tanaka.bsky.social

3/n

05.01.2025 16:02 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0
Preview
ICLR: In-Context Learning of Representations Recent work has demonstrated that semantics specified by pretraining data influence how representations of different concepts are organized in a large language model (LLM). However, given the open-end...

TL;DR: Given sufficient context, LLMs can suddenly shift from their concept representations to 'in-context representations' that align with the task structure!

Paper: arxiv.org/abs/2501.00070

2/n

05.01.2025 16:02 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0
Video thumbnail

New paper! โ€œIn-Context Learning of Representationsโ€

What happens to an LLMโ€™s internal representations in the large context limit?

We find that LLMs form โ€œin-context representationsโ€ to match the structure of the task given in context!

05.01.2025 16:02 โ€” ๐Ÿ‘ 7    ๐Ÿ” 2    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

@corefpark is following 15 prominent accounts