's Avatar

@diegodoimo.bsky.social

26 Followers  |  75 Following  |  6 Posts  |  Joined: 08.12.2024  |  1.4729

Latest posts by diegodoimo.bsky.social on Bluesky

This work is a collaboration with a team of talented researchers at the AreaSciencePark of Trieste, Italy.
Special thanks to @alexpietroserra.bsky.social, Alessio Ansuini and @albecazzaniga.bsky.social !

If you are @neuripsconf.bsky.social don't miss our poster tomorrow, Dec 11, at 11am!!
🧡6/6

10.12.2024 21:46 β€” πŸ‘ 2    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Post image

βš’οΈ We applied an advanced density-based clustering algorithm, showing its potential as an interpretability tool and in guiding novel strategies for the effective finetuning of LLMs.
🧡5/6

10.12.2024 19:54 β€” πŸ‘ 2    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Post image

In fine-tuning, answer-focused modes rapidly emerge midway through the network, just after the intrinsic dimension peak.
Early layers remain largely unchanged.
🧡4/6

10.12.2024 19:52 β€” πŸ‘ 2    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Post image

In few-shot learning, the prompt topic defines the modes of data distribution early in the network, and density modes are hierarchically organized based on the similarity of the subjects.
🧡3/6

10.12.2024 19:49 β€” πŸ‘ 2    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Post image

🎯 Key results: few-shot learning and fine-tuning show two distinct processing phases inside LLMs.

These phases are separated by a peak of the data intrinsic dimension and a sharp decrease in the separation of the probability modes.

Paper: arxiv.org/abs/2409.03662
🧡2/6

10.12.2024 19:48 β€” πŸ‘ 2    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Post image

Just landed in Vancouver to present @neuripsconf.bsky.social the results of our new work!

Few-shot learning and fine-tuning change the layers inside LLMs in a dramatically different way, even when they perform equally well on multiple-choice question-answering tasks.
🧡1/6

10.12.2024 19:47 β€” πŸ‘ 10    πŸ” 0    πŸ’¬ 1    πŸ“Œ 3

@diegodoimo is following 20 prominent accounts