Giuseppe Paolo's Avatar

Giuseppe Paolo

@giupaolo.bsky.social

Research scientist @Cognizant AI Lab Working on RL, Multi agent systems and Embodied AI. -> giupaolo.com ๐Ÿ“ท Photographer in my free time

73 Followers  |  447 Following  |  11 Posts  |  Joined: 21.11.2024  |  1.7151

Latest posts by giupaolo.bsky.social on Bluesky

Post image

๐Ÿš€Proud to share our work on the training dynamics in Transformers with Wassim Bouaziz & @viviencabannes.bsky.social @Inria @MetaAI

๐Ÿ“Easing Optimization Paths arxiv.org/pdf/2501.02362 (accepted @ICASSP 2025 ๐Ÿฅณ)

๐Ÿ“Clustering Heads ๐Ÿ”ฅhttps://arxiv.org/pdf/2410.24050

๐Ÿ–ฅ๏ธ github.com/facebookrese...

1/๐Ÿงต

04.02.2025 11:56 โ€” ๐Ÿ‘ 5    ๐Ÿ” 4    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 1
Dumb and Dumber gif: Jim Carrey plugging his ears while his companion babbles at him about something.
On top of Jim, it says "Me, Writing ICML Papers"
On top of his babbling companion, it says "DeepSeek Takes".

Dumb and Dumber gif: Jim Carrey plugging his ears while his companion babbles at him about something. On top of Jim, it says "Me, Writing ICML Papers" On top of his babbling companion, it says "DeepSeek Takes".

28.01.2025 18:42 โ€” ๐Ÿ‘ 62    ๐Ÿ” 2    ๐Ÿ’ฌ 2    ๐Ÿ“Œ 1

The bottom-up approach, starting from simpler systems, is much more interesting imo.

Studying how intelligence emerges naturally from simple rules is much more likely to teach us how intelligence arose in the first place.
2/2

19.01.2025 19:53 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

LLMs and today's AI research chase top-down intelligence through abstract reasoning. This path might get us to AGI faster but I donโ€™t think it will help us understand how intelligence emerged in nature.

1/2

19.01.2025 19:53 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

3- Ultimately, I believe we won't get to AGI until we solve the deeper challenge: giving true agency and meaning to these machines.

3/3

22.12.2024 13:49 โ€” ๐Ÿ‘ 1    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

2- @garymarcus.bsky.social raises a valid point: solving ARC through training isn't true AGI, and such claims are mainly hype. Real AGI should tackle these puzzles through reasoning alone, like humans do - not by optimizing for the benchmark. And I agree with it.

2/3

22.12.2024 13:49 โ€” ๐Ÿ‘ 1    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

I am usually quite critical of LLM hypey breakthroughs, but I donโ€™t see an issue with OpenAI training on ARC.
There is an issue, but itโ€™s more fundamental and multifaceted.

1- As itโ€™s often the case, the benchmark here has become the goal. @fchollet.bsky.social also said solving ARC != AGI

1/3

22.12.2024 13:49 โ€” ๐Ÿ‘ 2    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

By not doing research only on LLMs ;)

22.12.2024 13:44 โ€” ๐Ÿ‘ 3    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0
e introduce the effective horizon, a property of
MDPs that controls how difficult RL is. Our analysis is mo-
tivated by Greedy Over Random Policy (GORP), a simple
Monte Carlo planning algorithm (left) that exhaustively ex-
plores action sequences of length k and then uses m random
rollouts to evaluate each leaf node. The effective horizon
combines both k and m into a single measure. We prove
sample complexity bounds based on the effective horizon that
correlate closely with the real performance of PPO, a deep
RL algorithm, on our BRIDGE dataset of 155 deterministic
MDPs (right).

e introduce the effective horizon, a property of MDPs that controls how difficult RL is. Our analysis is mo- tivated by Greedy Over Random Policy (GORP), a simple Monte Carlo planning algorithm (left) that exhaustively ex- plores action sequences of length k and then uses m random rollouts to evaluate each leaf node. The effective horizon combines both k and m into a single measure. We prove sample complexity bounds based on the effective horizon that correlate closely with the real performance of PPO, a deep RL algorithm, on our BRIDGE dataset of 155 deterministic MDPs (right).

Kind of a broken record here but proceedings.neurips.cc/paper_files/...
is totally fascinating in that it postulates two underlying, measurable structures that you can use to assess if RL will be easy or hard in an environment

23.11.2024 18:18 โ€” ๐Ÿ‘ 151    ๐Ÿ” 28    ๐Ÿ’ฌ 8    ๐Ÿ“Œ 2

๐Ÿคฉ

23.11.2024 11:30 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

When I was 5, I loved science, which I took to be planets and magnets and chemicals and shit.

Soon I realized science was a collective human activity, and these were just the objects of its attention.

Anyway, whenever I read an OpEd on how science isn't political, I think "Are you 5 years old?"

22.11.2024 04:32 โ€” ๐Ÿ‘ 8079    ๐Ÿ” 1240    ๐Ÿ’ฌ 136    ๐Ÿ“Œ 65

I could note find any paper or work proposing AI frameworks like this.
I suppose cause it's hard, but maybe I am missing something?

21.11.2024 16:24 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

Since discovering @drmichaellevin.bsky.social work, I have been fascinated by hierarchical multi-agent systems in which the hierarchy develops across different ontological levels.
So hierarchy that goes
engineers -> the company they work for,
rather than
engineers -> their boss.

21.11.2024 16:24 โ€” ๐Ÿ‘ 3    ๐Ÿ” 1    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

Right on point, as I am getting more and more interested in this topic :)

21.11.2024 15:56 โ€” ๐Ÿ‘ 1    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

If you're an RL researcher or RL adjacent, pipe up to make sure I've added you here!
go.bsky.app/3WPHcHg

09.11.2024 16:42 โ€” ๐Ÿ‘ 70    ๐Ÿ” 26    ๐Ÿ’ฌ 52    ๐Ÿ“Œ 0

Not sure another social media app will increase my productivity.

But at least I can justify this one as useful for work.

21.11.2024 14:37 โ€” ๐Ÿ‘ 4    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

@giupaolo is following 20 prominent accounts