Yoav Goldberg's Avatar

Yoav Goldberg

@yoavgo.bsky.social

6,546 Followers  |  932 Following  |  1,295 Posts  |  Joined: 27.04.2023  |  2.1925

Latest posts by yoavgo.bsky.social on Bluesky

whats the difference in your view?

12.12.2025 17:17 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

i discuss this in the gist text. this is the more correct way to frame it imo (env provides observations, which agent interprets as rewards based on its goals), and it also opens up possible variations in how to think about learning from the env.

06.12.2025 00:44 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0
Preview
rl-wrong-about-rewards.md GitHub Gist: instantly share code, notes, and snippets.

I complain a lot about RL lately, and here we go again.

The CS view of RL is wrong in how it thinks about rewards, already at the setup level. Briefly, the reward computation should be part of the agent, not part of the environment.

More at length here:

gist.github.com/yoavg/3eb3e7...

05.12.2025 23:37 โ€” ๐Ÿ‘ 13    ๐Ÿ” 2    ๐Ÿ’ฌ 2    ๐Ÿ“Œ 0

yes it sucks to be the ICLR organizers today, totally agree

28.11.2025 00:01 โ€” ๐Ÿ‘ 2    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

given that the data is already out and a large jsonl file is rumored to be floating around (which seems very plausible to me), i think the moral thing to do now would be to make the breached data publicly available for all rather than trying to hide it.

27.11.2025 23:32 โ€” ๐Ÿ‘ 2    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

RL is ok. but the jump from
A) people can be thought of as agents who observe and environment, act, observe the outcome and update their beliefs

to:

B) lets model all things as a POMDP with a numeric reward function!

is just way too big for me

27.11.2025 20:13 โ€” ๐Ÿ‘ 4    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

the fascinating (to me) quality of hard-core RL researchers (e.g Sutton) is the ability to have an all encompassing view of RL as the basis of intelligence, while at the same time working on super low level stuff like tabular TD algorithms, and yet strongly believe these are actually the same thing

27.11.2025 16:32 โ€” ๐Ÿ‘ 20    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

ืžืกื›ื™ื ืœื’ืžืจื™

19.11.2025 02:07 โ€” ๐Ÿ‘ 1    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

ื•ื”ืืžื™ื™ืœ ื”ื–ื” (ืื ื™ ืžื ื™ื— ืฉืœืžื˜ืจืช ืฆื™ืœื•ืžื™ื ืจืฉืžื™ื™ื ืœืžื˜ืจื” ื›ืฉืœื”ื™ ืฉืœ ืืจื’ื•ืŸ ื›ืœืฉื”ื•) ื ืชืคืก ื›ืฆื™ื ื–ื•ืจ? (ืื ื™ ืคืฉื•ื˜ ืžื•ืคืชืข ื›ื™ ืื ื™ ืœื ื”ื™ื™ืชื™ ื—ื•ืฉื‘ ืขืœ ื–ื”)

18.11.2025 17:58 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

ื‘ืชื•ืจ ืื—ื“ ืฉืจื•ืื” ืขืฆืžื• ื›ืขืฉื•ื™ ืœื›ืชื•ื‘ ืžืฉื”ื• ื›ื–ื” ื‘ื˜ืขื•ืช ื•ืœื ืžื‘ื™ืŸ ืžื” ื”ืขื ื™ื™ืŸ, ืืฉืžื— ืื ืชื•ื›ืœื• ืœื”ืกื‘ื™ืจ ืžื” ื›ืœ ื›ืš ืžืงืคื™ืฅ ืคื”?

18.11.2025 16:24 โ€” ๐Ÿ‘ 1    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

ืœื ื”ื‘ื ืชื™ ืžื” ื”ืจืคืจื ืก ืœืฉื˜ืจ ืฉืœ ื›ืกืฃ ื•ื’ื ืœื ืจืื™ืชื™ ืืช ื”ื™ืจื•ืง ื“ื•ื•ืงื ื›ืจืคืจื•ืจ ืœืฆื”ืœ.. ืื‘ืœ ื›ืืžื•ืจ ืื•ืœื™ ื–ื” ื›ื™ ืื ื™ ื‘ืืžืช ืœื ืžืขืฆื‘ ืื– ืื ื™ ืœื ื—ื•ืฉื‘ ื‘ืžื•ื‘ื ื™ื ื”ืืœื•

18.11.2025 15:39 โ€” ๐Ÿ‘ 3    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

ื”ืื ืืช ืžืืžื™ื ื” ื‘ืขื™ืงืจื•ืŸ? ื›ื™ ื ืฉืžืข ืžืฆื™ื•ืฆื™ื ืื—ืจื™ื ืฉืœื ืžืžืฉ, ื•ืื–, ืžื” ืื›ืคืช ืœืš ื‘ืขืฆื ืขื“ ื›ืžื” ื–ื” ืžื“ื•ื™ื™ืง? ืื ื™ ืื™ืฉื™ืช ื›ืŸ ืžืืžื™ืŸ, ื•ืื›ืŸ ื”ื™ื™ืชื™ ืฉืžื— ืื ื–ื” ื™ืฉืชืคืจ ืœื”ื‘ื, ื•ืžืืžื™ืŸ ืฉืื›ืŸ ื›ืš ื™ื”ื™ื” ื›ื™ ื–ื” ื›ื•ืœื” ื‘ื•ืœื˜ ืฉืžืงื•ืฆืจ ื‘ืื•ืคืŸ ืœื ื‘ืจื•ืจ ืขืœ ืคื•ืกื˜ืจ.

18.11.2025 15:37 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 2    ๐Ÿ“Œ 0

ืœื ื”ืกืœื•ื’ืŸ ื”ืื™ื“ืืœื™ ืื‘ืœ ื’ื ื‘ืืžืช ืœื ื›ื–ื” ืžื•ืคืจืš. ื”ื›ืจื” ืžืคื•ืจืฉืช ื‘ื–ื›ื•ืช ืฉืœ ื™ืฉืจืืœ ืœื”ืชืงื™ื™ื ื›ื™ืฉื•ืช ืฆื™ื•ื ื™ืช, ืœืฆื“ ื”ืžื“ื™ื ื•ืช ื”ืขืจื‘ื™ื•ืช ื›ื•ืœืœ ื”ืคืœืกื˜ื™ื ื™ืช.

18.11.2025 15:11 โ€” ๐Ÿ‘ 3    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

ื›ืœื ืžืขืฆื‘, ื–ื” ื ืจืื” ืœื™ ืงืฆืช ื—ื•ื‘ื‘ื ื™ ืื‘ืœ ืกื”ื› ืžืžืฉ ื‘ืกื“ืจ. ืžื” ื”ื‘ืขื™ื”? ืžื” ื–ื” ื™ืจื•ืง ืœื ื ื›ื•ืŸ?

18.11.2025 15:10 โ€” ๐Ÿ‘ 5    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

what's the latest-and-greatest attempt to reverse-engineer and document the inner-working of claude-code?

17.11.2025 10:23 โ€” ๐Ÿ‘ 1    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

(hmm i guess we can amend to "increase in the proportion of knowledge we believe to be true")

17.11.2025 07:00 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

i think memory is never "free", in the sense that the real bottleneck is not storage, but the ability to retrieve the right thing, while not retrieving a wrong (out of date) thing by mistake.

but assuming we do delete facts, is deleting considered learning in your definition?

17.11.2025 06:59 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

is "increase" necessary? or is "change" enough? (although i guess that in an ideal form, you dont "forget" a wrong fact but add the fact that it is wrong, so you may consider it as increasing...)

16.11.2025 20:11 โ€” ๐Ÿ‘ 1    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

yes, following instructions in prompt is not learning. but if a wrapping systems stores items to inject in future prompts, then you can consider the system as learning.

16.11.2025 20:00 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

it will be in-context-induction, and the storing and retention from external memory would be learning.

16.11.2025 19:34 โ€” ๐Ÿ‘ 1    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

the storage, if it happens, is the learning part. the inference process is not learning.

16.11.2025 19:14 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0
Post image

or as i wrote two years ago:

gist.github.com/yoavg/59d174...

16.11.2025 17:27 โ€” ๐Ÿ‘ 1    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

i dont think it is a very useful view. at a very minimum we see extremely elaborate neighbor-matching and interpolation mechanisms, so the "glorified" part should be elaborated on and studied.

16.11.2025 17:25 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 2    ๐Ÿ“Œ 0

i agree, where is "storing" in the above case?

16.11.2025 17:04 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

ah, cool!

16.11.2025 15:54 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

indeed kNN is also not learning. its just a classification method. if you want to consider kNN as a learning method, then the learning part is just "store these pairs as is".

16.11.2025 15:54 โ€” ๐Ÿ‘ 1    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

i am not sure that it is (or rather, if everything is retrieval, then this term is useless)

16.11.2025 15:20 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 2    ๐Ÿ“Œ 0

if we want to study the phenomena, a non-misleading name may be better than a misleading one

16.11.2025 15:10 โ€” ๐Ÿ‘ 2    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

to me "learning" *requires* that something is stored for later use. again i dont care *where* it is stored, but *that* it is stored.

16.11.2025 15:09 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

this term is more accurate, but does not help in what i have in mind, which is to have a better name to the process that happens in ICL. some suggested "induction", which is OK but also not perfect (because the model both induces and applies).

16.11.2025 15:07 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

@yoavgo is following 20 prominent accounts