Csaba Szepesvari's Avatar

Csaba Szepesvari

@skiandsolve.bsky.social

⛷️ ML Theorist carving equations and mountain trails | πŸš΄β€β™‚οΈ Biker, Climber, Adventurer | 🧠 Reinforcement Learning: Always seeking higher peaks, steeper walls and better policies. https://ualberta.ca/~szepesva

1,187 Followers  |  223 Following  |  92 Posts  |  Joined: 30.11.2024  |  2.3364

Latest posts by skiandsolve.bsky.social on Bluesky

..actually, not only standard notation, but also to be able to speak about the loss (=log-loss) used to train today's LLMs.

10.07.2025 15:19 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

No, it is not information retrieval. It is deducing new things from old things. You can do this by running a blind breadth-first (unintelligent) search producing all proofs of all possible statements. Just don't want errors. But this is not retrieval. It is computation.

10.07.2025 05:21 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

Of course approximations are useful. The paper is narrowly focused on deductive reasoning which seem to require the exactness we talk about. The point is that regardless of whether you use quantum mechanics or the Newtonian one, you don't want your derivations mistake-ridden.

10.07.2025 05:19 β€” πŸ‘ 2    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

Worst-case vs. average case: yes!
But I would not necessarily connect these to minimax vs. Bayes.

10.07.2025 05:14 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

Yeah, admittedly, not a focus point of the paper. How about if the model produces a single response, the loss is the zero-one loss. Then the model better choose the label with the highest probability label, which is OK. Point of having mu: Not much point, just matching standard notation..

10.07.2025 05:13 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

I am curious about these examples.. (and yes, I can construct a few, too, but I want to add more)

10.07.2025 04:51 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

No, this is not correct: Learning 1[A>B] interestingly has the same complexity (provably). This is because 1[A>B] is in the "orbit" of 1[A>=B]. So the symmetric learning who is being taught 1[A>B] need to figure out it is not taught 1[A>=B].

10.07.2025 04:50 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

Maybe. I am asking for much less here from the machines. I am asking for them just to be correct (or stay silent). No intelligence, just good old fashioned computation.

09.07.2025 02:44 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

the solution is found..

09.07.2025 02:42 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

Yes, transformers do not have "working memory". Also, I don't believe in that using them in AR mode is powerful enough for challenging problems. In a way, without "working memory", external "loop", we say the model should solve problems by free association ad infinitum or at least until

09.07.2025 02:42 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

On the paper: Interesting but indeed there is little in common. On the problem studied in the paper: Would not a slightly more general statistical framework solve your problem? Ie measure error differently than through the prediction loss (AR models: parameters, spectral measure, etc.).

09.07.2025 02:39 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

Yeah, I don't see the exactness happening that much on its own through statistical learning. Neither experimentally, nor theoretically. We have an example for illustrating this: use the uniform distribution for good coverage, teach transformers to compare m-bit integers using GD. Need 2^m examples.

09.07.2025 02:39 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 3    πŸ“Œ 0

Yeah, we cite this and this was a paper that got me started on this project!

09.07.2025 02:32 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Preview
Beyond Statistical Learning: Exact Learning Is Essential for General Intelligence Sound deductive reasoning -- the ability to derive new knowledge from existing facts and rules -- is an indisputably desirable aspect of general intelligence. Despite the major advances of AI systems ...

First position paper I ever wrote. "Beyond Statistical Learning: Exact Learning Is Essential for General Intelligence" arxiv.org/abs/2506.23908 Background: I'd like LLMs to help me do math, but statistical learning seems inadequate to make this happen. What do you all think?

08.07.2025 02:21 β€” πŸ‘ 51    πŸ” 9    πŸ’¬ 4    πŸ“Œ 1

Our seminars are back. If you missed Max's talk, it is on YouTube and today I will host Jeongyeol from UWM who will talk about the curious case of why latent MDPs though scary at first sight might be tractable! Link to the seminar homepage:
sites.google.com/view/rltheor...

06.05.2025 17:09 β€” πŸ‘ 23    πŸ” 3    πŸ’¬ 0    πŸ“Œ 0

Glad to see someone remembers these:)

04.04.2025 02:05 β€” πŸ‘ 7    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

should be distinguished. The reason they should not is because they are indistinguishable. So at least those need to be collapsed. So yes, one can start with redundant models, where it will appear you could have epistemic uncertainty, but this is easy to rule out. 2/2

20.03.2025 22:49 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

I guess with a worst-case hat on, we just all die:) In other words, indeed, the distinction is useful inasmuch as the modelling assumptions are valid. And there the mixture of two Diracs over 0 and 1 actually is a bad example, because that says that two models that are identical as distributions 1/x

20.03.2025 22:47 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

I guess I stop here:) 5/5

20.03.2025 22:43 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

Well, yes, to the degree that the model you use correctly reflects what's going on. Example with drug trials, randomized patient allocation. Result is effectiveness. Meaning of aleatoric and epistemic uncertainty should be clear and they help with explaining outcomes of the trial. 4/x

20.03.2025 22:41 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

One observes 1, there is epistemic uncertainty (the model could be the first or the second). Of course, nothing is black and white like this ever. And we talk about models here. Models are.. made up.. Usual blurb about usefulness of models. Should you care about this distinction? 3/x

20.03.2025 22:35 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

Epistemic uncertainty refers to whether given the data (and prior information), we can surely identify the data generating model. Example: Model class has two distributions; one has support {0,1}, the other has support {1}. One observes 0. There is no epistemic uncertainty. 2/X

20.03.2025 22:33 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

I don't get this:
In the context of this terminology, data comes from a model. Aleatoric uncertainty refers to the case when this model is a Dirac! In the second case, the model is a mixture of two Dirac's. This is not a Dirac. Hence, there is aleatoric uncertainty. 1/X

20.03.2025 22:31 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
NSERC - Latest News - Launch of the new Harmonized Tri-agency Scholarship and Fellowship programs As announced in Budget 2024, the scholarship and fellowship programs administered by the three federal research funding agencies – the Canadian Institutes of Health Research (CIHR), the Natural Sciences and Engineering Research Council (NSERC), and the Social Sciences and Humanities Research Council (SSHRC) – have been streamlined into a new harmonized talent program called the Canada Research Training Awards Suite (CRTAS) that will open for applications in summer 2025.

This is a very significant development - more fellowships, harmonized and typically higher stipends, and international students can apply

#CanPoli

www.nserc-crsng.gc.ca/NewsDetail-D...

15.03.2025 17:53 β€” πŸ‘ 37    πŸ” 16    πŸ’¬ 2    πŸ“Œ 3

Dylan J. Foster, Zakaria Mhammedi, Dhruv Rohatgi: Is a Good Foundation Necessary for Efficient Reinforcement Learning? The Computational Role of the Base Model in Exploration https://arxiv.org/abs/2503.07453 https://arxiv.org/pdf/2503.07453 https://arxiv.org/html/2503.07453

11.03.2025 07:26 β€” πŸ‘ 5    πŸ” 5    πŸ’¬ 1    πŸ“Œ 0

But also we are how we act! So it's up to us all to behave so as to make statement true.

11.03.2025 16:37 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Preview
From the nonononoyes community on Reddit: He was there for a while Explore this post and more from the nonononoyes community

Who says mountain car is a toy problem? www.reddit.com/r/nonononoye...

09.03.2025 17:46 β€” πŸ‘ 6    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

Yes, another gem from Rich!

07.03.2025 02:58 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
TURING AWARD WINNER Richard S. Sutton in Conversation with Cam Linke | No Authorities in Science
YouTube video by Amii TURING AWARD WINNER Richard S. Sutton in Conversation with Cam Linke | No Authorities in Science

www.youtube.com/watch?v=9_Pe... An interview with Rich. The humility of Rich is truly inspiring: "There are no authorities in science". I wish people would listen and live by this.

06.03.2025 20:50 β€” πŸ‘ 39    πŸ” 13    πŸ’¬ 2    πŸ“Œ 1

That's all good: Bubbles join when they get up high into the blue sky:)

06.03.2025 20:42 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

@skiandsolve is following 20 prominent accounts