I was following this one during the COVID pandemic, but it has been inactive for quite some time. The original talks' recordings are amazing, though!
16.06.2025 14:20 β π 1 π 0 π¬ 1 π 0@manfreddiaz.bsky.social
Ph.D. Candidate at Mila and the University of Montreal, interested in AI/ML connections with economics, game theory, and social choice theory. https://manfreddiaz.github.io
I was following this one during the COVID pandemic, but it has been inactive for quite some time. The original talks' recordings are amazing, though!
16.06.2025 14:20 β π 1 π 0 π¬ 1 π 0Yeah, it's been a period for all of us simultaneously! I have also been pretty busy with thesis/job search. Hopefully, it will be back running in the Fall term!
05.06.2025 15:35 β π 1 π 0 π¬ 0 π 0@aamasconf.bsky.social 2025 was very special for us! We had the opportunity. to present a tutorial on general evaluation of AI agents, and we got a best paper award! Congrats, @sharky6000.bsky.social and the team! π
23.05.2025 14:23 β π 11 π 1 π¬ 0 π 0In the afternoon we will be giving a tutorial on general evaluation of AI agents.
sites.google.com/view/aamas20... 10/N
Announcing our latest arxiv paper:
Societal and technological progress as sewing an ever-growing, ever-changing, patchy, and polychrome quilt
arxiv.org/abs/2505.05197
We argue for a view of AI safety centered on preventing disagreement from spiraling into conflict.
Congrats, Seth!
01.05.2025 22:07 β π 1 π 0 π¬ 1 π 0First LessWrong post! Inspired by Richard Rorty, we argue for a different view of AI alignment, where the goal is "more like sewing together a very large, elaborate, polychrome quilt", than it is "like getting a clearer vision of something true and deep"
www.lesswrong.com/posts/S8KYwt...
The quality of London's museums is just amazing! Enjoy!
16.04.2025 01:50 β π 3 π 0 π¬ 0 π 0In case folks are interested, here's a video of a talk I gave at MIT a couple weeks ago: youtu.be/FmN6fRyfcsY?...
01.04.2025 20:50 β π 8 π 3 π¬ 0 π 0Our new evaluation method, Soft Condorcet Optimization is now available open-source! π
Both the sigmoid (smooth Kendall-tau) and Fenchel-Young (perturbed optimizers) versions.
Also, an optimized C++ implementation that is ~40X faster than the Python one. π€©β‘
github.com/google-deepm...
Working at the intersection of social choice and learning algorithms?
Check out the 2nd Workshop on Social Choice and Learning Algorithms (SCaLA) at @ijcai.bsky.social this summer.
Submission deadline: May 9th.
I attended last year at AAMAS and loved it! π
sites.google.com/corp/view/sc...
If the AAMAS website is a good reference for this, it may not be, but uncertain atm.
06.03.2025 05:34 β π 1 π 0 π¬ 1 π 0Come to understand ML evaluation from first principles! We have put together a great AAMAS tutorial covering statistics, probabilistic models, game theory, and social choice theory.
Bonus: a unifying perspective of the problem leveraging decision-theoretic principles!
Join us on May 19th!
Re #2: The key finding there is that the stationary points of SCO contain the margin matrix but, as I said in the note, there is still more work to do!
04.03.2025 19:31 β π 1 π 0 π¬ 1 π 0Thanks! I have been meaning to update the manuscript to standalone without the main paper but instead I may have change the content to a different format π. Coming soon!
04.03.2025 19:30 β π 1 π 0 π¬ 2 π 0Ah, I see the confusion... I never used the "identically distributed assumption," only the independence assumption (from 8 to 9).
25.02.2025 19:58 β π 1 π 0 π¬ 0 π 0I'm not sure if I understood your question correctly, but yes? As the post you shared says, "Voila! We have shown that minimizing the KL divergence amounts to finding the maximum likelihood estimate of ΞΈ." Maybe I am missing your point π¬
25.02.2025 19:48 β π 0 π 0 π¬ 2 π 0Elo drives most LLM evaluations, but we often overlook its assumptions, benefits, and limitations. While working on SCO, we wanted to understand the SCO-Elo distinction, so I looked and uncovered some intriguing findings and documented them in these notes. I hope you find them valuable!
25.02.2025 02:29 β π 2 π 1 π¬ 0 π 0Looking for a principled evaluation method for ranking of *general* agents or models, i.e. that get evaluated across a myriad of different tasks?
Iβm delighted to tell you about our new paper, Soft Condorcet Optimization (SCO) for Ranking of General Agents, to be presented at AAMAS 2025! π§΅ 1/N
I had the convexity results for the online pairwise update (Section B.1.1.1) in my notes (manfreddiaz.github.io/assets/pdf/s...), but it is not clear to me if they hold for the other non-online settings. Worth taking a more detailed pass over the paper!
20.02.2025 20:10 β π 2 π 0 π¬ 0 π 0That's a nice finding, @sacha2.bsky.social! @sharky6000.bsky.social I skimmed over it, and it seems neat! There is an important distinction, though. They work with the "online" Elo regime, departing from the traditional gradient/batch gradient descent updates. (e.g., FIDE doesn't use online updates)
20.02.2025 20:10 β π 2 π 0 π¬ 1 π 0lol π
12.02.2025 20:31 β π 3 π 0 π¬ 0 π 0Not that Michael Jordan, but this one en.wikipedia.org/wiki/Michael...
12.02.2025 20:29 β π 3 π 0 π¬ 1 π 0I believe this example conveys, as Prof. Jordan hinted, the need for fresh conceptual frameworks that shift our perspective, help us avoid conceptual confusion, and increase our ability to build the future of AI. I believe ML-SoA provides such framework, but Iβd love to hear more perspectives!
11.02.2025 20:57 β π 4 π 0 π¬ 1 π 0Through the ML-SoA design-centric view, reward hacking, misspecification and similar issues originate from designers' bounded rationality. The surprising βalignment problemsβ (www.youtube.com/watch?v=tlOI...) reflect ML designers' inability to comprehend the consequences of the programs they write.
11.02.2025 20:57 β π 3 π 0 π¬ 1 π 0For instance, in the AI safety literature, we hear phrases such as βthe agent hacked the rewardβ, βthe robot gamed the rewardβ, and similar statements. But does an ML model really "hack" or "game" anything? Or is this just a misleading metaphor? Chances are, itβs the latter.
11.02.2025 20:57 β π 2 π 0 π¬ 1 π 0None of these benefits is more critical than removing, to a greater extent, the anthropomorphism that has existed in AI, currently exacerbated by the rise of LLMs, whose principal peril is the conceptual confusion it sometimes causes. Let me clarify what I mean by "conceptual confusion".
11.02.2025 20:57 β π 1 π 0 π¬ 1 π 0We recently argued here (openreview.net/forum?id=LNY...) that examining ML through this perspective, which I call Machine Learning Through the Sciences of the Artificial (ML-SoA), offers solid foundations for understanding AI and ML in a broader context and introduces multiple benefits.
11.02.2025 20:57 β π 3 π 0 π¬ 1 π 0Back in the 1960, Herbert Simon defined in "The Sciences of the Artificial" an artificial entity as a product of human creation. In this context, AI shares foundations with other scientific theories that study human-designed structures such as institutions, markets, economies, and elections.
11.02.2025 20:57 β π 3 π 0 π¬ 1 π 0Last week, Michael I. Jordan's insightful talk at the AI Action Summit (www.youtube.com/live/W0QLq4q...) reminded us of the meaningful connections between AI, ML, economics, game theory, and mechanism design. But I'd argue the relationship goes deeperβit's profound, historical, and foundational. β¬οΈ
11.02.2025 20:57 β π 5 π 0 π¬ 1 π 1