1-panel SMBC comic update with the caption "Science pro tip: you can prove anything you want using regression to the mean" where a scientist explains he can up student grades by smearing peanut butter and googly eyes on the foreheads of the worse performing ones.
If you oppose this it's because you want line go down, but I want line go up.
COMIC β www.smbc-comics.com/comic/protocol
PATREON β www.patreon.com/c/ZachWeiner...
STORE β smbc-store.myshopify.com
01.08.2025 22:30 β π 38 π 5 π¬ 2 π 1
I was the AEA's President Elect as the first Trump term began. We worried about government statistics then, and appointed a committee that included members (such as Google's Chief Economist) who might be in a position to help make reliable statistics available if the government went dark.
#econsky
02.08.2025 23:53 β π 182 π 70 π¬ 3 π 1
Tried this out; generally very impressed with the quality of the feedback!
25.07.2025 01:15 β π 13 π 1 π¬ 0 π 0
This was also part of a fun exploration we did at the same time in how to incorporate modern AI tools in the process of doing theoretical and experimental science. We learned a lot about that too --- we'll write more at some point.
24.07.2025 13:41 β π 3 π 0 π¬ 0 π 0
The paper is here: arxiv.org/abs/2507.09683 This is joint work with @mkearnsphilly.bsky.social and Emily Ryu, from a fun visit in June.
24.07.2025 13:41 β π 2 π 1 π¬ 1 π 0
Nevertheless, we show that in any DAG information will be aggregated at the end of a path of length D up to excess error ~ 1/sqrt{D}. And there are distributions such that no DAG with depth D can support information aggregation to error < 1/D. So depth is a fundamental parameter.
24.07.2025 13:41 β π 0 π 0 π¬ 1 π 0
No! Even though y = x1 + x2, x1 is uninformative on its own, since it is independent of y. Hence A's predictor will be the constant function f_A(x1) = 0. This is all B sees, so learns nothing about x1. The best predictor he can learn from x2 has MSE 1/2 -- information was lost.
24.07.2025 13:41 β π 0 π 0 π¬ 1 π 0
An example to see why the problem is delicate/interesting: Suppose there are two agents, A and B, arranged in a line A->B. A observes x1 and B observes x2. x1 and y are iid standard Gaussians, and x2 = y-x1. So there is a perfect linear predictor: y = x1 + x2. Can B learn it?
24.07.2025 13:41 β π 0 π 0 π¬ 1 π 0
When is this enough for "information aggregation" to take place? i.e. when will someone in the network be able to produce a predictor that is as accurate as if they had trained a regression model on ALL features, even though nobody sees all features at once?
24.07.2025 13:41 β π 0 π 0 π¬ 1 π 0
Nobody sees other's observations, just their parents' predictions. They can use these as features (together with their direct observations) to solve a regression problem. Their own predictions will be observed by their children. This is how information spreads in the network.
24.07.2025 13:41 β π 0 π 0 π¬ 1 π 0
Suppose we have a regression problem that many people want to solve, but information is distributed across agents --- different agents see different subsets of features. And the agents are embedded in a network (DAG). When one makes predictions, they are observed by its children.
24.07.2025 13:41 β π 14 π 1 π¬ 1 π 0
Agentic LLM tooling like Windsurf is amazing. But you can already tell from the interface (you can pick from two dozen models from half a dozen different model providers, seamlessly) that AI is going to be a commodity. Not at all clear that LLM developers will get the surplus.
11.06.2025 23:40 β π 3 π 0 π¬ 0 π 0
Amazing set of hires, congratulations!
03.06.2025 22:22 β π 0 π 0 π¬ 0 π 0
Decimating (many times over!) scientific funding and (worse!) restricting international scientists from coming here will give up our enviable status as the world center for science and technology. Even if these policies are reversed 4 years down the line it might be too late. 5/5
31.05.2025 13:21 β π 5 π 0 π¬ 1 π 0
But for the same reason it was near impossible for anyone to overtake the US while we were the agreed upon meeting point for scientific talent, it will be near impossible to regain this status if the center moves elsewhere. This is why the current moment is so dangerous. 4/
31.05.2025 13:21 β π 0 π 0 π¬ 1 π 0
This is why for the last 75 years, the United States has dominated scientific discovery and why the world-beating tech companies were founded and headquartered in the US: this has been where the talent is. This was a huge source of American economic (and military) power. 3/
31.05.2025 13:21 β π 0 π 0 π¬ 1 π 0
This makes it hard for any other country to compete: while the US has been the global center for science, no other country has even been able to retain their -own- best students domestically, let alone attract the best students worldwide. 2/
31.05.2025 13:21 β π 0 π 0 π¬ 1 π 0
The United States has had a tremendous advantage in science and technology because it has been the consensus gathering point: the best students worldwide want to study and work in the US because that is where the best students are studying and working. 1/
31.05.2025 13:21 β π 11 π 1 π¬ 1 π 0
What is the complaint? It sounds like it is finding plenty of duplicates, just as it says on the box.
29.04.2025 17:41 β π 3 π 0 π¬ 0 π 0
In some sense the bigger deal is perhaps the lower bound --- I don't think it was previously known that you couldn't get bounds diminishing at a polynomial rate with epsilon even for high dimensions, though of course this was suspected given the lack of algorithms.
15.04.2025 22:32 β π 1 π 0 π¬ 1 π 0
A surprising new parameter regime, but right, not a pure improvement. An exponentially better dependence on the dimension at a cost of an exponentially worse dependence on the error parameter. It mirrors recent new swap regret algorithms (one by the same author).
15.04.2025 22:31 β π 1 π 0 π¬ 1 π 0
Referring to Godels incompleteness theorem in any context other than mathematical logic is also a red flag
09.04.2025 20:06 β π 16 π 0 π¬ 2 π 0
Thanks for the pointer, I'll check it out!
09.04.2025 12:45 β π 1 π 0 π¬ 0 π 0
The paper is here: arxiv.org/abs/2504.06075 and is joint work with the excellent @ncollina.bsky.social , @iraglobusharris.bsky.social , @surbhigoel.bsky.social , Varun Gupta, and Mirah Shi!
09.04.2025 12:11 β π 6 π 2 π¬ 0 π 0
Just as our last paper generalized Aumann's agreement theorem, this paper tractably generalizes "information aggregation" theorems for Bayesian reasoners. Our results lift back to the classic Bayesian setting and give the first distribution free information aggregation theorems.
09.04.2025 12:11 β π 5 π 0 π¬ 1 π 0
For any function classes H(A), H(B), H(J) satisfying this weak learning condition, we show how two parties can collaborate to be as accurate at H(J), by only needing to solve a small number of squared error regression problems on their own data over H(A) and H(B) respectively.
09.04.2025 12:11 β π 2 π 0 π¬ 1 π 0
MIT postdoc, incoming UIUC CS prof
katedonahue.me
Marketing & Economics Professor the Wharton School of @UPenn. Scholar of digital economy, technology, media. @NBER.org Fellow. www.pinaryildirim.com
Professor @ UVA Law School, writing about discrimination law and theory, bribery and corruption.
Assistant Professor of Machine Learning, Carnegie Mellon University (CMU)
Building a Natural Science of Intelligence π§ π€β¨
Prev: ICoN Postdoctoral Fellow @MIT, PhD @Stanford NeuroAILab
Personal Website: https://cs.cmu.edu/~anayebi
Chief Economics Correspondent for The New York Times. Adjunct at CUNY Newmark. Ex: FiveThirtyEight, WSJ. He/him.
Email: ben.casselman@nytimes.com
Signal: @bencasselman.96
πΈ: Earl Wilson/NYT
Physicist Turned Psychologist | Senior Researcher in #STEMed | Meta-Analysis Nerd | https://d-miller.github.io/
Also posts about π§ͺ science funding to focus my attention.
Personal account. I donβt speak for my employer or any other orgs.
Ecology, evolution, and social dynamics at the University of Pennsylvania. He/him
https://akcay.theoretical.bio
Assistant Professor of Operations Research and Statistics, MIT. Interested in social computing, human-in-the-loop of AI decision making, recommendation systems, and AI policy.
Econometrics, Statistics, Computational Economics, etc
http://donskerclass.github.io
π³οΈββ§οΈ
Scientist, Inventor, author of the NTQR Python package for AI safety through formal verification of unsupervised evaluations. On a mission to eliminate Majority Voting from AI systems. E Pluribus Unum.
Assistant Professor at University of Pennsylvania.
Robot Learning.
https://www.seas.upenn.edu/~dineshj/
The world's leading venue for collaborative research in theoretical computer science. Follow us at http://YouTube.com/SimonsInstitute.
Linguist, professor, Africanist, aspiring mad scientist/inventor/youtuber/occasional shitposter. Interested in meaning, sounds, the navigation therebetween, and how that all plays out for humans across history and geography.
Algorithms for Toddlers (https://youtu.be/nnLOi3ia210) | Algorithms for Teenagers (https://tinyurl.com/2cnp39cf) | Algorithms for Grown Ups (http://dblp.org/pid/11/10308)
CS Professor at UMass Amherst. Theoretical computer science, numerical linear algebra, machine learning.