It worked.
27.01.2026 14:53 β π 1 π 0 π¬ 1 π 0@joshgans.bsky.social
Professor at University of Toronto
It worked.
27.01.2026 14:53 β π 1 π 0 π¬ 1 π 0Good policy direction from Brookings on how to deal with entry level job disruption that may come from AI. www.brookings.edu/articles/to-...
25.01.2026 22:40 β π 1 π 0 π¬ 0 π 1Excellent
24.01.2026 16:40 β π 1 π 0 π¬ 0 π 0Sure you have read short histories of innovation before but this one is the shortest.
22.01.2026 21:48 β π 1 π 0 π¬ 0 π 0The strategy of proposing to destroy the status quo and then claiming victory for restoring the status quo is an interesting one.
21.01.2026 21:17 β π 9 π 0 π¬ 2 π 0You've heard about the @MarkJCarney speech. You should listen to it all. It is one of great speeches of all time. We will talk of this for decades. www.youtube.com/watch?v=CQOr...
21.01.2026 11:36 β π 21 π 6 π¬ 1 π 0In strategic management and wanting to move to Canada. We have a senior slot this year in Strategic Management. Job ad and application details are here jobs.utoronto.ca/job/Toronto-...
19.01.2026 20:17 β π 3 π 0 π¬ 0 π 0Bottom line: Jagged intelligence isn't a bug that scaling will fixβit's a fundamental feature of how these models work. Better adoption strategies require either much denser knowledge coverage, effective calibration, or users developing expertise about where models fail.
19.01.2026 13:04 β π 2 π 0 π¬ 0 π 0For reasoning-intensive tasks, the model predicts that extended reasoning (like chain-of-thought) helps most in the largest knowledge gapsβexactly where users encounter problems most often. This explains why reasoning modes improve user experience disproportionately.
19.01.2026 13:04 β π 1 π 0 π¬ 1 π 0The paper shows that reducing variation in gap sizes (making performance more regular) is more valuable than just increasing average quality. A model with consistent mediocre performance can be more useful than one with erratic brilliance.
19.01.2026 13:04 β π 1 π 0 π¬ 1 π 0"Mastery" means learning where the model is reliable through repeated use. But this is harder than it sounds. The reliability landscape has complex patterns, and you need extensive experience across many tasks to build an accurate mental map.
19.01.2026 13:04 β π 2 π 0 π¬ 1 π 0Scaling up models (adding more knowledge points) improves average quality but doesn't eliminate jaggedness. You get denser coverage, but gaps remain. The fundamental unevenness persists even as models get larger and more capable.
19.01.2026 13:04 β π 1 π 0 π¬ 1 π 0Calibration helps but has limits. If users could see reliability scores for each specific task, they'd delegate more effectively. But even with perfect calibration, the inspection paradox means users still encounter errors at higher rates than the model's average.
19.01.2026 13:04 β π 1 π 0 π¬ 1 π 0The model shows that "blind adoption"βusing AI without checking local reliabilityβworks only when coverage is extremely dense. For most current models, blind adoption leads to unacceptable error rates even when average quality seems high.
19.01.2026 13:04 β π 1 π 0 π¬ 1 π 0This creates a threshold problem. Users adopt AI when global quality signals (like benchmark scores) look good enough. But the errors they actually experience are amplified by the inspection paradox, leading to disappointment and distrust.
19.01.2026 13:04 β π 2 π 0 π¬ 1 π 0Think of it like waiting for a bus. Longer gaps between buses mean you spend more time waiting in those gaps. Similarly, you encounter AI's biggest knowledge gaps most frequentlyβmaking the model seem worse than its average performance.
19.01.2026 13:04 β π 1 π 0 π¬ 1 π 0Here's the key insight: users experience AI weaknesses more than its strengths due to an "inspection paradox." When you try many tasks, you're statistically more likely to hit the large gaps in knowledge rather than the small ones.
19.01.2026 13:04 β π 2 π 0 π¬ 1 π 0The paper models AI knowledge as scattered points on a landscape. Between these knowledge points, the AI extrapolatesβessentially guessing based on nearby information. The further you get from a knowledge point, the less reliable the output becomes.
19.01.2026 13:04 β π 1 π 0 π¬ 1 π 0Generative AI shows a puzzling pattern: it excels at some tasks but fails at seemingly similar ones. This "jagged" performance creates a fundamental problem for users trying to decide when to rely on AI. New research explains why this happens and what it means.
19.01.2026 13:04 β π 2 π 1 π¬ 1 π 0New paper out today. A thread. www.nber.org/papers/w3471...
19.01.2026 13:04 β π 8 π 1 π¬ 1 π 1I think that was its greatest episode.
18.01.2026 12:22 β π 1 π 0 π¬ 0 π 0In light of the current global situation, I believed that AI could be used to create a playable simulation that allows you to map out the nuances of the military and other trade-offs associated with a US invasion of Greenland. You can play it here claude.ai/public/artif...
17.01.2026 21:08 β π 2 π 0 π¬ 0 π 0Prediction Machines
16.01.2026 22:30 β π 1 π 0 π¬ 0 π 0This is incredible.
(More legible version here www.npr.org/sections/the...)
Americans. Sigh. www.nytimes.com/live/2026/01...
16.01.2026 11:35 β π 0 π 0 π¬ 0 π 0I'm with Ben Edelman here. The lawsuit documents appear to focus on what was odd in HBS' review of his candidacy. Not sure you can learn about your own tenure process from reading this. (And always better to write something or teach someone than to worry about the process.)
15.01.2026 19:13 β π 5 π 1 π¬ 0 π 0So junior faculty are always wanting more information on the tenure process, and we usually canβt give them more than advice. Whatever else is happening in Edelman v. Harvard, the documents are a window on this. www.edelman-v-harvard.org/documents/
14.01.2026 15:19 β π 3 π 0 π¬ 1 π 1I am pleased to announce that the OPEN ACCESS link for The Microeconomics of AI is now available. Click here to download for free. direct.mit.edu/books/oa-mon...
14.01.2026 14:18 β π 20 π 7 π¬ 1 π 1Anyhow, here is the paper. It doesn't involve any fancy techniques. Indeed, it could have easily been written in the 1950s. www.nber.org/papers/w34669
12.01.2026 13:18 β π 1 π 0 π¬ 0 π 0Is this just a theoretical curiosity? It certainly is that. But I think that a range of different prices based on income-related factors does arise in things like car sales and pharmaceuticals (especially internationally).
12.01.2026 13:18 β π 1 π 0 π¬ 1 π 0