Odd, or signal that ChatGPT-5 is lower quality, but cheaper to run, because it decides when a high-token test-time compute run is triggered..
09.08.2025 02:38 β π 0 π 0 π¬ 0 π 0@mjyang.bsky.social
Research faculty at Leeds School of Business, CU-Boulder. https://mjyang.com/ Researching Scientific Entrepreneurship, Technological Uniqueness, and AI #scientificentrepreneurship, #strategy, #innovation
Odd, or signal that ChatGPT-5 is lower quality, but cheaper to run, because it decides when a high-token test-time compute run is triggered..
09.08.2025 02:38 β π 0 π 0 π¬ 0 π 0This also raises huge issues for replicability of research based on older ChatGPT models..
08.08.2025 18:02 β π 3 π 0 π¬ 0 π 0Urg
07.08.2025 23:21 β π 0 π 0 π¬ 0 π 0Confirmed, geez
07.08.2025 23:08 β π 1 π 0 π¬ 0 π 0There are two Bs!
07.08.2025 21:49 β π 82 π 12 π¬ 1 π 0ChatGPT-5 is here.
07.08.2025 19:05 β π 2 π 0 π¬ 1 π 0You should read @noahpinion.blogsky.venki.dev on the data centre investments and whether they will lead to a financial crisis. I think he downplays the risk. It is potentially very high. www.noahpinion.blog/p/will-data-...
03.08.2025 19:11 β π 9 π 5 π¬ 1 π 0Disruption of the traditional MBA is here, whether we want it or not..
26.07.2025 23:12 β π 1 π 0 π¬ 0 π 0Consider: Google has a 90% market share in web and mobile searches. And has Gemini LLM results on top by default on any search. Are 1 trillion processed tokens really a sign of usefulness? (Same question about MSFT forcing LLMs on their developers..)
We see "revealed preference", just not for LLMs.
I have been thinking about this a lot, lately, after finding the troubling productivity impacts of AI on older firms ( conference.nber.org/conf_papers/... ), combined with prior work showing the impacts of digitization on older workers (www.nber.org/papers/w28094).
#EconSky
Importantly, the negative productivity effects are concentrated at old manufacturing establishments and firms. We can narrow down the mechanism as well, by showing that roughly half of the productivity loss is driven by removal of (human-supporting) structured management practices.
17.07.2025 17:28 β π 0 π 0 π¬ 0 π 0We use 2 separate datasets from the US Census Bureau and 3 separate identification strategies (matching, first-difference, IV) to establish credible causal effects. We find as of 2021 strongly negative productivity effects in the short-run and some evidence for potentially positive long-run effects.
17.07.2025 17:23 β π 0 π 0 π¬ 1 π 0Naked self-promotion: My fantastic co-author @kmcelheran.bsky.social will present our empirical paper on "Industrial AI" (self-driving forklifts, self-optimizing production lines, autonomous quality control) and its effect on productivity in US manufacturing.
www.nber.org/conferences/...
Its not me, its the LLMs: in domains including coding, math, mapping, logic and now simple physics, they overfit and are unable to generalize. Still waiting for ChatGPT-5 btw..
14.07.2025 14:17 β π 0 π 0 π¬ 0 π 0I agree that teachers (=experts) using LLMs has clear benefits. I think the evidence of benefits of LLMs for students is less consistent and more context dependent. LLMs are not a hoax, but also clearly over-hyped.
30.06.2025 13:56 β π 0 π 0 π¬ 0 π 0Part of the avalanche of recent papers showing incoherence of LLMs: @sendhil.bsky.social and coauthors show that LLM answers contradict their reasoning in the majority of cases, among 7 current LLM/LRM models. Another datapoint on "LLM reasoning is just pretense".
arxiv.org/abs/2506.21521
Finished "Empire of AI" by @karenhao.bsky.social: important & compelling read for anybody worried about the broader social consequences of LLMs, even if we never reach AGI. I was initially skeptical about the "empire/colonialism" framing but now appreciate it.
www.amazon.com/s?k=empire+o...
More evidence on the educational costs of LLMs: users underperform peers using traditional web-search.
papers.ssrn.com/sol3/papers....
Yup, that OpEd reminds me of the recent MIT-brain study on LLM use..
25.06.2025 18:34 β π 1 π 0 π¬ 0 π 0I guess it depends on how long "short term" is? Even if data center investments have broken even, the opportunity cost really should be measured by a long-run average of alternative investments at their social benefits..
25.06.2025 18:29 β π 0 π 0 π¬ 0 π 0Very impressive study showing that LLM-use reduces brain activity for an essay writing task, over a 4-month time horizon. Letting college students use LLMs can significantly reduce educational benefits, not increase them.
19.06.2025 18:50 β π 0 π 0 π¬ 0 π 0More and more, I'm thinking the answer is "no". However, that doesn't prevent employers to think that it can and for example not sure junior software developers anymore. More importantly, AI is sometimes cheaper than hiring entry-level skilled employees..
19.06.2025 14:13 β π 0 π 0 π¬ 0 π 0At the same time, he says "AI is NOT a fad" and emphasizes that this (LLM-supported code generation) will "never go away". I agree, more generally for higher-ed than just software engineering.
19.06.2025 13:59 β π 0 π 0 π¬ 0 π 0I like this guy's idea of "AI Hype Derangement Syndrome", and might refer to it. Ask yourself: If MSFT, META etc really could create 30% of their code through LLMs, why isn't there more high-quality LLM-created code in open source software? Or why is "AI-code" banned on many open-source platforms?
19.06.2025 13:38 β π 0 π 0 π¬ 1 π 0LLM-generated code is highly likely to generate cybersecurity risks. Yet, a majority of IT professionals say they believe that LLM-generated code is more secure than human code. Maybe because MSFT and others claim "up to 30%" of their code is LLM-based? Wait a sec..
youtu.be/NxOehfOfB6c?...
P.S.: Is it telling that joke "rebuttals" of the Apple paper (as co-written by Alex Lawsen and Antropic's Claude) have gone viral as a real rebuttal? Or does this just show that social media users believe what they want to believe?
18.06.2025 16:08 β π 0 π 0 π¬ 0 π 0Sad to see, but true: LLM companies are systematically gaming leaderboards, to make LLMs seem smarter than they are. This is why the Apple paper from 2 weeks ago had so much impact: it is part of a string of work that shows the misleading claims of (leading) LLM companies.
arxiv.org/abs/2504.20879
Having this reported by Sabine Hossenfelder (former professor of physics and now full-time YouTuber + QC skeptic), makes this news even more potentially credible. Although you should always take announcements by people who try to sell you something with a lot of skepticism!
17.06.2025 21:10 β π 0 π 0 π¬ 0 π 0Interesting new developments on the Quantum Computing (QC) front: IBM reports that it developed new error-correction to cut the required physical qbits for one logical qbit by 90%. Commercially useful QCs might be with us by the end of the decade!
youtu.be/ZNQIF_q3NVA?...