MJ Yang's Avatar

MJ Yang

@mjyang.bsky.social

Research faculty at Leeds School of Business, CU-Boulder. https://mjyang.com/ Researching Scientific Entrepreneurship, Technological Uniqueness, and AI #scientificentrepreneurship, #strategy, #innovation

455 Followers  |  1,128 Following  |  157 Posts  |  Joined: 21.11.2024  |  2.0922

Latest posts by mjyang.bsky.social on Bluesky

Odd, or signal that ChatGPT-5 is lower quality, but cheaper to run, because it decides when a high-token test-time compute run is triggered..

09.08.2025 02:38 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

This also raises huge issues for replicability of research based on older ChatGPT models..

08.08.2025 18:02 β€” πŸ‘ 3    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Post image

Urg

07.08.2025 23:21 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Post image

Confirmed, geez

07.08.2025 23:08 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

There are two Bs!

07.08.2025 21:49 β€” πŸ‘ 82    πŸ” 12    πŸ’¬ 1    πŸ“Œ 0
Post image

ChatGPT-5 is here.

07.08.2025 19:05 β€” πŸ‘ 2    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Preview
Will data centers crash the economy? This time let's think about a financial crisis before it happens.

You should read @noahpinion.blogsky.venki.dev on the data centre investments and whether they will lead to a financial crisis. I think he downplays the risk. It is potentially very high. www.noahpinion.blog/p/will-data-...

03.08.2025 19:11 β€” πŸ‘ 9    πŸ” 5    πŸ’¬ 1    πŸ“Œ 0

Disruption of the traditional MBA is here, whether we want it or not..

26.07.2025 23:12 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

Consider: Google has a 90% market share in web and mobile searches. And has Gemini LLM results on top by default on any search. Are 1 trillion processed tokens really a sign of usefulness? (Same question about MSFT forcing LLMs on their developers..)

We see "revealed preference", just not for LLMs.

26.07.2025 21:32 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

I have been thinking about this a lot, lately, after finding the troubling productivity impacts of AI on older firms ( conference.nber.org/conf_papers/... ), combined with prior work showing the impacts of digitization on older workers (www.nber.org/papers/w28094).
#EconSky

23.07.2025 17:39 β€” πŸ‘ 2    πŸ” 1    πŸ’¬ 0    πŸ“Œ 0

Importantly, the negative productivity effects are concentrated at old manufacturing establishments and firms. We can narrow down the mechanism as well, by showing that roughly half of the productivity loss is driven by removal of (human-supporting) structured management practices.

17.07.2025 17:28 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

We use 2 separate datasets from the US Census Bureau and 3 separate identification strategies (matching, first-difference, IV) to establish credible causal effects. We find as of 2021 strongly negative productivity effects in the short-run and some evidence for potentially positive long-run effects.

17.07.2025 17:23 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Preview
SI 2025 Digital Economics and Artificial Intelligence

Naked self-promotion: My fantastic co-author @kmcelheran.bsky.social will present our empirical paper on "Industrial AI" (self-driving forklifts, self-optimizing production lines, autonomous quality control) and its effect on productivity in US manufacturing.

www.nber.org/conferences/...

17.07.2025 17:22 β€” πŸ‘ 3    πŸ” 1    πŸ’¬ 1    πŸ“Œ 0

Its not me, its the LLMs: in domains including coding, math, mapping, logic and now simple physics, they overfit and are unable to generalize. Still waiting for ChatGPT-5 btw..

14.07.2025 14:17 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

I agree that teachers (=experts) using LLMs has clear benefits. I think the evidence of benefits of LLMs for students is less consistent and more context dependent. LLMs are not a hoax, but also clearly over-hyped.

30.06.2025 13:56 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Preview
Potemkin Understanding in Large Language Models Large language models (LLMs) are regularly evaluated using benchmark datasets. But what justifies making inferences about an LLM's capabilities based on its answers to a curated set of questions? This...

Part of the avalanche of recent papers showing incoherence of LLMs: @sendhil.bsky.social and coauthors show that LLM answers contradict their reasoning in the majority of cases, among 7 current LLM/LRM models. Another datapoint on "LLM reasoning is just pretense".

arxiv.org/abs/2506.21521

29.06.2025 16:49 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Preview
Amazon.com : empire of ai

Finished "Empire of AI" by @karenhao.bsky.social: important & compelling read for anybody worried about the broader social consequences of LLMs, even if we never reach AGI. I was initially skeptical about the "empire/colonialism" framing but now appreciate it.

www.amazon.com/s?k=empire+o...

28.06.2025 19:10 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Experimental Evidence of the Effects of Large Language Models versus Web Search on Depth of Learning The effects of using large language models (LLMs) versus traditional web search on depth of learning are explored. Results from four online and laboratory exper

More evidence on the educational costs of LLMs: users underperform peers using traditional web-search.

papers.ssrn.com/sol3/papers....

28.06.2025 02:11 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

Yup, that OpEd reminds me of the recent MIT-brain study on LLM use..

25.06.2025 18:34 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

I guess it depends on how long "short term" is? Even if data center investments have broken even, the opportunity cost really should be measured by a long-run average of alternative investments at their social benefits..

25.06.2025 18:29 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

Very impressive study showing that LLM-use reduces brain activity for an essay writing task, over a 4-month time horizon. Letting college students use LLMs can significantly reduce educational benefits, not increase them.

19.06.2025 18:50 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

More and more, I'm thinking the answer is "no". However, that doesn't prevent employers to think that it can and for example not sure junior software developers anymore. More importantly, AI is sometimes cheaper than hiring entry-level skilled employees..

19.06.2025 14:13 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

At the same time, he says "AI is NOT a fad" and emphasizes that this (LLM-supported code generation) will "never go away". I agree, more generally for higher-ed than just software engineering.

19.06.2025 13:59 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

I like this guy's idea of "AI Hype Derangement Syndrome", and might refer to it. Ask yourself: If MSFT, META etc really could create 30% of their code through LLMs, why isn't there more high-quality LLM-created code in open source software? Or why is "AI-code" banned on many open-source platforms?

19.06.2025 13:38 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Generative AI and Code Security β€” recent developments as of June 2025
YouTube video by Spec Again: Reinventing Modern Software Careers Generative AI and Code Security β€” recent developments as of June 2025

LLM-generated code is highly likely to generate cybersecurity risks. Yet, a majority of IT professionals say they believe that LLM-generated code is more secure than human code. Maybe because MSFT and others claim "up to 30%" of their code is LLM-based? Wait a sec..

youtu.be/NxOehfOfB6c?...

19.06.2025 13:27 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Post image

P.S.: Is it telling that joke "rebuttals" of the Apple paper (as co-written by Alex Lawsen and Antropic's Claude) have gone viral as a real rebuttal? Or does this just show that social media users believe what they want to believe?

18.06.2025 16:08 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Preview
The Leaderboard Illusion Measuring progress is fundamental to the advancement of any scientific field. As benchmarks play an increasingly central role, they also grow more susceptible to distortion. Chatbot Arena has emerged ...

Sad to see, but true: LLM companies are systematically gaming leaderboards, to make LLMs seem smarter than they are. This is why the Apple paper from 2 weeks ago had so much impact: it is part of a string of work that shows the misleading claims of (leading) LLM companies.

arxiv.org/abs/2504.20879

18.06.2025 16:01 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

Having this reported by Sabine Hossenfelder (former professor of physics and now full-time YouTuber + QC skeptic), makes this news even more potentially credible. Although you should always take announcements by people who try to sell you something with a lot of skepticism!

17.06.2025 21:10 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Surprise Progress in Quantum Computing
YouTube video by Sabine Hossenfelder Surprise Progress in Quantum Computing

Interesting new developments on the Quantum Computing (QC) front: IBM reports that it developed new error-correction to cut the required physical qbits for one logical qbit by 90%. Commercially useful QCs might be with us by the end of the decade!

youtu.be/ZNQIF_q3NVA?...

17.06.2025 17:15 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Preview
A knockout blow for LLMs? LLM β€œreasoning” is so cooked they turned my name into a verb

Yup, "LLM reasoning" is just pretense. But OpenAI, Google and Anthropic still charge extra for these "frontier" models. My question: What will LLM companies come up next to try and convince us that AI scaling does not run into diminishing returns?

09.06.2025 00:55 β€” πŸ‘ 3    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

@mjyang is following 20 prominent accounts