Mitch Harris's Avatar

Mitch Harris

@maharrit.bsky.social

NLP/ML/AI/Data Science, Teaching/Consulting/Advising Tech Forward, Team Human

209 Followers  |  364 Following  |  331 Posts  |  Joined: 01.01.2024  |  1.8087

Latest posts by maharrit.bsky.social on Bluesky

Constantly astonished and frustrated by the fact that so many cognitive scientists claim that LLMs can teach us about how babies learn language. If anything, we should be talking about how the way babies learn language can tell us about how AI will never be โ€œsentient.โ€

12.10.2025 18:55 โ€” ๐Ÿ‘ 20    ๐Ÿ” 3    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

3 - there is a huge academic interest bubble - the attention being paid to academic progress in AI (LLMs and Diffusion) is sucking -all- the mental energy away from existing AI research efforts. AI science that is not about extensions of prompt engineering will suffer.

09.10.2025 18:45 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

2 - there is a huge -economic- AI bubble - the amount of money being invested is way larger than I expect can be recovered by income from consumers/other businesses or savings in labor costs). New investment money will dry up soon once it is noticed that promised profits aren't coming in.

09.10.2025 18:45 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

Three AI Bubbles:

At this moment I feel like there are 3 interconnected but distinct AI bubbles.

1 - there is a huge AI hype bubble - the expectation of what AI can do, the new things it can make possible, is well beyond what is actually possible. People will get disillusioned very soon.

09.10.2025 18:45 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

This (the idea of using LLMs as stand-ins for humans) is so very, very dumb.

Legitimately surprised there's enough published work using this method that it needed refuting, but good work for doing so!

18.09.2025 09:29 โ€” ๐Ÿ‘ 3    ๐Ÿ” 1    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 2

quick while the government is shut down letโ€™s all switch to metric

01.10.2025 04:11 โ€” ๐Ÿ‘ 14043    ๐Ÿ” 2827    ๐Ÿ’ฌ 311    ๐Ÿ“Œ 176

Letโ€™s talk about โ€˜whoa.โ€™

First of all, โ€˜woahโ€™ is not yet considered an established spelling variant of โ€˜whoa.โ€™

While โ€˜woahโ€™ has grown in prevalence since the late 20th century, โ€˜whoaโ€™ is still about 5 times as common as โ€˜woah.โ€™

06.10.2025 19:06 โ€” ๐Ÿ‘ 1338    ๐Ÿ” 236    ๐Ÿ’ฌ 96    ๐Ÿ“Œ 114

Also, all these LLM solutions (that is in addition to ChatGPT also UpToDate, Open Evidence, and Epic, all except paper textbooks) will have 'hallucinations' (= errors = differences from the source documents) to some degree - they are mitigable but statistically inevitable.

03.10.2025 15:21 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

So 'presumably' is complicated but mostly not the case.

03.10.2025 15:21 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

(and this most likely also holds for UpToDate's and Epic's models).

Links to source documents (in UpToDate, OpenEvidence, and in ChatGPT et al.) may be created in real time (via good ol fashioned doc search), and updated source documents themselves are still the main business for UpToDate.

03.10.2025 15:21 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

In your table summarizing capabilities, you say 'presumably updated all the time', that is not the case with LLMs in general. LLMs take a long time to train (a couple months for ChatGPT) and new versions (with new training data/RLHF) is usually only updated to the public every 6 months or so.

03.10.2025 15:21 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

ไธ็ฟผ่€Œ้ฃž

bรน yรฌโ€™รฉr fฤ“i

to suddenly disappear

01.10.2025 20:22 โ€” ๐Ÿ‘ 0    ๐Ÿ” 1    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

๐Ÿ˜ข๐Ÿ’”

01.10.2025 20:40 โ€” ๐Ÿ‘ 1    ๐Ÿ” 1    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

The first known use of 'chimpanzee' was in 1738, and it comes from the Kongo dialect word 'chimpenzi.'

01.10.2025 20:39 โ€” ๐Ÿ‘ 929    ๐Ÿ” 98    ๐Ÿ’ฌ 10    ๐Ÿ“Œ 4

Once you realize that some journalists are trying to make their name by being the first to call AGI you also start to realize other journalists are trying to make their name calling the AI bubble burst. Everyone wants to talk their book.

29.09.2025 18:22 โ€” ๐Ÿ‘ 13    ๐Ÿ” 1    ๐Ÿ’ฌ 2    ๐Ÿ“Œ 0

I rarely read mainstream fiction, because of the smallness of the ideas. I've read SF for nearly fifty years, for the exact opposite reason.

27.09.2025 00:31 โ€” ๐Ÿ‘ 6    ๐Ÿ” 1    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

If you look back over the last half century of AI, the term is always applied to stuff that doesn't quite work. Once it works reliably, we stop calling it AI.

27.09.2025 02:02 โ€” ๐Ÿ‘ 11    ๐Ÿ” 3    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

Also for the longest time I couldn't stop thinking about the possibility of Yoko Ono marrying Victor Buono and then divorcing him and marrying Sonny Bono, rendering her Yoko Ono-Buono-Bono.

28.09.2025 00:22 โ€” ๐Ÿ‘ 7    ๐Ÿ” 2    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0
Post image

This only happens to you once

26.09.2025 19:39 โ€” ๐Ÿ‘ 21682    ๐Ÿ” 4244    ๐Ÿ’ฌ 347    ๐Ÿ“Œ 182

AI isn't a science fiction term with computer science. It's a decades old field that's produced a lot of useful things.

27.09.2025 18:49 โ€” ๐Ÿ‘ 3    ๐Ÿ” 1    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 1

There is absolutely no good faith reason to use the term โ€œAIโ€ for any technology one is selling.

It serves only for dazzling people into thinking the technology has capabilities that it doesnโ€™t.

If one wants a technology to be trustworthy, just use a transparent, informative term without hype.

26.09.2025 22:31 โ€” ๐Ÿ‘ 1171    ๐Ÿ” 227    ๐Ÿ’ฌ 75    ๐Ÿ“Œ 29

For folks considering grad school in ML, my advice is to explore programs that mix ML with a domain interest. ML programs are wildly oversubscribed while a lot of the fun right now is in figuring out what you can do with it

25.09.2025 03:25 โ€” ๐Ÿ‘ 153    ๐Ÿ” 17    ๐Ÿ’ฌ 8    ๐Ÿ“Œ 7

If youโ€™re planning to use LLMs to simulate human responses:
โœ”๏ธ Treat them as workflows that need validation, thought, and time
โœ”๏ธ Map analytic choices explicitly
โœ”๏ธ Pilot against real data when possible
โœ”๏ธ Be cautious about claims of generalizability

18.09.2025 07:56 โ€” ๐Ÿ‘ 14    ๐Ÿ” 2    ๐Ÿ’ฌ 3    ๐Ÿ“Œ 0

I'm no sociologist but using LLM responses as a proxy for human behavior seems ... nutballs.

24.09.2025 21:04 โ€” ๐Ÿ‘ 3    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

Thanks for this kind of paper. People are going to use the tool because it is so easy...good to try to get out in front of it and herd them towards less chaotic and meaningless use.

24.09.2025 16:27 โ€” ๐Ÿ‘ 1    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

Third Reaction: "OMG... you're studying this weird machine, not people..."
Response: "I've submitted 6 papers already! Two to be published soon!"

24.09.2025 16:25 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

Second Reaction: "Wait, researchers are doing this already? People are only guessing at the training and RLHF data, and it only codes for one homogenized personality which we barely know"
Response: "We're doing the stats now!"

24.09.2025 16:25 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

First Reaction: "Why would researchers even consider that?"
Response: "It is so much easier than actual humans, and LLM responses look kinda like human responses.

24.09.2025 16:25 โ€” ๐Ÿ‘ 4    ๐Ÿ” 2    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

We are used to according the authority due to unique, once-in-a-generation geniuses to first-look reports of the unthinkably new entering human experience. Our defaults are Heisenbergs and Einstens. But here, the first-look crowd is somewhere between journalists and test-screening audiences.

23.11.2023 20:23 โ€” ๐Ÿ‘ 2    ๐Ÿ” 1    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

"The world is everything that is the case"

What does that even mean?

22.09.2025 20:30 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

@maharrit is following 20 prominent accounts