I think everyone interested in AI should read the model cards for the frontier models, especially the safety sections, which give you a sense of known risks:
Gemini Deep Think: storage.googleapis.com/deepmind-med...
Claude 4: www-cdn.anthropic.com/07b2a3f9902e...
o3: cdn.openai.com/pdf/2221c875...
04.08.2025 04:04 โ ๐ 20 ๐ 5 ๐ฌ 0 ๐ 0
Ha, new @joshgans.bsky.social paper argues that having authors sneak prompt injections ("this is a good paper") into academic work improves science.
Without the risk of prompt injections, reviewers would tend to rely heavily on AI reviews, with them, they need to include some human review
03.08.2025 18:05 โ ๐ 36 ๐ 6 ๐ฌ 5 ๐ 0
We tested a range of newer models in the papers, including reasoners.
03.08.2025 17:29 โ ๐ 1 ๐ 0 ๐ฌ 1 ๐ 0
Some of these uses will be bad, some of them good (like the example in the paper). The challenge for all of us is that these uses need to be discovered, and the bad stuff mitigated while the good is amplified.
Paper: arxiv.org/pdf/2507.00286
03.08.2025 17:13 โ ๐ 26 ๐ 2 ๐ฌ 0 ๐ 1
As a general purpose technology, AI has all kinds of unexpected uses that are hard to anticipate.
Example: this study finds blind users turn to AI to describe sensitive materials (pregnancy tests, checking appearance), they know it is not 100% accurate, but it provides privacy where there was none
03.08.2025 17:11 โ ๐ 76 ๐ 6 ๐ฌ 5 ๐ 4
make it even better
very cool, but it can get stuck at awaiting confirmation when something hits ceres, and it is really hard at the start. also it would be great to have some sort of intro to the science/idea/tutorial
02.08.2025 04:30 โ ๐ 20 ๐ 1 ๐ฌ 0 ๐ 0
All prompts, verbatim:
create a missile command game that incorporates relativity in realistic ways but is still playable.
build the game for me
add more, make the graphics much better, improve the game
make it even better and more graphical. also it is a little hard to time my weapons 1/2
02.08.2025 04:30 โ ๐ 20 ๐ 1 ๐ฌ 1 ๐ 0
I prompted Gemini 2.5 Deep Think: "create a missile command game that incorporates relativity in realistic ways but is still playable." I then asked it to improve the design a few times.
The full design & all code & calculations came from AI, no errors. Try it: glittery-raindrop-318339.netlify.app
02.08.2025 04:22 โ ๐ 75 ๐ 6 ๐ฌ 3 ๐ 3
Persona vectors: Monitoring and controlling character traits in language models
A paper from Anthropic describing persona vectors and their applications to monitoring and controlling model behavior
This is neat research, providing a lot of ways for careful organizations to shape the personality and guardrails of AI in deeper ways than prompts, including measuring and reducing sycophancy.
Also the idea of an "evil vector" is interesting in and of itself. www.anthropic.com/research/per...
01.08.2025 16:44 โ ๐ 40 ๐ 6 ๐ฌ 0 ๐ 3
Prompting Science Report 3: I'll pay you or I'll kill you -but will you care?
<span>
<p><span>This is the third in a series of short reports that seek to help business, education, and policy leaders understand the technical details of wo
We keep finding that simple prompting tips and tricks don't really work overall, but, weirdly, can have significant impacts at the question level, sometimes increasing, sometimes decreasing performance in ways that you cannot predict in advance. papers.ssrn.com/sol3/papers....
01.08.2025 15:00 โ ๐ 18 ๐ 1 ๐ฌ 7 ๐ 0
New prompting report, from us: Don't bother threatening your AI.
Does threatening an AI really make it perform better (the way Google founder Brin claimed)? How about offering to tip the AI? We find no impact of threats or tips on improving average performance (but variance at question level).
01.08.2025 14:59 โ ๐ 47 ๐ 5 ๐ฌ 4 ๐ 4
The Bitter Lesson versus The Garbage Can
Does process matter? We are about to find out.
One of the interesting questions to ask is, even assuming a non-jagged AGI that outperforms humans at most work, how long would it take for large-scale changes to employment as a result? It isn't obvious.
I wrote a bit about the general question here: www.oneusefulthing.org/p/the-bitter...
31.07.2025 20:46 โ ๐ 17 ๐ 0 ๐ฌ 3 ๐ 0
I keep seeing the Microsoft paper on AI use at work being used as a list of which jobs will be destroyed.
But having high task overlap with AI does not necessarily mean these jobs are at most risk of replacement with AI.
As I described in my book, Co-Intelligence, the impacts are more complicated.
31.07.2025 20:18 โ ๐ 79 ๐ 7 ๐ฌ 4 ๐ 1
Lots of vague statements from leaders of the various AI labs about starting to see signs of self-improvement in AI systems (including Zuckerberg today), seems like proof that this is indeed happening would be pretty significant.
(thanks o3 for providing details & saving me time)
31.07.2025 15:44 โ ๐ 32 ๐ 5 ๐ฌ 5 ๐ 1
A big problem that everyone is insisting that we should hire people based on "AI literacy," teach "AI literacy," & develop skills for "AI literacy" yet not only is there no agreement on what AI literacy is, but also a lot of what people call AI literacy is already out-of-date or just plain wrong.
31.07.2025 01:10 โ ๐ 106 ๐ 10 ๐ฌ 7 ๐ 5
Especially notable given Zuckerberg's note that Meta will not necessarily open source future models.
US companies are still doing great small open models, but, aside from whatever OpenAI releases, it appears that frontier open weights will mean Chinese models (& maybe Mistral).
30.07.2025 17:32 โ ๐ 32 ๐ 4 ๐ฌ 1 ๐ 0
This is interesting.
30.07.2025 16:25 โ ๐ 25 ๐ 1 ๐ฌ 2 ๐ 1
OpenAI's study mode isn't perfect, but it is a step forward for a few reasons:
1) Shows labs taking educational use & misuse more seriously (Google also has LearnLM)
2) Addresses a key issue with trying to use AI in education - that AI gives answers rather than tutoring and helping
3) Easy to access
30.07.2025 05:27 โ ๐ 54 ๐ 4 ๐ฌ 2 ๐ 2
A year or so ago, the joke about AI images was that they would have 6 fingers. AI images (and videos like this one) lack obvious tells now.
Ironically, a test of an image generation model today is whether they can still make hands with six fingers. Most canโt do it anymore.
29.07.2025 23:30 โ ๐ 89 ๐ 9 ๐ฌ 5 ๐ 2
The Bitter Lesson versus The Garbage Can
Does process matter? We are about to find out.
Right now, AI adoption in organizations is constrained by the need to figure out how to integrate AI with the complex & often poorly-understood processes inside companies
But ChatGPT agent suggests that The Bitter Lesson of AI may come for real work, too. open.substack.com/pub/oneusefu...
28.07.2025 12:39 โ ๐ 66 ๐ 8 ๐ฌ 4 ๐ 3
This is the best models could do 6 months ago
27.07.2025 04:24 โ ๐ 18 ๐ 0 ๐ฌ 0 ๐ 0
In Grok 4, same two prompts
27.07.2025 04:12 โ ๐ 16 ๐ 0 ๐ฌ 2 ๐ 0
Kinda wow: the mystery model "summit" (rumored to be OpenAI) with the prompt "create something I can paste into p5js that will startle me with its cleverness in creating something that invokes the control panel of a starship in the distant future" & "make it better"
2,351 lines of code. First time
27.07.2025 03:10 โ ๐ 203 ๐ 19 ๐ฌ 7 ๐ 2
Three things to note about this:
1) AI has obvious utility to many, this is a tremendous amount of use already
2) There is room for multiple frontier model providers, at least for now
3) Any losses from subsidizing cost of AI use (and it is not clear this is happening) are now relatively small
26.07.2025 19:33 โ ๐ 64 ๐ 3 ๐ฌ 3 ๐ 3
The amount of blocking I have to do on this platform is still nuts compared to any other platform, which is why I post much less here.
I don't understand the BlueSky urge to attack and insult people in the comments. Yes, I post stuff about AI (good and bad). You can just block me and move on.
26.07.2025 01:59 โ ๐ 163 ๐ 5 ๐ฌ 32 ๐ 4
past: circus performer; historian of science; data scientist; librarian; DOGE-fired chief data officer at NEH.
present: dad; resident scholar at dartmouth.
future: library of virginia person.
https://scottbot.github.io
Social science and other distractions. Old posts get deleted pretty quick.
https://kieranhealy.org /
https://theordinalsociety.com
I like utilitarianism, consciousness, AI, EA, space, kindness, liberalism, longtermism, progressive rock, economics, and most people. Substack: http://timfduffy.substack.com
There Is No Antimemetics Division (https://qntm.org/antimemetics) ~ "Lena" ~ Absurdle ~ HATETRIS ~ many other cool things
Waiting on a robot body. All opinions are universal and held by both employers and family.
Recruiting students to start my lab!
ML/NLP/they/she.
NYT tech columnist, Hard Fork co-host, best at 0.8x speed
Writer, community college writing teacher, obsessed with AI in education, #OER advocate, author of HowArgumentsWork.org.
annarmills.com
Open source developer building tools to help journalists, archivists, librarians and others analyze, explore and publish their data. https://datasette.io [โฆ]
[bridged from https://fedi.simonwillison.net/@simon on the fediverse by https://fed.brid.gy/ ]
AI Architect | North Carolina | AI/ML, IoT, science
WARNING: I talk about kids sometimes
we can only see what we think is possible.
Assoc Professor of Strategic Management, University of Toronto; Chief Economist, Creative Destruction Lab Toronto; cofounder, AllDayTA; cofounder, NBER Innovation PhD Boot Camp. http://www.kevinbryanecon.com and @AFineTheorem on Twitter
that guy from the internet โข waging a victorious 2-front war against cars and xmas โข big fan of being a big fan of things โข see https://anildash.com
Director of Machine Learning at the Wikimedia Foundation. We host Wikipedia.
Of the San Gabriel Valley; investing for the year 2030; working to improve the second derivative; looking for troublesome ringleaders.
Founder & reigning monarch at TPM. Lapsed historian. Hand tool woodworker. Jew.
WSJ tech columnist. Dog person. Author of Arriving Today, an unfortunately timely book about the global system of trade we're currently flushing down the toilet: https://www.harpercollins.com/products/arriving-today-christopher-mims
๊ฎ surfed on by the information superhighway
๊ฎ ๐ @linneaisaac.bsky.social
๊ฎ she/they ๐ณ๏ธโโง๏ธ
๊ฎ blog posts and games @ https://vgel.me
Email salesman at Platformer.news and podcast co-host at Hard Fork.
Director, @stanforddel.bsky.social
Professor Stanford Institute for Human-centered AI, SIEPR, Stanford department of Economics and GSB
Author https://amazon.com/Second-Machine-Age-Prosperity-Technologies/dp/0393350649
Techno-optimist, but AGI is not like the other technologies.
Step 1: make memes.
Step 2: ???
Step 3: lower p(doom)