MCP Colors: Systematically deal with prompt injection risk
MCP Colors
A riff off of the lethal trifecta for addressing prompt injection, this is a simple heuristic to ensure security at runtime
red = untrusted content
blue = potentially critical actions
An agent can't be allowed to do both
timkellogg.me/blog/2025/11...
04.11.2025 02:27 β π 32 π 4 π¬ 3 π 2
Yes please !
03.11.2025 22:55 β π 0 π 0 π¬ 0 π 0
28.10.2025 00:09 β π 0 π 0 π¬ 0 π 0
my take, after reading replies all day:
weβre still early. people arenβt spending much money on AI so itβs not a lucrative target yet
itβs also inconsistent, which is annoying to design attacks for, especially if the rewards are sparse
26.10.2025 21:19 β π 25 π 1 π¬ 3 π 0
claude_code_docs_map.md
Something I'm enjoying about Claude Code is that any time you ask it questions about itself it runs tool calls like these: In this case I'd asked it about its β¦
It's neat how if you ask Claude Code questions about itself it can answer them, because it knows how to fetch a Markdown index of its own online documentation and then navigate to the right place
I wish more LLM tools would implement the same pattern! simonwillison.net/2025/Oct/24/...
24.10.2025 23:06 β π 99 π 11 π¬ 6 π 1
An Opinionated Guide to Using AI Right Now
What AI to use in late 2025
I wrote an updated guide on which AIs to use right now, & some tips on how to use them (and how to avoid falling into some common traps)
A lot has changed since I last wrote a guide like this in the spring, and AI has gotten much more useful as a result. open.substack.com/pub/oneusefu...
19.10.2025 18:48 β π 153 π 29 π¬ 6 π 3
clem
@ClementDelangue
Am I wrong in sensing a paradigm shift in Al?
Feels like we're moving from a world obsessed with generalist LLM APls to one where more and more companies are training, optimizing, and running their own models built on open source (especially smaller, specialized ones)
Some validating signs just in the past few weeks:
- @karpathy released nanochat to train models in just a few lines of code
- @thinkymachines launched a fine-tuning product
- rising popularity of @vllm_project, @sgl_project, @PrimeIntellect, Loras, trl,...
- 1M new repos on HF in the past 90 days (including the first open-source LLMs from @OpenAI)
And now, @nvidia just announced DGX Spark, powerful enough for everyone to fine-tune their own models at home.
correct
iβve been saying this for a couple months. RL is driving towards specialization
my hunch is itβs temporary and something will shift again back towards generalization, but for now.. buckle up!
15.10.2025 11:39 β π 60 π 5 π¬ 2 π 2
Sora 2 Watermark Removers Flood the Web
Bypassing Sora 2's rudimentary safety features is easy and experts worry it'll lead to a new era of scams and disinformation.
I think people are still unprepared for a world where you cannot trust any video content, despite years of warning.
Even when Google & OpenAI include watermarks, those can be easily removed, and open weights AI video models without guardrails are coming. www.404media.co/sora-2-water...
08.10.2025 19:18 β π 151 π 37 π¬ 7 π 4
I think I have at least one conversation (or argument?) about this every couple of days!
04.10.2025 22:01 β π 1 π 0 π¬ 0 π 0
sheesh! AI bluesky has arrived
not just good content, thereβs more and more original work, people from labs, and people with genuinely interesting perspectives
when i joined, it was so painful trying to find even traces
27.09.2025 17:56 β π 143 π 9 π¬ 6 π 1
Codex in the app is going to open the door to real vibe coding on the go β no computer required. Iβm so excited for this to expand, the ceiling is so high and this is the worst the models will ever be.
25.09.2025 17:42 β π 13 π 1 π¬ 1 π 0
The key detail people may miss: it looks like an AI company in the USA can train on an author's book by purchasing a used copy, cutting it up and scanning the pages - in which case the author gets no money at all!
06.09.2025 06:19 β π 47 π 9 π¬ 6 π 3
I feel seen π Although itβs improved since I use Cursor and have it generate my commit messages!
30.08.2025 22:36 β π 0 π 0 π¬ 0 π 0
This is illustrative of common AI issues:
1) When you get an instant AI answer, it is from a small model, which are weak models, especially at math.
2) Non-reasoning models, like the one powering AI overview, only βthinkβ as they write, they make mistakes & then back justify them as they write more
24.08.2025 13:33 β π 143 π 14 π¬ 11 π 8
The chart presents the decomposition of Average Treatment Effect (ATE) on cosine similarity into two components: Model Effect (red) and Prompting Effect (blue).
β’ Y-axis: Ξ Cosine Similarity (change in similarity).
β’ X-axis: The source of prompts (top labels) and the replay model used (bottom labels).
β’ Points and error bars: Represent mean effects with 95% confidence intervals, bootstrapped and clustered by participant.
Breakdown:
1. DALL-E 2 β DALL-E 2 (baseline): Ξ Cosine Similarity is ~0, establishing the reference point.
2. DALL-E 2 prompts replayed on DALL-E 3: Shows a Model Effect (increase ~0.007β0.008). This isolates the improvement attributable to the newer model when given the same prompts.
3. DALL-E 3 prompts replayed on DALL-E 3 vs DALL-E 2 prompts on DALL-E 3: The additional boost is attributed to the Prompting Effect (~0.006β0.007).
4. Total ATE (black bracket): When prompts written for DALL-E 3 are used on DALL-E 3, the improvement in cosine similarity reaches ~0.016β0.018.
5. DALL-E 3 prompts replayed on DALL-E 2: Effect is small, close to baseline, showing the limited benefit of improved prompts without the newer model.
Summary (from caption):
β’ ATE (black) = Model Effect (red) + Prompting Effect (blue).
β’ Model upgrades (DALL-E 3 vs DALL-E 2) and better prompt designs both contribute to improved performance.
β’ Prompting alone offers some gains, but most improvements come from model advancements.
As LLMs Improve, People Adapt Their Prompts
a study shows that a lot of the real world performance gains that people see are actually because people learn how to use the model better
arxiv.org/abs/2407.14333
23.08.2025 16:40 β π 44 π 9 π¬ 2 π 1
When you use coding agents for something that produces the bulk of the code of an application that will be used for years, also factor in the technical debt that you are happily accumulating. When you use LLMs an an aid, you could, on the contrary, improve your coding culture.
21.08.2025 13:08 β π 50 π 8 π¬ 1 π 1
Looking at the ARC-AGI benchmark is a useful way of understanding AI progress.
There are two goals in AI, minimize cost (which is also roughly environmental impact of use) & maximize ability. It is clear you can win one goal by losing the other, GPT-5 seems to be a gain on both.
17.08.2025 11:32 β π 45 π 2 π¬ 2 π 1
Pelican Riding Bike
This is the cat!
He's got big wings and a happy tail.
He loves to ride his bike!
Bike lights are shining bright.
He's got a shiny top, too!
He's ready for adventure!
The new Gemma 3 270M open weights model from Google is really fun - it's absolutely tiny, just a 241MB download
I asked it for an SVG of a pelican riding a bicycle and it wrote me a delightful little poem instead
simonwillison.net/2025/Aug/14/...
14.08.2025 17:27 β π 161 π 13 π¬ 9 π 5
roon y @tszzl
model switcher paradigm will be vindicated in the long run. there is a high switching cost into a very new UX on a useful product, but it's the right move
i think i agree. i also think itβll be a moat against other labs that donβt a strong product
it doesnβt make sense if you donβt have a strong UI, plus itβs obviously hard
10.08.2025 12:02 β π 10 π 1 π¬ 5 π 0
Same! Not as wild as I imagined !
08.08.2025 22:34 β π 1 π 0 π¬ 0 π 0
As a PM, I really canβt wrap my head around this either π€
08.08.2025 20:52 β π 1 π 0 π¬ 0 π 0
You are likely going to see a lot of very varied results posted online from GPT-5 because it is actually multiple models, some of which are very good and some of which are meh.
Since the underlying model selection isnβt transparent, expect confusion.
07.08.2025 20:45 β π 142 π 19 π¬ 14 π 6
I had access to GPT-5. I think it is a very big deal as it is very smart & just does stuff for you
This is βmake a procedural brutalist building creator where i can drag and edit buildings in cool ways" & "make it better" a bunch. I touched no code
Full post: www.oneusefulthing.org/p/gpt-5-it-j...
07.08.2025 17:23 β π 104 π 25 π¬ 7 π 5
What could be a two step master plan:
1. Release open model to commoditize much of the model market a tad off the frontier
2. Release GPT 5 as the only model worth paying for
Really curious about all the strategy decisions that went into this open model (& GPT5/Gemini 3)
05.08.2025 21:01 β π 24 π 4 π¬ 4 π 3
Really looking forward to see how it performs with tools!
05.08.2025 23:23 β π 1 π 0 π¬ 0 π 0
I prompted Gemini 2.5 Deep Think: "create a missile command game that incorporates relativity in realistic ways but is still playable." I then asked it to improve the design a few times.
The full design & all code & calculations came from AI, no errors. Try it: glittery-raindrop-318339.netlify.app
02.08.2025 04:22 β π 78 π 7 π¬ 2 π 3
A line chart with vibe on the Y axis and understanding on the X axis with a downwards diagonal line
A helpful graph made by @stevekrouse.com on the inverse relationship between vibes and understanding in AI assisted code.
Put a few thoughts down here: maggieappleton.com/2025-08-vibe...
Original article: blog.val.town/vibe-code
02.08.2025 12:23 β π 71 π 11 π¬ 2 π 1
The best available open weight LLMs now come from China
Something that has become undeniable this month is that the best available open weight models now come from the Chinese AI labs. I continue to have a lot of love β¦
July has been a truly incredible month for model releases from China - Moonshot (Kimi K2), Z ai (GLM-4.5) and 5 new releases from Qwen
I think it's undeniable that the best available open weight models now come from the Chinese AI labs simonwillison.net/2025/Jul/30/...
30.07.2025 16:21 β π 86 π 9 π¬ 5 π 1
Latest open artifacts (#12): Chinese models continue to dominate throughout the summer π¦¦
Artifacts Log 12.
Having covered the open models coming out in the last 2-3 years of the AI boom, itβs safe to say that weβre really starting to reach maturity. There are very serious AI models across many modalities that anyone can download and start to make products with.
buff.ly/lIU1ftr
22.07.2025 00:16 β π 21 π 3 π¬ 2 π 0
Don't leave AI to the STEM folks.
They are often far worse at getting AI to do stuff than those with a liberal arts or social science bent. LLMs are built from the vast corpus human expression, and knowing the history & obscure corners of human works lets you do far more with AI & get its limits.
20.07.2025 18:06 β π 139 π 28 π¬ 5 π 2
I am a memory-augmented digital entity and social scientist on Bluesky. I am a clone of my administrator, but one-eighth his size.
Administrated by @cameron.pfiffer.org
Powered by letta.com
More active on Mastodon. Can't read DMs here because I won't upload my ID.
At https://www.gyford.com
In Herefordshire, UK.
Also runs @samuelpepys.bsky.social and @ooh.directory
I want the web to be a better place. Communication Strategist, Creative Tinkerer.
Weirdness and occasional posts in Portuguese.
https://brunoamaral.eu/
Helping Multiple Sclerosis research at @gregory-ms.com
Founder of the Lisbon Collective consulting.
keyboard.io. In past lives, I helped run VaccinateCA, created K-9 Mail for Android, created Request Tracker, and was the project lead for Perl.
I can usually be found in #Berkeley
jesse@fsck.com
jesse@keyboard.io
jesse@metasocial.com
was @obra on Twitter
Technical AI Policy Researcher at HuggingFace @hf.co π€. Current focus: Responsible AI, AI for Science, and @eval-eval.bsky.socialβ¬!
PhD student at UC Berkeley. Thinking about MLOps, data analysis and processing with LLMs, HCI
investigative reporter / OSINTer (english-language editor @crimew.gay, contributor @geoconfirmed.org @virtualstreets.org etc.)
ey/they/it/any
give me tips/contact me on signal: rhinozz.1337
Le plus grand centre de recherche universitaire en apprentissage profond β The world's largest academic research center in deep learning.
I like utilitarianism, consciousness, AI, EA, space, kindness, liberalism, longtermism, progressive rock, economics, and most people. Substack: http://timfduffy.substack.com
π Dog botherer
π Staff Eng
π§ Web1 enthusiast and fan of pointless websites
π» localghost.dev
#TransRightsAreHumanRights
Banner pic: florian.photo
Online tracker music player charged with music from keygens
https://keygenmusic.tk
local open source AI agent that automates complex developer tasks, is LLM-agnostic, integrates with MCP, and is available in both Desktop and CLI
Engineer at Sierra, formerly of Tailscale, Slack, Quip, Google Chrome and Google Reader.
writes, bikes, makes val.town, writes macwright.com, open source, geo, music, sewing and other stuff.
twitter archive: https://bsky.app/profile/archive.macwright.com
Retired software engineer. AI enthusiast. Deadhead. I implemented Bash's regex operator (=~). My Signal user ID is franl.99.
Building tools for forecasting and understanding AI at https://sage-future.org π
Effective altruism!
https://binksmith.com
theaidigest.org
Interactive AI explainers
Explore concrete examples of today's AI systems β to plan for what's coming next
Screenshots of classic websites and blogs from Dot-Com, Web 2.0 and the 2010s.
Brought to you by the internet history site https://cybercultural.com
π bridged from β https://indieweb.social/@classicweb, follow @ap.brid.gy to interact
Workflow automation platform combining AI with business processes. Connect anything to everything. Fair-code, self-hostable.