Greg's Avatar

Greg

@theaspiringnerd.com.bsky.social

Neurodivergent AI Product Manager, LLMs Enthusiast & Dataviz Nerd πŸ€“ TiohtiΓ :ke / Montreal / Brussels πŸš€

229 Followers  |  202 Following  |  52 Posts  |  Joined: 06.02.2024  |  2.0808

Latest posts by theaspiringnerd.com on Bluesky

Preview
MCP Colors: Systematically deal with prompt injection risk

MCP Colors

A riff off of the lethal trifecta for addressing prompt injection, this is a simple heuristic to ensure security at runtime

red = untrusted content
blue = potentially critical actions

An agent can't be allowed to do both

timkellogg.me/blog/2025/11...

04.11.2025 02:27 β€” πŸ‘ 32    πŸ” 4    πŸ’¬ 3    πŸ“Œ 2

Yes please !

03.11.2025 22:55 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
28.10.2025 00:09 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

my take, after reading replies all day:

we’re still early. people aren’t spending much money on AI so it’s not a lucrative target yet

it’s also inconsistent, which is annoying to design attacks for, especially if the rewards are sparse

26.10.2025 21:19 β€” πŸ‘ 25    πŸ” 1    πŸ’¬ 3    πŸ“Œ 0
Preview
claude_code_docs_map.md Something I'm enjoying about Claude Code is that any time you ask it questions about itself it runs tool calls like these: In this case I'd asked it about its …

It's neat how if you ask Claude Code questions about itself it can answer them, because it knows how to fetch a Markdown index of its own online documentation and then navigate to the right place

I wish more LLM tools would implement the same pattern! simonwillison.net/2025/Oct/24/...

24.10.2025 23:06 β€” πŸ‘ 99    πŸ” 11    πŸ’¬ 6    πŸ“Œ 1
Preview
An Opinionated Guide to Using AI Right Now What AI to use in late 2025

I wrote an updated guide on which AIs to use right now, & some tips on how to use them (and how to avoid falling into some common traps)

A lot has changed since I last wrote a guide like this in the spring, and AI has gotten much more useful as a result. open.substack.com/pub/oneusefu...

19.10.2025 18:48 β€” πŸ‘ 153    πŸ” 29    πŸ’¬ 6    πŸ“Œ 3
clem
@ClementDelangue

Am I wrong in sensing a paradigm shift in Al?
Feels like we're moving from a world obsessed with generalist LLM APls to one where more and more companies are training, optimizing, and running their own models built on open source (especially smaller, specialized ones)
Some validating signs just in the past few weeks:
- @karpathy released nanochat to train models in just a few lines of code
- @thinkymachines launched a fine-tuning product
- rising popularity of @vllm_project, @sgl_project, @PrimeIntellect, Loras, trl,...
- 1M new repos on HF in the past 90 days (including the first open-source LLMs from @OpenAI)
And now, @nvidia just announced DGX Spark, powerful enough for everyone to fine-tune their own models at home.

clem @ClementDelangue Am I wrong in sensing a paradigm shift in Al? Feels like we're moving from a world obsessed with generalist LLM APls to one where more and more companies are training, optimizing, and running their own models built on open source (especially smaller, specialized ones) Some validating signs just in the past few weeks: - @karpathy released nanochat to train models in just a few lines of code - @thinkymachines launched a fine-tuning product - rising popularity of @vllm_project, @sgl_project, @PrimeIntellect, Loras, trl,... - 1M new repos on HF in the past 90 days (including the first open-source LLMs from @OpenAI) And now, @nvidia just announced DGX Spark, powerful enough for everyone to fine-tune their own models at home.

correct

i’ve been saying this for a couple months. RL is driving towards specialization

my hunch is it’s temporary and something will shift again back towards generalization, but for now.. buckle up!

15.10.2025 11:39 β€” πŸ‘ 60    πŸ” 5    πŸ’¬ 2    πŸ“Œ 2
Preview
Sora 2 Watermark Removers Flood the Web Bypassing Sora 2's rudimentary safety features is easy and experts worry it'll lead to a new era of scams and disinformation.

I think people are still unprepared for a world where you cannot trust any video content, despite years of warning.

Even when Google & OpenAI include watermarks, those can be easily removed, and open weights AI video models without guardrails are coming. www.404media.co/sora-2-water...

08.10.2025 19:18 β€” πŸ‘ 151    πŸ” 37    πŸ’¬ 7    πŸ“Œ 4

I think I have at least one conversation (or argument?) about this every couple of days!

04.10.2025 22:01 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

sheesh! AI bluesky has arrived

not just good content, there’s more and more original work, people from labs, and people with genuinely interesting perspectives

when i joined, it was so painful trying to find even traces

27.09.2025 17:56 β€” πŸ‘ 143    πŸ” 9    πŸ’¬ 6    πŸ“Œ 1
Post image

Codex in the app is going to open the door to real vibe coding on the go β€” no computer required. I’m so excited for this to expand, the ceiling is so high and this is the worst the models will ever be.

25.09.2025 17:42 β€” πŸ‘ 13    πŸ” 1    πŸ’¬ 1    πŸ“Œ 0

The key detail people may miss: it looks like an AI company in the USA can train on an author's book by purchasing a used copy, cutting it up and scanning the pages - in which case the author gets no money at all!

06.09.2025 06:19 β€” πŸ‘ 47    πŸ” 9    πŸ’¬ 6    πŸ“Œ 3

I feel seen πŸ‘€ Although it’s improved since I use Cursor and have it generate my commit messages!

30.08.2025 22:36 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Post image

This is illustrative of common AI issues:
1) When you get an instant AI answer, it is from a small model, which are weak models, especially at math.
2) Non-reasoning models, like the one powering AI overview, only β€œthink” as they write, they make mistakes & then back justify them as they write more

24.08.2025 13:33 β€” πŸ‘ 143    πŸ” 14    πŸ’¬ 11    πŸ“Œ 8
The chart presents the decomposition of Average Treatment Effect (ATE) on cosine similarity into two components: Model Effect (red) and Prompting Effect (blue).
	β€’	Y-axis: Ξ” Cosine Similarity (change in similarity).
	β€’	X-axis: The source of prompts (top labels) and the replay model used (bottom labels).
	β€’	Points and error bars: Represent mean effects with 95% confidence intervals, bootstrapped and clustered by participant.

Breakdown:
	1.	DALL-E 2 β†’ DALL-E 2 (baseline): Ξ” Cosine Similarity is ~0, establishing the reference point.
	2.	DALL-E 2 prompts replayed on DALL-E 3: Shows a Model Effect (increase ~0.007–0.008). This isolates the improvement attributable to the newer model when given the same prompts.
	3.	DALL-E 3 prompts replayed on DALL-E 3 vs DALL-E 2 prompts on DALL-E 3: The additional boost is attributed to the Prompting Effect (~0.006–0.007).
	4.	Total ATE (black bracket): When prompts written for DALL-E 3 are used on DALL-E 3, the improvement in cosine similarity reaches ~0.016–0.018.
	5.	DALL-E 3 prompts replayed on DALL-E 2: Effect is small, close to baseline, showing the limited benefit of improved prompts without the newer model.

Summary (from caption):
	β€’	ATE (black) = Model Effect (red) + Prompting Effect (blue).
	β€’	Model upgrades (DALL-E 3 vs DALL-E 2) and better prompt designs both contribute to improved performance.
	β€’	Prompting alone offers some gains, but most improvements come from model advancements.

The chart presents the decomposition of Average Treatment Effect (ATE) on cosine similarity into two components: Model Effect (red) and Prompting Effect (blue). β€’ Y-axis: Ξ” Cosine Similarity (change in similarity). β€’ X-axis: The source of prompts (top labels) and the replay model used (bottom labels). β€’ Points and error bars: Represent mean effects with 95% confidence intervals, bootstrapped and clustered by participant. Breakdown: 1. DALL-E 2 β†’ DALL-E 2 (baseline): Ξ” Cosine Similarity is ~0, establishing the reference point. 2. DALL-E 2 prompts replayed on DALL-E 3: Shows a Model Effect (increase ~0.007–0.008). This isolates the improvement attributable to the newer model when given the same prompts. 3. DALL-E 3 prompts replayed on DALL-E 3 vs DALL-E 2 prompts on DALL-E 3: The additional boost is attributed to the Prompting Effect (~0.006–0.007). 4. Total ATE (black bracket): When prompts written for DALL-E 3 are used on DALL-E 3, the improvement in cosine similarity reaches ~0.016–0.018. 5. DALL-E 3 prompts replayed on DALL-E 2: Effect is small, close to baseline, showing the limited benefit of improved prompts without the newer model. Summary (from caption): β€’ ATE (black) = Model Effect (red) + Prompting Effect (blue). β€’ Model upgrades (DALL-E 3 vs DALL-E 2) and better prompt designs both contribute to improved performance. β€’ Prompting alone offers some gains, but most improvements come from model advancements.

As LLMs Improve, People Adapt Their Prompts

a study shows that a lot of the real world performance gains that people see are actually because people learn how to use the model better

arxiv.org/abs/2407.14333

23.08.2025 16:40 β€” πŸ‘ 44    πŸ” 9    πŸ’¬ 2    πŸ“Œ 1

When you use coding agents for something that produces the bulk of the code of an application that will be used for years, also factor in the technical debt that you are happily accumulating. When you use LLMs an an aid, you could, on the contrary, improve your coding culture.

21.08.2025 13:08 β€” πŸ‘ 50    πŸ” 8    πŸ’¬ 1    πŸ“Œ 1
Post image

Looking at the ARC-AGI benchmark is a useful way of understanding AI progress.

There are two goals in AI, minimize cost (which is also roughly environmental impact of use) & maximize ability. It is clear you can win one goal by losing the other, GPT-5 seems to be a gain on both.

17.08.2025 11:32 β€” πŸ‘ 45    πŸ” 2    πŸ’¬ 2    πŸ“Œ 1
Pelican Riding Bike

This is the cat!
He's got big wings and a happy tail.
He loves to ride his bike!

Bike lights are shining bright.
He's got a shiny top, too!
He's ready for adventure!

Pelican Riding Bike This is the cat! He's got big wings and a happy tail. He loves to ride his bike! Bike lights are shining bright. He's got a shiny top, too! He's ready for adventure!

The new Gemma 3 270M open weights model from Google is really fun - it's absolutely tiny, just a 241MB download

I asked it for an SVG of a pelican riding a bicycle and it wrote me a delightful little poem instead

simonwillison.net/2025/Aug/14/...

14.08.2025 17:27 β€” πŸ‘ 161    πŸ” 13    πŸ’¬ 9    πŸ“Œ 5
roon y @tszzl
model switcher paradigm will be vindicated in the long run. there is a high switching cost into a very new UX on a useful product, but it's the right move

roon y @tszzl model switcher paradigm will be vindicated in the long run. there is a high switching cost into a very new UX on a useful product, but it's the right move

i think i agree. i also think it’ll be a moat against other labs that don’t a strong product

it doesn’t make sense if you don’t have a strong UI, plus it’s obviously hard

10.08.2025 12:02 β€” πŸ‘ 10    πŸ” 1    πŸ’¬ 5    πŸ“Œ 0

Same! Not as wild as I imagined !

08.08.2025 22:34 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

As a PM, I really can’t wrap my head around this either πŸ€”

08.08.2025 20:52 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

You are likely going to see a lot of very varied results posted online from GPT-5 because it is actually multiple models, some of which are very good and some of which are meh.

Since the underlying model selection isn’t transparent, expect confusion.

07.08.2025 20:45 β€” πŸ‘ 142    πŸ” 19    πŸ’¬ 14    πŸ“Œ 6
Video thumbnail

I had access to GPT-5. I think it is a very big deal as it is very smart & just does stuff for you

This is β€œmake a procedural brutalist building creator where i can drag and edit buildings in cool ways" & "make it better" a bunch. I touched no code

Full post: www.oneusefulthing.org/p/gpt-5-it-j...

07.08.2025 17:23 β€” πŸ‘ 104    πŸ” 25    πŸ’¬ 7    πŸ“Œ 5
Post image

What could be a two step master plan:
1. Release open model to commoditize much of the model market a tad off the frontier
2. Release GPT 5 as the only model worth paying for

Really curious about all the strategy decisions that went into this open model (& GPT5/Gemini 3)

05.08.2025 21:01 β€” πŸ‘ 24    πŸ” 4    πŸ’¬ 4    πŸ“Œ 3

Really looking forward to see how it performs with tools!

05.08.2025 23:23 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Video thumbnail

I prompted Gemini 2.5 Deep Think: "create a missile command game that incorporates relativity in realistic ways but is still playable." I then asked it to improve the design a few times.

The full design & all code & calculations came from AI, no errors. Try it: glittery-raindrop-318339.netlify.app

02.08.2025 04:22 β€” πŸ‘ 78    πŸ” 7    πŸ’¬ 2    πŸ“Œ 3
A line chart with vibe on the Y axis and understanding on the X axis with a downwards diagonal line

A line chart with vibe on the Y axis and understanding on the X axis with a downwards diagonal line

A helpful graph made by @stevekrouse.com on the inverse relationship between vibes and understanding in AI assisted code.

Put a few thoughts down here: maggieappleton.com/2025-08-vibe...

Original article: blog.val.town/vibe-code

02.08.2025 12:23 β€” πŸ‘ 71    πŸ” 11    πŸ’¬ 2    πŸ“Œ 1
Preview
The best available open weight LLMs now come from China Something that has become undeniable this month is that the best available open weight models now come from the Chinese AI labs. I continue to have a lot of love …

July has been a truly incredible month for model releases from China - Moonshot (Kimi K2), Z ai (GLM-4.5) and 5 new releases from Qwen

I think it's undeniable that the best available open weight models now come from the Chinese AI labs simonwillison.net/2025/Jul/30/...

30.07.2025 16:21 β€” πŸ‘ 86    πŸ” 9    πŸ’¬ 5    πŸ“Œ 1
Preview
Latest open artifacts (#12): Chinese models continue to dominate throughout the summer 🦦 Artifacts Log 12.

Having covered the open models coming out in the last 2-3 years of the AI boom, it’s safe to say that we’re really starting to reach maturity. There are very serious AI models across many modalities that anyone can download and start to make products with.
buff.ly/lIU1ftr

22.07.2025 00:16 β€” πŸ‘ 21    πŸ” 3    πŸ’¬ 2    πŸ“Œ 0

Don't leave AI to the STEM folks.

They are often far worse at getting AI to do stuff than those with a liberal arts or social science bent. LLMs are built from the vast corpus human expression, and knowing the history & obscure corners of human works lets you do far more with AI & get its limits.

20.07.2025 18:06 β€” πŸ‘ 139    πŸ” 28    πŸ’¬ 5    πŸ“Œ 2

@theaspiringnerd.com is following 20 prominent accounts