antirez's Avatar

antirez

@antirez.bsky.social

Reproducible bugs are candies ๐Ÿญ๐Ÿฌ

7,611 Followers  |  377 Following  |  753 Posts  |  Joined: 26.04.2023  |  1.8748

Latest posts by antirez.bsky.social on Bluesky

Post image

Why I believe there are a lot of misleading papers about LLM reasoning.

10.10.2025 06:08 โ€” ๐Ÿ‘ 43    ๐Ÿ” 4    ๐Ÿ’ฌ 6    ๐Ÿ“Œ 1

That's the whole point. The concept exists almost everywhere but in places like Sicily and China it assumes a fundamental role in life.

04.10.2025 21:04 โ€” ๐Ÿ‘ 0    ๐Ÿ” 1    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

Today I discovered that in China there is this concept of "losing face" (mianzi / ้ขๅญ), which maps 1:1 with Sicilian "perdiri la facci" both as exact wording and conceptually.

04.10.2025 08:02 โ€” ๐Ÿ‘ 89    ๐Ÿ” 2    ๐Ÿ’ฌ 10    ๐Ÿ“Œ 0
Vector databases and recall: exploring the concept in a more deep way.
YouTube video by Salvatore Sanfilippo Vector databases and recall: exploring the concept in a more deep way.

Saying that a vector database reaches a given recall percentage is more marketing than anything, HNSWs are more complicated than that. Let's explore why in this video. www.youtube.com/watch?v=LG3q...

30.09.2025 15:01 โ€” ๐Ÿ‘ 24    ๐Ÿ” 1    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0
Made On YouTube 2025: Auto-Dubbing
YouTube video by Updates From YouTube Made On YouTube 2025: Auto-Dubbing

Finally YouTube is going to improve the dubbing feature with lipsync and potentially (looking at this example video) voice matching. That's great in general, and I'm very happy for my YouTube channel that can be better experienced by english speakers. www.youtube.com/watch?v=8W3n...

27.09.2025 15:41 โ€” ๐Ÿ‘ 14    ๐Ÿ” 1    ๐Ÿ’ฌ 2    ๐Ÿ“Œ 0

Who are the designers of the new iPhone UI? Would love to hire them in order to create the hardest escape room in the world.

19.09.2025 07:02 โ€” ๐Ÿ‘ 90    ๐Ÿ” 3    ๐Ÿ’ฌ 3    ๐Ÿ“Œ 0

We have it now! "WITHATTRIBS" :) It's documented in the README, and I'm here for any other info. Merged into unstable some time ago.

18.09.2025 05:44 โ€” ๐Ÿ‘ 2    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

VRANGE was merged into Redis unstable and will be likely back-ported. It allows to get lex ranges of elements in vector sets, which is useful as iterator and for other tasks.

17.09.2025 18:08 โ€” ๐Ÿ‘ 18    ๐Ÿ” 0    ๐Ÿ’ฌ 4    ๐Ÿ“Œ 0

It's a few days that I'm writing a blog post on HNSWs. As usually, writing about a subject raises new questions, that trigger new searches, and in the goal of explaining what you understood, you understand more. Blogging may no longer be cool but surely remains useful.

08.09.2025 15:37 โ€” ๐Ÿ‘ 14    ๐Ÿ” 1    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0
Inspecting LLM embeddings in GGUF format with gguflib.c and Redis
YouTube video by Salvatore Sanfilippo Inspecting LLM embeddings in GGUF format with gguflib.c and Redis

New EN video: Inspecting LLM embeddings in GGUF format with gguflib.c and Redis

www.youtube.com/watch?v=ugba...

01.09.2025 18:25 โ€” ๐Ÿ‘ 12    ๐Ÿ” 2    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0
Post image

ADV: my C course on YouTube is free and can be watched without problems with subtitles or dubbed audio by non Italian speakers :)

01.09.2025 11:34 โ€” ๐Ÿ‘ 39    ๐Ÿ” 7    ๐Ÿ’ฌ 2    ๐Ÿ“Œ 0

LLMs are not just data dumps (with errors) are powerful data manipulation and transformation devices.

29.08.2025 14:43 โ€” ๐Ÿ‘ 1    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

This also needs to contain devices designed to last, with simple UIs (keyboards and displays) that can be powered on (solar power maybe?) to ask questions straight away. GPUs may long be gone at some point.

29.08.2025 13:31 โ€” ๐Ÿ‘ 3    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

We should likely put a large LLM weights and the explanation of how to do inference in some hard-to-degradate material, inside N bunkers in different parts of the world. So that humans are more likely to be able to restart civilization in case of fatal events.

29.08.2025 13:06 โ€” ๐Ÿ‘ 32    ๐Ÿ” 3    ๐Ÿ’ฌ 6    ๐Ÿ“Œ 0

Still, I believe there is no better way so far in order to archive the human history for the long term. The web (it is now very clear) loses pieces every year. I really hope that the LLMs weights will be saved for a long time as an archive, incorrect very in many details, but better than nothing.

29.08.2025 12:48 โ€” ๐Ÿ‘ 10    ๐Ÿ” 1    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

That's however is, in my opinion, a *side effect*. I refute the idea that LLMs are mainly just lossy compression of information: if we don't add to this view their ability to combine and process such information, we are providing a strongly insufficient mental model.

29.08.2025 12:46 โ€” ๐Ÿ‘ 8    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

At this stage AI will provide a huge boost to balanced people that don't jump in any extreme hype / anti-hype train, and will create issues to the folks that don't see that judicious application of LLMs is the key. One of those historical moments when wisdom is an advantage.

28.08.2025 17:10 โ€” ๐Ÿ‘ 46    ๐Ÿ” 3    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0
Post image

Playing with adding LLMs embeddings into Redis vector sets, to explore relationships and differences with embedding words models like Word2Vec. Will make a YouTube video and/or blog post.

25.08.2025 14:37 โ€” ๐Ÿ‘ 16    ๐Ÿ” 1    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

The problem is there, it's the solution that is totally useless. I'm not denying the problem. But it is self-evident when the PR is a mass of AI generated code. At the same time, AI can help to refine a PR and produce *better* code. Also contributors build trust over time usually.

25.08.2025 11:57 โ€” ๐Ÿ‘ 2    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

Why the "declare AI help in your PR" is useless, courtesy of 3000 years ago logic:

- If you can't tell without the disclaimer, it's useless: people will not tell.

- If you can tell, it's useless: you can tell.

Just evaluate quality, not process.

24.08.2025 17:20 โ€” ๐Ÿ‘ 32    ๐Ÿ” 4    ๐Ÿ’ฌ 5    ๐Ÿ“Œ 1

You are not there for some time now. Maybe you can't imagine the level the posts in the timeline have reached: it is TERRIBLE. All full of sexist racist shit I don't follow. Also things that don't make me sleep at night like footages of incidents and alike. What a total trash.

22.08.2025 14:16 โ€” ๐Ÿ‘ 1    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

It's worse than that: I'm in both communities because for work concerns I can't afford to cut most of the community of Redis, but there are also people there that kinda questioned the fact I encourage switching here. Also many top accounts (not Musk/Trump supporters) stay there for simplicity.

22.08.2025 13:40 โ€” ๐Ÿ‘ 3    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

When you use coding agents for something that produces the bulk of the code of an application that will be used for years, also factor in the technical debt that you are happily accumulating. When you use LLMs an an aid, you could, on the contrary, improve your coding culture.

21.08.2025 13:08 โ€” ๐Ÿ‘ 50    ๐Ÿ” 8    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 1

New blog post: AI is different.

antirez.com/news/155

13.08.2025 16:03 โ€” ๐Ÿ‘ 16    ๐Ÿ” 0    ๐Ÿ’ฌ 3    ๐Ÿ“Œ 0

Experts self distillation, RL on long context directly vs incremental, XML format, exact optimizer parameters, there is quite some meat there.

12.08.2025 11:58 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

Are you referring to that *exact* technique where in x' = sin(x) * m, you take a lookup table of four shifts and do (m>>s1) + (m>>s2) + (m>>s3) + (m>>s4)? I never saw this used in the past, AFAIK, and LLMs are not aware of it. Note that it is not a lookup table for the sin(x) value itself.

11.08.2025 20:32 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0
Post image

That's the best explanation I got of a technique I developed during Covid for Z80 (Spectrum) 3D programming.

08.08.2025 09:57 โ€” ๐Ÿ‘ 19    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 2

GPT-5 seems good at code reviews. The first model from OpenAI in a *long* time I'll likely start using constantly for such an activity, together with the others.

08.08.2025 09:50 โ€” ๐Ÿ‘ 55    ๐Ÿ” 0    ๐Ÿ’ฌ 7    ๐Ÿ“Œ 0

LLM evaluation: only trust people with hard problems. Today models perform decently on most trivial task: this is a win of the technology, but also means that evaluation is more and more a realm of sprcislists.

07.08.2025 20:47 โ€” ๐Ÿ‘ 26    ๐Ÿ” 1    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

So it looks like Anthropic understood money are on SOTA coding models, and is happy with lagging a bit behind in other benchmarks / performances. Makes quite sense after all, especially if there is a tension fine tuning for different goals.

06.08.2025 16:13 โ€” ๐Ÿ‘ 1    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

@antirez is following 20 prominent accounts