What UI is this π§
07.10.2025 22:37 β π 1 π 0 π¬ 1 π 0@isolyth.dev.bsky.social
ππ€ππ€ βββββββββββββββββββββββββββββββββββββββ There's never been a better time to have a problem https://isolyth.dev
What UI is this π§
07.10.2025 22:37 β π 1 π 0 π¬ 1 π 0The graphic at the top of thinkingmachines.ai/blog/lora/ is a giant SVG, defined right in the page. I noticed this because 2/3 of the note obsidian web clipper wanted to create was just SVG code
07.10.2025 19:29 β π 2 π 0 π¬ 0 π 0PROJECT ANNOUNCEMENT: Streamplace is building a unified AT Protocol code generation tool! It eats lexicons and creates code in a variety of languages, as well as documentation.
To lead this project, who better than the creator of @pdsls.dev, @juli.ee!
Read more in the announcement post:
yes! at least partially. longposted about it on other site here: x.com/voooooogel/s...
07.10.2025 04:39 β π 27 π 1 π¬ 1 π 3i wrote a custom llm sampler for llama-3.1-8b so it could only say words that are in the bible
07.10.2025 04:35 β π 378 π 62 π¬ 12 π 14People often ask why we donβt act on off-platform evidence when investigating reports. Hereβs an example of how complex that can get. 1/9
06.10.2025 20:46 β π 144 π 22 π¬ 101 π 88if you're curious about the architecture and mechanics of LLMs, this site has a really excellent explorable interactive visualization. it helps build intuition for how massive these models are, what 'interpretability' means, and the complexity involved here
bbycroft.net/llm
Interesting paper suggesting a mechanism for why in-context learning happens in LLMs.
They show that LLMs implicitly apply an internal low-rank weight update adjusted by the context. Itβs cheap (due to the low-rank) but effective for adapting the modelβs behavior.
#MLSky
arxiv.org/abs/2507.16003
I'm not aware, you may be able to find something on r/locallama
On Gemini doing better at long context: π€· I think Google might be doing the deepseek attention thing, which, seemingly (with thinking) improves long context performance a lot, for some reason
The pixel fold is neat because you can put graphene on it and the cameras are great but their obsession over curved corner displays makes the outside look so goofy π also after Samsung made the z fold so absurdly thin (its as thick as a regular phone when closed!) it's weird how thick the pf10 is
06.10.2025 13:13 β π 0 π 0 π¬ 0 π 0That is possible, but if you *also* have to scale a bunch to use large contexts, it's not a super economical or really practical way forward. Also I can't see it in the reply view but I don't think Kimi or GLM (both are at like a trillion parameters I think?) do all that well
06.10.2025 13:10 β π 1 π 0 π¬ 1 π 0OpenAI is having a livestream tomorrow... Could be them?
06.10.2025 05:55 β π 0 π 0 π¬ 0 π 0The state of LLMs is messy: Some AI features (like vision) lag others (like tool use) while others have blind spots (imagegen and clocks). And the expensive "heavy thinking" models are now very far ahead of all the other AIs that most people use, capable of real work
None of this is well-documented
Ehhhh, not really? Gpt-5 does worse than Gemini here. It holds up pretty well but it does start to decline noticeably towards the end, if models were actually able to be scaled to 100m or whatever I would expect to see straight hundreds at relatively tiny contexts like this
05.10.2025 22:00 β π 4 π 0 π¬ 1 π 0It starts to get real small once you start feeding in lots of data :( also LLMs suck at actually using past like, 128k, even on the best models like Gemini. Most can't really perform well past 64k
05.10.2025 21:42 β π 1 π 0 π¬ 1 π 0Or like, here is an absolute psycho just happily recounting his behavior in a github comment. You can go look at a whole pile of Claude Code issues and see similar, because of, uh, Exactly Who Is Using It This Way: put kindly, people who want employees and should not have them.
05.10.2025 18:32 β π 44 π 2 π¬ 2 π 4The web can be made personal again.
blog.muni.town/personal-dat...
100 million in a lifetime? I guess but it often takes multiple tokens to encode a word, and LLMs get literally ALL of their input from tokens, so comparing it just to the words humans hear doesn't seem right. A moderate codebase can already consume most of a million just on its own :/
05.10.2025 18:30 β π 1 π 0 π¬ 1 π 0Update: @vgel.me has written a blog post about why this happens from a mechanistic interpretability standpoint!
vgel.me/posts/seahor...
A little proof of concept I've been working on, it's a version of Bluesky that you can run locally, even on your laptop, that pulls just the data you care about from the firehose, indexes it, and provides you an interface to view and interact with posts.
04.10.2025 21:28 β π 887 π 180 π¬ 19 π 16I think the problem with a lot anti-AI people on here is that they actually believe this very first sentence.
04.10.2025 03:15 β π 50 π 7 π¬ 5 π 0I gotta be honest, I think it's too late. They got there first. I think it's like Q-Tips and Kleenex now.
πββοΈ "I asked chajeepity."
π€¦ββοΈ "Which chajeepity?"
πββοΈ "Claude."
π€¦ββοΈ "That's a pretty good chajeepity. Dumb name though."
Use For You instead of Discover
Use For You instead of Discover
Use For You instead of Discover
Use For You instead of Discover
Use For You instead of Discover