Here's a stab at it: github.com/simonw/llm/p...
19.11.2025 01:56 β π 0 π 0 π¬ 0 π 0@daturkel.bsky.social
Machine learning, recommenders and personalization, music, books. danturkel.com / Brooklyn, NY
Here's a stab at it: github.com/simonw/llm/p...
19.11.2025 01:56 β π 0 π 0 π¬ 0 π 0Alright I've tried combining my github.com/daturkel/llm... with a template gist.github.com/daturkel/4fc... but I cannot for the life of me get claude to consistently stop saying "Based on the webpage content," or "I'll fetch that webpage and summarize it for you" :(
19.11.2025 01:12 β π 1 π 0 π¬ 0 π 0ha it was the process of writing the bash script that made me want to see if I could simplify. the script workaround works fine for now but I might hack around to see if it's possible to make this work in a template
18.11.2025 23:15 β π 1 π 0 π¬ 0 π 0@simonwillison.net Is there a way in llm to have a template pass $input to a fragment loader? E.g. if I want a url summarizer, I want llm to load the input url as a fragment and pass that as context. Just setting $input as a fragment doesn't work.
18.11.2025 22:59 β π 0 π 0 π¬ 1 π 0The market is saturated, but for products which differentiate themselves with features or performance, people seem to be willing to convert at the drop of a hat. Lots of cursor fans switched to claude code seemingly overnight. This feels like a marketing failure (then again, i haven't used q either)
30.06.2025 21:17 β π 1 π 0 π¬ 0 π 0I got web search with exa working pretty easily too: github.com/daturkel/llm...
25.05.2025 21:23 β π 0 π 0 π¬ 0 π 0@simonwillison.net I'm working on rag tool for llms own embeddings db. It works, though I've had trouble getting good results out of it (it really doesn't work well with symbex --nl output; it works ok with docs). Curious what you think
github.com/daturkel/llm...
gist.github.com/daturkel/b27...
I may give it a shot if you're open to it!
17.05.2025 20:39 β π 1 π 0 π¬ 1 π 0@simonwillison.net I made an llm fragment plugin to use a library (Trafilatura) to locally convert a site to markdown rather than using the Jina API. The experience was super easy and fun.
Is it possible to pass a fragment into a chat though?
github.com/daturkel/llm...
Loved this post from Dillon Davis at Airbnb on how they evolved their location retrieval system from heuristics to a NN-powered bandit.
Somehow I'd never heard of Monte Carlo dropout, but it looks like a great way to get uncertainty estimates from NNs (eg for bandits).
medium.com/airbnb-engin...
I thought Succession was based on the Murdochs but it turns out that the Murdochs are based on Succession.
09.12.2024 23:23 β π 2 π 0 π¬ 0 π 0even apple is selling genai as the way to half-ass a gift for your spouse. there are so many appealing non-dystopian use-cases...why pick this? youtu.be/A0BXZhdDqZM?...
03.12.2024 03:10 β π 1 π 0 π¬ 0 π 0Your marketing worked on meβI was going to grab maybe 6 or 7 and then I decided screw it I'll get the bundle.
29.11.2024 23:02 β π 0 π 0 π¬ 0 π 0Someone can take my photo in public and sell it, and we generally are on board with this. But we probably wouldn't want a billion dollar corporation to monetize photos taken of all people in public at all times. Same principle, different vibes.
29.11.2024 03:20 β π 1 π 0 π¬ 0 π 0Unless your site is very complicated I think you should have more or less the same experience. It's python, not a punch-card powered mainframe!
29.11.2024 03:09 β π 1 π 0 π¬ 0 π 0Benedict Evans has a good essay about how principles don't always hold up at massive scale, which has been useful for me in thinking through some of these issues. www.ben-evans.com/benedictevan...
29.11.2024 03:08 β π 0 π 0 π¬ 1 π 0It does have hot reloading, and it's fairly fast. docs.getpelican.com/en/latest/pu...
29.11.2024 02:47 β π 0 π 0 π¬ 1 π 0You probably found that Pelican and Nikola are the two dominant python choices. They're not as popular as Hugo and some others, but I've been using Pelican for a few years now with no problem. I find it a bit more hackable when I'm fluent in the language, which I'm not for Go etc
29.11.2024 00:00 β π 1 π 0 π¬ 1 π 0That makes sense, I actually recall someone even saying that the research seemed to be outpacing the accessible implementations for sparse neural embeddings. Maybe once it's in sentence transformers, splade will be everywhere π
28.11.2024 12:08 β π 0 π 0 π¬ 0 π 0This is not how the industry will win the hearts and minds of its critics...
28.11.2024 04:17 β π 2 π 0 π¬ 1 π 0Bluesky might not have some of the specific subsets of trolls that are now rampant on Twitter, but I don't think they'll solve the fact that people can be much meaner behind a screen and keyboard than they would be "irl" π€
28.11.2024 02:43 β π 2 π 0 π¬ 1 π 0If all of this means better standards going forward for opt-outs, that could be a win for everyone.
28.11.2024 02:39 β π 4 π 0 π¬ 1 π 0I think the fairly indiscriminate use of internet data to train commercial generative models has brought awareness to and shifted opinion against similar (if more innocuous) practices that have existed for a while, and the industry may have to reckon with the resulting bad vibes.
28.11.2024 02:38 β π 5 π 0 π¬ 1 π 0Whether I agree with it or not (and I'm still thinking it through), I think the reactions are to the principle of the collection and publication of these datasets in general, not exclusively to the HF dataset. I doubt the posters you're referring to would be thrilled about this one.
28.11.2024 02:35 β π 2 π 0 π¬ 1 π 0Hey Tom, thanks so much for your work on sentence-transformers, it's a terrific library.
When I was at SIGIR this year, SPLADE and sparse embeddings came up a *lot* in talks, but no one was able to point to much adoption of these methods in industry. Have you seen it? If not, why do you think?
A screenshot of a python snippet demonstrating a configuration pattern.
Yeah, it's not perfect and it's a little verbose, but it gracefully handles a few different scenarios (see attached).
21.11.2024 03:50 β π 1 π 0 π¬ 1 π 0(also, love the blog post! this is a hard and common problem and I don't think anyone's solved it super well, including our design described above)
21.11.2024 00:22 β π 2 π 0 π¬ 1 π 0we add methods to our pydantic model which return a dict of all the kwargs needed in a certain situation, so you end up with something like
some_fn_or_obj(**cfg.get_kwargs_for_some_fn_or_obj).
I prefer not passing the entire config object in order to reduce the size of the interface 2/2
We manage configs with hydra.cc structured configs (which are pydantic models) but we don't use its ability to dynamically instantiate objects. Instead, since multiple objects and functions may need overlapping fields from the config... 1/2
21.11.2024 00:21 β π 1 π 0 π¬ 1 π 0