's Avatar

@dferrer.bsky.social

ML Scientist (derogatory), Ex-Cosmologist, Post Large Scale Structuralist

172 Followers  |  162 Following  |  66 Posts  |  Joined: 17.07.2024  |  1.9912

Latest posts by dferrer.bsky.social on Bluesky

You know, I think this totally doable (though I probably wont have it close the window). I think I've found my next hobby project. Not going to be at the same quality of writing, obviously, but Qwen 3 can be prodded into being deranged enough to get the right mood.

31.07.2025 22:09 β€” πŸ‘ 5    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

I guess that would be a fitting job for AI to completely obsolete.

30.07.2025 20:39 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

There is another point here about subjectivity, and LLMs having no direct apprehension of the world. That's true! It's really interesting what that could mean for concepts like "knowledge". But saying we know *conclusively* what that means is pretending we've solved Philosophy of Mind.

30.07.2025 20:39 β€” πŸ‘ 2    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

I'm not even saying that's wrong. Maybe understanding has some Qualia component an LLM can never achieve. I will happily read someone's publication arguing that. But I also want to hear how they know *people* have this "understanding." Instead, it's usually "it can't understand because it's random."

30.07.2025 20:39 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

We can say "yes, it contains a map, but it doesn't *understand* it!" Ok, sure, but if "understand" doesn't encompass "an abstract representation of the concepts in the text it was fed which it can use in novel contexts to solve problems," I'm going to need to know what we *do* mean by it.

30.07.2025 20:39 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

It feels a lot more parsimonious to say training the model has produced a structure somewhere that is like a map. For models where you can get the full reasoning trace this is even more compelling. The answers start by noting states in that general region, then check them individually.

30.07.2025 20:39 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

All of the big models released this year can answer this question well without tool calling. We can try to tell ourselves that this is just because the model has memorized every possible description of geographic relations between states, though this isn't a common benchmark or thing people discuss

30.07.2025 20:39 β€” πŸ‘ 2    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

I'm pretty sure I'd barely have an advantage over chance identifying LLM generated ML papers mixed with real ones from 6 months ago. There is very little signal to use there beyond memorizing what papers are real exhaustively, which will always be weak unless they come up *a lot* in the corpus.

23.07.2025 20:34 β€” πŸ‘ 4    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

There was a website ca. 2010 that used a CFG to generated High Energy Physics and Mathematical Physics paper titles ("Super-paramagnetic Black Saturn: A Review"). Even as a grad student in the field it was hard to tell the fake from the real most of the time. I think you're on to something here.

23.07.2025 20:15 β€” πŸ‘ 6    πŸ” 1    πŸ’¬ 1    πŸ“Œ 0

It's possible, but most academic ML / SWE training paths really don't give you the stats / applied math knowledge you need to do the science role well. You *can* learn that yourself, though. It's going to be quite a bit of work for most people.

10.07.2025 11:53 β€” πŸ‘ 2    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

The path into the positions is often different too. A lot of "engineer" roles come up through SWE and don't have significant stats or applied math training. A lot of the "science" side comes out of math / stats / physics and can be pretty shaky on SWE concepts.

10.07.2025 11:48 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

This really is a lot of it. It's also not a "soft" PhD position where you aren't really going to use a lot of the training. The difference in amount of math you need between "making a novel architecture that is even semi-functional" and "implementing the prod version of that architecture" is huge.

10.07.2025 11:48 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

They differ in training a little, but also less than you’d think (less RL for vision). There’s been broad convergence in how these models work for the last 5 years.

27.06.2025 22:15 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

A lot of this is also the collapse in meaning of β€œhallucination” from technical term-of-art for gen AI into popular meaning of β€œan AI is wrong”. But also modern vision networks use Transformer (SWin and its like) past the first few layer or two just like LLMs. Attention is too good to ignore.

27.06.2025 22:15 β€” πŸ‘ 2    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

I had a student come up after a talk ~2 years ago and tell me that "[Prompt Injection] is easy to solve---you just parse the grammar and determine what is a quote and what isn't." I love that LLMs have un-siloed semiotics from abstract thought experiments into direct engineering relevance.

17.06.2025 21:52 β€” πŸ‘ 2    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

The relationship between language, meaning, and thought is philosophically... controversial. You could easily spend a whole PhD doing nothing but arguing for *or* against every clause you have written here. That's without even mentioning LLMs.

17.06.2025 21:42 β€” πŸ‘ 7    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

I’ve seen a lot of signs but definitely feel the pride vibe too. Not complaining thoughβ€”just hoping the weather keeps holding

14.06.2025 18:09 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

Spectra also can’t be red all the way to zero, but it’s very common for them to be red down to the lowest frequencies sampled. Images in particular that have increasing power for a broad part of the sampled frequencies look *weird*

11.06.2025 12:11 β€” πŸ‘ 13    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

If you had some domain with spectra that were very blue you’d see the reverse. The reason reason real data usually aren’t is separate and *also* fascinating. Obv. finite variance signals can’t be blue asymptotically, but there are very few where they are even over an interesting chunk of freq. space

11.06.2025 12:11 β€” πŸ‘ 15    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

The reasons diffusion does this for images and time series are really interesting. It comes out of the stats of the signal that’s being modeled, not the method. Image power spectra are usually pretty close to pink. Correctly sampling the low frequencies first is optimal. It doesn’t have to be so.

11.06.2025 12:11 β€” πŸ‘ 19    πŸ” 0    πŸ’¬ 2    πŸ“Œ 1

Exactly! And the movie hates them for it. The ways it's a bad movie are so much more "predictive" than the world it shows.

11.06.2025 02:41 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

So much of what was detestable about comedies of that time seemed to have burnt its way out of Hollywood in the 2010s. It's a pity that cruelty, cynicism, and contempt for virtue have exploded back into vogue again---like the villain at the end of a slasher movie. It was behind us the whole time.

11.06.2025 02:35 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

I think this is ironically due to the way the movie is so problematic / wrong, not what it "got right". The mean-spirited, openly misogynist, weirdly eugenics-tinged vision of a future ruled by "idiots" *does* have surprising currency with the present. Just mostly not how it thinks it will.

11.06.2025 02:35 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
No, Idiocracy Is Not A Documentary
YouTube video by Sarah Z No, Idiocracy Is Not A Documentary

Heard someone say (again) today that "Idiocracy was a documentary," and I thought back to the excellent @sarahz.bsky.social's video on it: www.youtube.com/watch?v=o52z...
I don't disagree with her. I think her points hold up. But the current moment and the movie still seem to have "resonance".

11.06.2025 02:35 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

Not per capita. In total. It’s embarrassing that so many otherwise reasonable people keep saying this. You can keep hating AI! We can have that conversation another time. But repeating the water use thing is like conservatives complaining the NSF spent $40 k researching frog genders. It is *tiny*

04.06.2025 12:56 β€” πŸ‘ 3    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

People seem to think the point is that banning ChatGPT wouldn’t reduce water use as much as all of of us deciding to stop eating meat. It’s not. Banning AI entirely would have the same impact as a single school not using meat any more. Banning new training is like *just you* deciding not to eat beef

04.06.2025 12:56 β€” πŸ‘ 3    πŸ” 1    πŸ’¬ 1    πŸ“Œ 0

People *vastly* overestimate the costβ€”in water *and* powerβ€”of this industry. That’s the point. GTA 6 is almost certainly going to cause more environmental harm than ChatGPT.

04.06.2025 12:56 β€” πŸ‘ 2    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

There’s this incredible, persistent innumeracy in the replies to this and the original where people say β€œbut I can eat beef! LLMs are useless!” Or β€œas if we shouldn’t just stop eating beef.” *Training* one of the mostly costly models to produce to date used maybe *1* entire cow.

04.06.2025 12:56 β€” πŸ‘ 5    πŸ” 0    πŸ’¬ 2    πŸ“Œ 0

ACAB includes defense attorneys advising you not to talk to the cops. Anyone who limits my freedom is a cop. Anyone. ACAB even includes the person who scrawled the β€œDon’t Open Dead Inside” sign on the door holding the zombies back.
You may not like it, but this is what peak Collectivism looks like.

28.05.2025 22:27 β€” πŸ‘ 20    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

This was amusingly *easy* compared to my previous attempts at tone forcing this way. It’s almost like there’s a vast corpus of this sort of β€œcritique” to draw on.
The harder part was getting it not to do *overt* parody. Models need a lot of convincing to double down on β€œPydantic and the Death Drive”

28.05.2025 21:43 β€” πŸ‘ 2    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

@dferrer is following 20 prominent accounts