Nafnlaus ๐Ÿ‡ฎ๐Ÿ‡ธ ๐Ÿ‡บ๐Ÿ‡ฆ ๐Ÿ‡ฌ๐Ÿ‡ช's Avatar

Nafnlaus ๐Ÿ‡ฎ๐Ÿ‡ธ ๐Ÿ‡บ๐Ÿ‡ฆ ๐Ÿ‡ฌ๐Ÿ‡ช

@nafnlaus.bsky.social

Mastodon: @nafnlaus@fosstodon.org Twitter: @enn_nafnlaus URL: https://softmaxdroptableartists.bandcamp.com/ #Energy #EVs #Ukraine #AI #Horticulture #Research

6,437 Followers  |  740 Following  |  33,071 Posts  |  Joined: 03.05.2023
Posts Following

Posts by Nafnlaus ๐Ÿ‡ฎ๐Ÿ‡ธ ๐Ÿ‡บ๐Ÿ‡ฆ ๐Ÿ‡ฌ๐Ÿ‡ช (@nafnlaus.bsky.social)

It's time for the latest episode of "Everybody Hates Iran"...

A reminder that Saudi Arabia's military budget is nearly as high as Russia's (though they don't get as good of a buy for their money)

01.03.2026 16:44 โ€” ๐Ÿ‘ 1    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

Issam al-Da'alis (2025, Palestine)*+
Ahmed al-Rahawi (2025, Yemen)*+
Ali Khamenei (2026, Iran)

+: Nuance to the degree of state-links of the killing
*: Nuance to whether the target leads a "state"

But apart from these (and others).... ๐Ÿ™ƒ

01.03.2026 16:38 โ€” ๐Ÿ‘ 4    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

Thomas Sankara (1987, Burkina Faso)+
Renรฉ Moawad (1989, Lebanon)+
Rafic Hariri (2005, Lebanon)+
Ahmed Yassin (2004, Hamas/Palestine)*
Ismail Haniyeh (2024, Hamas/Palestine)*
Hassan Nasrallah (2024, Hezbollah/Lebanon)*
Yahya Sinwar (2024, Hamas/Palestine)*

01.03.2026 16:38 โ€” ๐Ÿ‘ 2    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

Ngo Dinh Diem (1963, South Vietnam)
Renรฉ Schneider (1970, Chile)+
Wasfi Tal (1971, Jordan)+
Salvador Allende (1973, Chile)
Ibrahim al-Hamdi (1977, North Yemen)+
Hafizullah Amin (1979, Afghanistan)
Omar Torrijos (1981, Panama)+
Bachir Gemayel (1982, Lebanon)+
Rashid Karami (1987, Lebanon)+
...

01.03.2026 16:38 โ€” ๐Ÿ‘ 2    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

Zhang Zuolin (1928, China)
Armand Cฤƒlinescu (1939, Romania)
Josรฉ Abad Santos (1942, Philippines)
Hazza' al-Majali (1960, Jordan)
Patrice Lumumba (1961, Democratic Republic of the Congo)+
Rafael Trujillo (1961, Dominican Republic)
Sylvanus Olympio (1963, Togo)+
Abd al-Karim Qasim (1963, Iraq)+
...

01.03.2026 16:38 โ€” ๐Ÿ‘ 2    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0
a man giving a thumbs up in front of a periodic table for polonium 209 ALT: a man giving a thumbs up in front of a periodic table for polonium 209
01.03.2026 11:51 โ€” ๐Ÿ‘ 1    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

100% on all five of those.

01.03.2026 01:14 โ€” ๐Ÿ‘ 1    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

I think it's an interception behind it.

01.03.2026 01:12 โ€” ๐Ÿ‘ 1    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

I read as "Epic Furry"

28.02.2026 19:27 โ€” ๐Ÿ‘ 1    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

I'm not going to carry on this conversation any further.

To anyone reading this thread who actually wants to know how LLMs work, read the linked pages in the order provided

(I didn't include anything for attention, but just google Word2Vec & GloVe for a primitive version)

28.02.2026 19:06 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0
Vector database - Wikipedia

Without all of them, you do not have a LLM. The closest real-world thing to what you're describing is a vector database.

en.wikipedia.org/wiki/Vector_...

28.02.2026 19:02 โ€” ๐Ÿ‘ 1    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

Your description of latent spaces was pretty good (apart from describing it as linguistic). But LLMs don't "work" on latent spaces. Latent spaces are one of three separate things that come together to enable Transformers, alongside attention and DNNs. You have to understand all of them.

28.02.2026 19:01 โ€” ๐Ÿ‘ 1    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

FFNs are what *do the work* of finding the next latent position. Which is often quite different than the input hidden state they're provided (add + norm, which we do to avoid the disappearing gradient problem, helps carry forward the earlier hidden state positions)

28.02.2026 19:01 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

You described LLMs solving problems by "whatever is near in the latent space". As if the word that if "mother" is encoded in a hidden state then the next word that comes should be "mama", "mammy", or any other position right next to "mother". It's not some sort of random latent walk.

28.02.2026 19:01 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

Why does it matter that you leave out the largest part of how LLMs work?

Really?

28.02.2026 18:46 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

The fact that what you're describing isn't LLMs at all? You're describing taking latent spaces and then just jumping to nearby latent positions, as if the FFNs don't exist at all - when in reality they make up the vast majority of the parameters of any model.

The next latent comes from DNNs.

28.02.2026 18:22 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

(And the other bots aren't shown later because they're never in frame)

I think there is a bit of an assumption that you're familiar with FIRST. :)

28.02.2026 18:08 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0
FIRST Robotics Finals 2009
YouTube video by Lunar XPRIZE FIRST Robotics Finals 2009

The first frame with the blueprint is a diagram of their robot. Their robot has two parts which immediately disconnect (frame 3).

If you mean frame 2, FIRST competitions can take different forms - here was the last competition before this was drawn:

www.youtube.com/watch?v=bDo5...

28.02.2026 18:08 โ€” ๐Ÿ‘ 1    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

It's not merely "fact lookup", it's "(fuzzy) logical inference performance". Facts are indeed stored in the FFNs, but so are the conditions in which those facts are the result of the input. And the logical rules on which deductions are made each layer are mind-bogglingly complex.

28.02.2026 17:59 โ€” ๐Ÿ‘ 1    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0
Why Neural Networks can learn (almost) anything
YouTube video by Emergent Garden Why Neural Networks can learn (almost) anything

Which in turn are built off DNNs (Transformers FFNs are DNNs), so to understand how *they* learn and perform fuzzy logical deductions, I'd recommend:

www.youtube.com/watch?v=0Qcz...

28.02.2026 17:56 โ€” ๐Ÿ‘ 2    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0
Preview
Zoom In: An Introduction to Circuits By studying the connections between neurons, we can find meaningful algorithms in the weights of neural networks.

To understand how circuits are built up from the base, I recommend:

distill.pub/2020/circuit...

28.02.2026 17:55 โ€” ๐Ÿ‘ 2    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0
Scaling Monosemanticity: Extracting Interpretable Features from Claude 3 Sonnet

Then understanding that each of these very high-level circuits is built on simpler circuits, iteratively down with each layer. E.g.:

transformer-circuits.pub/2024/scaling...

28.02.2026 17:55 โ€” ๐Ÿ‘ 2    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0
Preview
On the Biology of a Large Language Model We investigate the internal mechanisms used by Claude 3.5 Haiku โ€” Anthropic's lightweight production model โ€” in a variety of contexts, using our circuit tracing methodology.

To understand LLMs, I recommend first starting here for the highest-level view:

transformer-circuits.pub/2025/attribu...

28.02.2026 17:53 โ€” ๐Ÿ‘ 2    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

I wouldn't share that thread, IMHO.

bsky.app/profile/nafn...

28.02.2026 17:52 โ€” ๐Ÿ‘ 4    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

LLMs are not "a hyperdimensional map of language use". *Latents* are a hyperdimensional map of *concepts*. This map is generated from (among other things, but usually no longer exclusively) language, but there's nothing linguistic about it.

But again, LLMs are not merely latents.

28.02.2026 17:51 โ€” ๐Ÿ‘ 2    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

FFNs don't encode into the subsequent latent "whatever is nearest"; they function as detector-generators - *detecting* combinations of concepts in the latent, and then *generating* whatever concepts are the logical deductions from the combinations of concepts detected.

28.02.2026 17:50 โ€” ๐Ÿ‘ 2    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

This is not how LLMs work. You seem to understand latent spaces but you're forgetting about the FFNs. LLMs are not simply latent spaces; latent spaces just hold the conceptual representation that the FFNs work on.

28.02.2026 17:48 โ€” ๐Ÿ‘ 2    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 1

It's-a-me! Dario! I'm 'a gonna be 'a responsible with'a me'a models!

28.02.2026 01:28 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

Maybe we should ask his sister about his character...

28.02.2026 01:15 โ€” ๐Ÿ‘ 1    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0
Post image

Btw, if you are following the Anthropic debacle, and Minnesota is fresh in your mind, consider the limits your government was not willing to budge on, according to the Anthropic CEO: www.anthropic.com/news/stateme...

28.02.2026 00:26 โ€” ๐Ÿ‘ 242    ๐Ÿ” 82    ๐Ÿ’ฌ 10    ๐Ÿ“Œ 10