Youβve probably heard this before, but chances are that you know a decent amount of math, just in a different language. E.g. if you can code well, clearly youβre good at symbol manipulation (most of coding is that, in a sense).
Math is like a different (and at times esoteric) programming language.
09.02.2026 22:08 β
π 3
π 0
π¬ 0
π 0
β¦ [2/2] and (b) how easy is it to extract, or βrenderβ those answers.
The true craft of scientific writing is figuring out these functions. Everything else (LLMs or not) is irrelevant.
09.02.2026 21:49 β
π 0
π 0
π¬ 0
π 0
I actually think this is the wrong way to think about (scientific) writing.
The purpose of a scientific document is to help answer questions about some topic.
The quality of the document is a function of (a) how large is the space of questions you cover, β¦ [1/2]
09.02.2026 21:49 β
π 0
π 0
π¬ 1
π 0
I defended my PhD today, and would like to take this moment to thank the folks who made it worth the journey -- ya'll are awesome.
14.01.2026 15:42 β
π 1
π 0
π¬ 0
π 0
IMO the practical value of impossibility theorems often lie in formulating those assumptions β they might reveal what conditions are probably violated in practice (if the statement of the proof is intuitively off). Eg. the IIA assumption in Arrowβs Impossibility Theorem.
13.01.2026 21:30 β
π 0
π 0
π¬ 0
π 0
I do wonder how much of that fat tail is lost in simulation, and what the downstream effects are. Model autophagy disorder is not something youβd want for models making life or death decisions.
07.11.2025 06:46 β
π 1
π 0
π¬ 0
π 0
The purpose of a document is to help an LLM answer questions about said document.
06.07.2025 15:33 β
π 1
π 0
π¬ 0
π 0
Imagine where we would be today if an overpowered Nazi Germany did not have US and USSR to counterbalance. Or just watch Man in the High Castle.
29.01.2025 22:40 β
π 1
π 0
π¬ 0
π 0
Dario Amodei β On DeepSeek and Export Controls
On DeepSeek and Export Controls
Re: Darioβs post (darioamodei.com/on-deepseek-...) β the current state of the US has left me scared shitless about a unipolar world. A single, powerful pole with compromised institutions is the worst possible outcome for human civilization.
29.01.2025 22:36 β
π 3
π 0
π¬ 1
π 0
Iβm low key disappointed r1 doesnβt swear in its CoT.
Tired: βWaitβ¦β
Wired: βhol the fuck upβ
25.01.2025 23:10 β
π 1
π 0
π¬ 0
π 0
And in that world, the quality of the weapon is how cheap it runs for how good it is.
I wonβt be surprised if nation states eventually train and host their own models. Heck, some LLM shops seem to be betting on that, e.g. Mistral and AlephAlpha with βEuropean LLMsβ.
17.01.2025 22:56 β
π 0
π 0
π¬ 0
π 0
Weβre approaching an era of memetic warfare where LLMs are the weapons. Weβre not there yet β the values espoused by Chinese LLMs arenβt all that different from American ones β but thatβs for now.
But once LLMs become our primary interface with the outside world, itβs bound to happen.
17.01.2025 22:56 β
π 1
π 0
π¬ 1
π 0
Looks nice! Some FastAPI endpoints + a docker image should help adoption. :)
17.01.2025 22:02 β
π 0
π 0
π¬ 0
π 0
The @jina.ai team is low key cracked. No yapping just shipping. π«‘
12.01.2025 22:20 β
π 0
π 0
π¬ 0
π 0
Finished the first season of Severance. Brb implementing memory for my LLM agents.
08.01.2025 22:50 β
π 0
π 0
π¬ 0
π 0
This is fascinating work, congratulations!
Question: the point that architectural constraints (locality + equivariance) are sufficient is well demonstrated.
But do you think it is necessary? I.e. would you expect a diffusion transformer to learn these constraints?
01.01.2025 10:41 β
π 1
π 0
π¬ 0
π 0
In other words, code 1 is more βmulti-agentβ than others.
What do I mean when I say βagentβ? A system that weβd like to abstract away like a black box (Rovelliβs definition). Of that, I count three in code 1, and 1 in both codes 2 and 3.
21.12.2024 22:39 β
π 2
π 0
π¬ 0
π 0
Because code 1 is most explicit about the structure of the computational graph. :)
bsky.app/profile/nasi...
21.12.2024 22:29 β
π 2
π 0
π¬ 1
π 0
If I were to force an answer, Iβd say code 1 (prompt chaining) has more agent energy than the others.
20.12.2024 23:35 β
π 0
π 0
π¬ 1
π 0
Claude is something special.
16.12.2024 22:39 β
π 0
π 0
π¬ 0
π 0
Have the token-level LLM predict βconcept tokensβ. The hidden states for these tokens go in to an adapter, and out come concept representations. Concept tokens attend to previous concept tokens, and perhaps also the span between itself and the previous concept token.
14.12.2024 05:43 β
π 0
π 0
π¬ 1
π 0
Very cool work from Meta AI: Large Concept Models. Idea: autoregress in the space of sentence level representations.
I think an interesting next step would be to layer this on conventional LLMs / token prediction models.
Hereβs how that could work: ‡οΈ
14.12.2024 05:43 β
π 1
π 0
π¬ 1
π 0
GitHub - microsoft/markitdown: Python tool for converting files and office documents to Markdown.
Python tool for converting files and office documents to Markdown. - microsoft/markitdown
Microsoft just released a tool that lets you convert Office files to Markdown. Never thought I'd see the day.
Google also added Markdown export to Google Docs a few months ago.
github.com/microsoft/markitdown
13.12.2024 20:25 β
π 524
π 127
π¬ 24
π 24
The idea is to model two things:
(a) if concepts fit together to make good art, and
(b) if people have already thought about that combination of concepts (βcognitive availabilityβ).
Seek out the combos for which (a) is true but (b) isnβt, and ask a text-to-image model to render that.
π§΅β€΅οΈ
13.12.2024 09:41 β
π 0
π 0
π¬ 1
π 0
My MSc advisee (& gang) cooked.
tl:dr β a cute technique to get machines to be more creative.
π§΅β€΅οΈ
13.12.2024 09:41 β
π 1
π 0
π¬ 1
π 0
If you arenβt already doing this: share that pdf with Claude.
07.12.2024 22:50 β
π 0
π 0
π¬ 0
π 0