Toni Kukurin's Avatar

Toni Kukurin

@tkukurin.bsky.social

life is short and I am in a hurry. tkukurin.github.io

24 Followers  |  144 Following  |  58 Posts  |  Joined: 30.07.2023  |  2.3308

Latest posts by tkukurin.bsky.social on Bluesky

agree on the latter point, cf bsky.app/profile/tkuk...

on the former - it's costly if one assumes grok invocation per page load. but IMO scaling a grok-twitter-recsys is amenable to good system dimensioning (e.g. model cascade + periodic batching)

from that perspective, doesn't seem preposterous 🀷

18.10.2025 16:45 β€” πŸ‘ 3    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

how would satya sell enterprise AI without text extraction

18.10.2025 15:35 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

our entire economy depends on document scanning for training data.

18.10.2025 15:31 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

iirc @neuroai.bsky.social make a nice remark at some point during the learning salon. paraphrasing from memory, "is the real AGI DeepMind's accumulated bias + methods or are we _truly_ building a general learner"

18.10.2025 09:34 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

"long-context" and "prompt eng" are flip-sides of the same underlying concept. _someone_ has to compress the representation, either in language- or high-d-vector space.

IMO the domain model is intelligence as part user (akin to "phds doing lin reg" in finance), part interface (RLHF), part model

21.09.2025 13:18 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

interesting but I guess not a relevant attack vector (assuming it exclusively influences user-local results).

becomes relevant in some intricate global-recommender-system situation but google already has years of dealing with this behind them.

17.08.2025 19:08 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

how do you here define _frontier work_? most work I do at least seems a combination of monotonous incantations and real insight. insight usually being ~ "this will affect outcomes I care about N steps down the road". && frequently this is preempted by me just-realizing which outcomes I care about :)

18.07.2025 11:50 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

but also sounds like a cool research project. rev-eng how much google3 code in gemini pre-training based on relative quality of sth like gemini vs gpt4 or claude.

(eg via diff-in-diffs perf on coding tasks nodejs vs swift)

(I assume gem was predominantly google3-trained based on its coding quirks)

13.07.2025 10:51 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Preview
Open-Endedness is Essential for Artificial Superhuman Intelligence In recent years there has been a tremendous surge in the general capabilities of AI systems, mainly fuelled by training foundation models on internetscale data. Nevertheless, the creation of openended...

novelty = f(audience, goal)?

there's a nice exposition of this topic in @rockt.ai and others' position paper arxiv.org/abs/2406.042...

which, to prove a meta-point, delivers a rather "obvious" message with exposition clarity deemed novel enough for ICML poster acceptance :)

12.07.2025 19:09 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

"dynamic" as discussed here sounds to me more akin to what people would refer to as "reasoning" fine-tuning nowadays?

RAG implies external memory; reasoning post-trained model generates more artefacts (/tokens/"writes") as a result of computation (which is also where "dynamic" makes a difference)

10.06.2025 14:02 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Preview
FineWeb: decanting the web for the finest text data at scale - a Hugging Face Space by HuggingFaceFW This application helps create large, high-quality datasets for training large language models by processing and filtering web data. Users can obtain datasets with optimized deduplication and filter...

not sure about a single unifying tutorial textbook; however closest might be something in the huggingface/anthropic/... blogpost genre... eg. what comes to mind in data filtering/collection space
huggingface.co/spaces/Huggi...

03.06.2025 18:04 β€” πŸ‘ 7    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

Humans demonstrating OOD failure, AI demonstrating IID success :)

bsky.app/profile/mela...

30.05.2025 17:06 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
A message from Ira Glass that every Artist should hear...
YouTube video by Krzysztof A. Janczak A message from Ira Glass that every Artist should hear...

"its normal and the most important thing you can do is do a lot of work" :)

it would be intriguing to see the progress if you're willing to share at some point!

29.05.2025 07:53 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

related and belated followup: opinion/rant on similar topic

> Every app has a different design [optimized based on your activity ...] each app trying to make you do different things in uniquely annoying ways [...] low-quality clickbait

anyhow we all know how it goes

25.05.2025 18:08 β€” πŸ‘ 2    πŸ” 1    πŸ’¬ 0    πŸ“Œ 0
Kurt Vonnegut Lecture
YouTube video by Case Western Reserve University Kurt Vonnegut Lecture

I mean, you've described a lot of content. "low entropy" does some moderately-heavy lifting, but in essence www.youtube.com/@kurzgesagt videos are also this way.

all of art ultimately exists on a particular grounded low entropy axis (nice example by Kurt Vonnegut youtu.be/4_RUgnC1lm8?...)

25.05.2025 10:55 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

agree re/feeds

what you'd normally do for a closed-platform is (1) scrape posts, (2) write-your-own ranking/clf. algo to get better content (tags are useful but easily co-opted, the platform's recsys is necessarily deficient due to partial obs).

_native_ support for scaling (2) I find exciting

25.05.2025 08:51 β€” πŸ‘ 2    πŸ” 0    πŸ’¬ 1    πŸ“Œ 1

πŸ’―, people complain to overuse of mathematical formalisms in papers but a less expressive language enforces precision.

(in which sense, I guess a similar analogy exists: internal monologue vs. writing)

21.05.2025 16:52 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

How was it, "if I had more time I would've written a shorter letter"... Cursor has all the time in the world, now to incentivize compression. Twitter v0 had the right idea :)

15.05.2025 16:58 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Preview
GitHub - twitter/the-algorithm: Source code for Twitter's Recommendation Algorithm Source code for Twitter's Recommendation Algorithm - twitter/the-algorithm

despite all those heroic efforts to open source "the algorithm"?! shocker

15.05.2025 07:24 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

How many turns into the conversation? Any way to share the full trace?

05.05.2025 05:45 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

composition (in standard unixy sense), decomposition (use part of functionality and pipe into another program), misappropriation (eg use excel as coloring tool), environment-driven personalization (e.g. use OS-level accessibility tools)

01.05.2025 21:05 β€” πŸ‘ 8    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

+1, I conceptualize as generalized state of flow beyond visceral fleeting moments. prob why env design (school, "WFH v onsite", etc) matters & likely down to energy expense: helper.ipam.ucla.edu/publications...

> Control energy for state transitions decreases over the course of repeated task trials

19.04.2025 16:08 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

are you curious about oai index/retrieval quality or aiming at something else?

18.04.2025 17:59 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

Could you share a few? I feel like Goodhart is well interpreted in CS/ML, curious for specific counterexamples.

10.04.2025 19:31 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Post image

the data transform + API integration is great, but how do I tell it about function composition :'(

28.03.2025 19:48 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

also great model of why modern LLMs work.

user intent is specified in the query.

informed re-representation (his example: "cat chased by dog" vs "dog chases cat") massively benefits info transmission even when no new information is generated.

20.03.2025 22:36 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

@andymatuschak.org's writing is relevant (both on learn-to-learn e.g. [1] and personal xp, e.g. timely [2]). I think immediate-feedback learning style should become default w/LMs. quick feedback loops is how you win any system :)

[1]: andymatuschak.org/prompts/
[2]: bsky.app/profile/andy...

10.03.2025 14:49 β€” πŸ‘ 3    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

what were the assignments about? encouraging review via daily journal habit (e.g. essay "what we learned today") since early ed seems immensely useful.

10.03.2025 08:59 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

Does it fully matter? Student learns to use AI tools. Could be two way learning. Teacher is around mainly to enforce meta cognition and reflection.

09.03.2025 07:46 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

yes! agree it's a different psych manifestation.

but AFAIK "decision fatigue" usually implies deteriorating decision quality.

so counterfactual is ~ "did my problem-solving deteriorate less after 8h study relative to my decision making ability after 8h work".

which... not sure if true.

07.03.2025 19:59 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

@tkukurin is following 20 prominent accounts