agree on the latter point, cf bsky.app/profile/tkuk...
on the former - it's costly if one assumes grok invocation per page load. but IMO scaling a grok-twitter-recsys is amenable to good system dimensioning (e.g. model cascade + periodic batching)
from that perspective, doesn't seem preposterous π€·
18.10.2025 16:45 β π 3 π 0 π¬ 1 π 0
how would satya sell enterprise AI without text extraction
18.10.2025 15:35 β π 0 π 0 π¬ 0 π 0
our entire economy depends on document scanning for training data.
18.10.2025 15:31 β π 0 π 0 π¬ 1 π 0
iirc @neuroai.bsky.social make a nice remark at some point during the learning salon. paraphrasing from memory, "is the real AGI DeepMind's accumulated bias + methods or are we _truly_ building a general learner"
18.10.2025 09:34 β π 1 π 0 π¬ 0 π 0
"long-context" and "prompt eng" are flip-sides of the same underlying concept. _someone_ has to compress the representation, either in language- or high-d-vector space.
IMO the domain model is intelligence as part user (akin to "phds doing lin reg" in finance), part interface (RLHF), part model
21.09.2025 13:18 β π 0 π 0 π¬ 0 π 0
interesting but I guess not a relevant attack vector (assuming it exclusively influences user-local results).
becomes relevant in some intricate global-recommender-system situation but google already has years of dealing with this behind them.
17.08.2025 19:08 β π 1 π 0 π¬ 0 π 0
how do you here define _frontier work_? most work I do at least seems a combination of monotonous incantations and real insight. insight usually being ~ "this will affect outcomes I care about N steps down the road". && frequently this is preempted by me just-realizing which outcomes I care about :)
18.07.2025 11:50 β π 0 π 0 π¬ 1 π 0
but also sounds like a cool research project. rev-eng how much google3 code in gemini pre-training based on relative quality of sth like gemini vs gpt4 or claude.
(eg via diff-in-diffs perf on coding tasks nodejs vs swift)
(I assume gem was predominantly google3-trained based on its coding quirks)
13.07.2025 10:51 β π 0 π 0 π¬ 0 π 0
Open-Endedness is Essential for Artificial Superhuman Intelligence
In recent years there has been a tremendous surge in the general capabilities of AI systems, mainly fuelled by training foundation models on internetscale data. Nevertheless, the creation of openended...
novelty = f(audience, goal)?
there's a nice exposition of this topic in @rockt.ai and others' position paper arxiv.org/abs/2406.042...
which, to prove a meta-point, delivers a rather "obvious" message with exposition clarity deemed novel enough for ICML poster acceptance :)
12.07.2025 19:09 β π 1 π 0 π¬ 0 π 0
"dynamic" as discussed here sounds to me more akin to what people would refer to as "reasoning" fine-tuning nowadays?
RAG implies external memory; reasoning post-trained model generates more artefacts (/tokens/"writes") as a result of computation (which is also where "dynamic" makes a difference)
10.06.2025 14:02 β π 0 π 0 π¬ 0 π 0
Humans demonstrating OOD failure, AI demonstrating IID success :)
bsky.app/profile/mela...
30.05.2025 17:06 β π 1 π 0 π¬ 0 π 0
YouTube video by Krzysztof A. Janczak
A message from Ira Glass that every Artist should hear...
"its normal and the most important thing you can do is do a lot of work" :)
it would be intriguing to see the progress if you're willing to share at some point!
29.05.2025 07:53 β π 0 π 0 π¬ 0 π 0
related and belated followup: opinion/rant on similar topic
> Every app has a different design [optimized based on your activity ...] each app trying to make you do different things in uniquely annoying ways [...] low-quality clickbait
anyhow we all know how it goes
25.05.2025 18:08 β π 2 π 1 π¬ 0 π 0
YouTube video by Case Western Reserve University
Kurt Vonnegut Lecture
I mean, you've described a lot of content. "low entropy" does some moderately-heavy lifting, but in essence www.youtube.com/@kurzgesagt videos are also this way.
all of art ultimately exists on a particular grounded low entropy axis (nice example by Kurt Vonnegut youtu.be/4_RUgnC1lm8?...)
25.05.2025 10:55 β π 0 π 0 π¬ 0 π 0
agree re/feeds
what you'd normally do for a closed-platform is (1) scrape posts, (2) write-your-own ranking/clf. algo to get better content (tags are useful but easily co-opted, the platform's recsys is necessarily deficient due to partial obs).
_native_ support for scaling (2) I find exciting
25.05.2025 08:51 β π 2 π 0 π¬ 1 π 1
π―, people complain to overuse of mathematical formalisms in papers but a less expressive language enforces precision.
(in which sense, I guess a similar analogy exists: internal monologue vs. writing)
21.05.2025 16:52 β π 1 π 0 π¬ 0 π 0
How was it, "if I had more time I would've written a shorter letter"... Cursor has all the time in the world, now to incentivize compression. Twitter v0 had the right idea :)
15.05.2025 16:58 β π 0 π 0 π¬ 0 π 0
How many turns into the conversation? Any way to share the full trace?
05.05.2025 05:45 β π 0 π 0 π¬ 1 π 0
composition (in standard unixy sense), decomposition (use part of functionality and pipe into another program), misappropriation (eg use excel as coloring tool), environment-driven personalization (e.g. use OS-level accessibility tools)
01.05.2025 21:05 β π 8 π 0 π¬ 1 π 0
+1, I conceptualize as generalized state of flow beyond visceral fleeting moments. prob why env design (school, "WFH v onsite", etc) matters & likely down to energy expense: helper.ipam.ucla.edu/publications...
> Control energy for state transitions decreases over the course of repeated task trials
19.04.2025 16:08 β π 1 π 0 π¬ 0 π 0
are you curious about oai index/retrieval quality or aiming at something else?
18.04.2025 17:59 β π 0 π 0 π¬ 1 π 0
Could you share a few? I feel like Goodhart is well interpreted in CS/ML, curious for specific counterexamples.
10.04.2025 19:31 β π 0 π 0 π¬ 0 π 0
the data transform + API integration is great, but how do I tell it about function composition :'(
28.03.2025 19:48 β π 0 π 0 π¬ 0 π 0
also great model of why modern LLMs work.
user intent is specified in the query.
informed re-representation (his example: "cat chased by dog" vs "dog chases cat") massively benefits info transmission even when no new information is generated.
20.03.2025 22:36 β π 0 π 0 π¬ 0 π 0
@andymatuschak.org's writing is relevant (both on learn-to-learn e.g. [1] and personal xp, e.g. timely [2]). I think immediate-feedback learning style should become default w/LMs. quick feedback loops is how you win any system :)
[1]: andymatuschak.org/prompts/
[2]: bsky.app/profile/andy...
10.03.2025 14:49 β π 3 π 0 π¬ 0 π 0
what were the assignments about? encouraging review via daily journal habit (e.g. essay "what we learned today") since early ed seems immensely useful.
10.03.2025 08:59 β π 1 π 0 π¬ 1 π 0
Does it fully matter? Student learns to use AI tools. Could be two way learning. Teacher is around mainly to enforce meta cognition and reflection.
09.03.2025 07:46 β π 1 π 0 π¬ 0 π 0
yes! agree it's a different psych manifestation.
but AFAIK "decision fatigue" usually implies deteriorating decision quality.
so counterfactual is ~ "did my problem-solving deteriorate less after 8h study relative to my decision making ability after 8h work".
which... not sure if true.
07.03.2025 19:59 β π 1 π 0 π¬ 1 π 0
Professor of Applied Econometrics and Policy Evaluation at the University of Fribourg/Freiburg (Switzerland) - causal analysis, statistics, econometrics, machine learning...and telemarking
I integrate insights and techniques from machine learning into the econometric toolbox.
https://gsb-faculty.stanford.edu/jann-spiess
teaching computers how to see
Faculty at βͺthe ELLIS Institute TΓΌbingen and Max Planck Institute for Intelligent Systems. Leading the AI Safety and Alignment group. PhD from EPFL supported by Google & OpenPhil PhD fellowships.
More details: https://www.andriushchenko.me/
Associate Professor, Department of Psychology, Harvard University. Computation, cognition, development.
AI + security | Stanford PhD in AI & Cambridge physics | techno-optimism + alignment + progress + growth | πΊπΈπ¨πΏ
web: http://maxim.ece.illinois.edu
substack: https://realizable.substack.com
Adds labels for repositories you contribute to (max 4)
Setup Instructions:
1. Follow this labeler (hint: click the "..." menu > "Follow")
2. Subscribe to the labeler
3. Like the labeler
All steps required. #3 sends a DM to continue setup
Pioneering a new generation of LLMs.
My opinions only here.
π¨βπ¬ RS DeepMind
Past:
π¨βπ¬ R Midjourney 1y π§βπ DPhil AIMS Uni of Oxford 4.5y
π§ββοΈ RE DeepMind 1y πΊ SWE Google 3y π TUM
π€ @nwspk
Marrying classical CV and Deep Learning. I do things, which work, rather than being novel, but not working.
http://dmytro.ai
Neuroimager @SPMIC-UoN.bsky.social Professor @uniofnottingham.bsky.social
Brain MRI methods developer. Father of two.
https://conilab.nottingham.ac.uk
Research scientist in machine learning at Google DeepMind, theatre director and co-founder of Improbotics, visiting researcher at Goldsmiths. Ex-Bell Labs, Microsoft, PhD at NYU. Work in AI for weather and climate & AI for artistic creativity. he/him.
Chief Scientist at the UK AI Security Institute (AISI). Previously DeepMind, OpenAI, Google Brain, etc.
PhD supervised by Tim RocktΓ€schel and Ed Grefenstette, part time at Cohere. Language and LLMs. Spent time at FAIR, Google, and NYU (with Brenden Lake). She/her.
ML research group @universityofoxford.bsky.social. Focussed on multi-agent, open-ended, meta and reinforcement learning as well as agent based models. More at http://foersterlab.com.
Professor at Wharton, studying AI and its implications for education, entrepreneurship, and work. Author of Co-Intelligence.
Book: https://a.co/d/bC2kSj1
Substack: https://www.oneusefulthing.org/
Web: https://mgmt.wharton.upenn.edu/profile/emollick
Senior Director, Research Scientist @ Meta FAIR + Visiting Prof @ NYU.
Pretrain+SFT: NLP from Scratch (2011). Multilayer attention+position encode+LLM: MemNet (2015). Recent (2024): Self-Rewarding LLMs & more!