But, as Iβve heard from others, the AIs often suggest possible connections, related results, or avenues to pursue that hadnβt occured to me. Unfortunately, these are usually dead ends.
29.12.2025 17:25 β π 1 π 0 π¬ 1 π 0@rdnowak.bsky.social
Director of the Center for the Advancement of Progress
But, as Iβve heard from others, the AIs often suggest possible connections, related results, or avenues to pursue that hadnβt occured to me. Unfortunately, these are usually dead ends.
29.12.2025 17:25 β π 1 π 0 π¬ 1 π 0Iβve found that the pro-version AIs are great for proving known theorems or new theorems that could be considered homework problems, but so far I have had no success using them to solve truly open/novel/challenging math problems.
29.12.2025 17:08 β π 2 π 0 π¬ 1 π 0Since I canβt get it out of my head, I wrote up my thoughts on @kevinbaker.bsky.social's critique of AI-automated science and the logical end of processes that can't self-correct.
15.12.2025 15:33 β π 44 π 9 π¬ 3 π 2Kevin Baker's essay is probably the best thing I have read in 2025.
15.12.2025 23:42 β π 1 π 0 π¬ 0 π 0Yes. Just write your thoughts in a rough and unpolished form, say rough paragraphs that contain terse points you want to make. then let 'er rip
31.10.2025 19:21 β π 1 π 0 π¬ 1 π 0Section 7 is a wonderful description of the process they went through.
25.10.2025 15:57 β π 1 π 0 π¬ 1 π 0something just isn't fully clicking. if you look at total yards and time of possession, they should have blown them out. well, better anyway to peak later in season, so let's hope that's what happens (like two seasons ago)
13.10.2025 02:09 β π 1 π 0 π¬ 0 π 0Packers get the win, but it wasn't pretty.
13.10.2025 00:45 β π 0 π 0 π¬ 1 π 0Thanks for participating and presenting your work!
08.09.2025 17:14 β π 2 π 0 π¬ 0 π 0Google promotes box shirts too
05.09.2025 18:19 β π 1 π 0 π¬ 0 π 0Pour into
27.08.2025 14:36 β π 2 π 0 π¬ 0 π 0Announcing the first workshop on Foundations of Language Model Reasoning (FoRLM) at NeurIPS 2025!
πSoliciting abstracts that advance foundational understanding of reasoning in language models, from theoretical analyses to rigorous empirical studies.
π Deadline: Sept 3, 2025
Nice article about my momβs new book shepherdexpress.com/culture/book...
04.08.2025 01:21 β π 3 π 0 π¬ 0 π 0βthe only way to predict or to control the functioning of such systems is by an intricate system of charms, spells, and incantationsβ
05.07.2025 16:01 β π 2 π 0 π¬ 0 π 0See you there!
21.06.2025 15:23 β π 2 π 0 π¬ 0 π 0More likely midges. The truest sign of a healthy ecosystem
16.05.2025 22:55 β π 1 π 0 π¬ 1 π 0Looking forward to a great MMLS!
25.04.2025 12:21 β π 3 π 0 π¬ 0 π 0This is collaboration with Ziyue Luo, @shroffness and @kevinlauka
07.02.2025 02:55 β π 0 π 0 π¬ 0 π 0Jifanβs on the industry job market now, and his expertise in efficient training, distillation, and data curation couldn't be more timely. Feel free to reach out to him at jifan@cs.wisc.edu.
π Paper: arxiv.org/abs/2410.02755
SIEVE improves upon existing quality filtering methods in the DataComp-LM challenge, producing better LLM pretraining data that led to improved model performance.
This work is part of Jifan's broader research on efficient ML training, from active learning to label-efficient SFT for LLMs.
Why does this matter? High-quality data is the bedrock of LLM training. SIEVE enables filtering trillions of web data for specific domains like medical/legal text with customizable natural language prompts.
07.02.2025 02:55 β π 0 π 0 π¬ 1 π 0SIEVE distills GPT-4's data filtering capabilities into lightweight models at <1% of the cost. Not just minor improvements - we're talking 500x more efficient filtering operations.
07.02.2025 02:55 β π 0 π 0 π¬ 1 π 0π§΅ Heard all the buzz around distilling from OpenAI models? Check out @jifanz's latest work SIEVE - showing how strategic distillation can make LLM development radically more cost-effective while matching quality.
07.02.2025 02:55 β π 4 π 0 π¬ 1 π 1Maybe Trump should have read my mom's book: "For the first six weeks, the embryo, whether XX or XY, coasts along in sexual ambiguity." p. 25
23.01.2025 00:25 β π 2 π 0 π¬ 0 π 0Task vectors are akin to punchcards: you feed them to your LLM and it implements specific tasks, without in-context demonstrations. Liu's new paper examines at what scale, where in the network and when during training do they emerge, and how to encourage their emergence.
arxiv.org/pdf/2501.09240
Good luck with that
04.01.2025 01:20 β π 0 π 0 π¬ 1 π 0p.s. we don't know for sure if I said this or not
04.01.2025 00:36 β π 1 π 0 π¬ 1 π 0Is the solution treating everything electronic as "fake"?
Maybe?