YouTube video by The Piano Experience
Glenn Gould plays Chopin Piano Sonata No. 3 in B minor Op.58
Gould mostly recorded baroque and early classical. He only recorded a single Chopin piece, as a one-off broadcast. But like many of his efforts, it's profoundly thought provoking, the end product as much Gould as it is Chopin. I love the last mvt (20:55+). www.youtube.com/watch?v=NAHE...
29.07.2025 17:44 β π 4 π 0 π¬ 0 π 0
I don't think those things seem boring. But most research directions honestly are quite boring, because they are geared towards people pleasing --- going with the herd, seeking approval from others, and taking no risks. It's a great way to avoid making a contribution that changes any minds.
29.07.2025 02:39 β π 1 π 0 π¬ 1 π 0
Whatever you do, just don't be boring.
28.07.2025 23:15 β π 4 π 0 π¬ 1 π 1
YouTube video by LoG Meetup NYC
It's Time to Say Goodbye to Hard (equivariance) Constraints - Andrew Gordon Wilson
I had a great time presenting "It's Time to Say Goodbye to Hard Constraints" at the Flatiron Institute. In this talk, I describe a philosophy for model construction in machine learning. Video now online! www.youtube.com/watch?v=LxuN...
22.07.2025 19:28 β π 13 π 2 π¬ 0 π 0
Excited to be presenting my paper "Deep Learning is Not So Mysterious or Different" tomorrow at ICML, 11 am - 1:30 pm, East Exhibition Hall A-B, E-500. I made a little video overview as part of the ICML process (viewable from Chrome): recorder-v3.slideslive.com#/share?share...
17.07.2025 00:16 β π 26 π 5 π¬ 0 π 1
While scaling laws typically predict the final loss, we show that good scaling rules enable accurate predictions of entire loss curves of larger models from smaller ones! Shikai Qiu did an amazing job leading the paper, in collaboration with L. Xiao, J. Pennington, A. Agarwala. 3/3
08.07.2025 14:45 β π 2 π 0 π¬ 0 π 0
In particular, scaling collapse allows us to transfer insights from experiments conducted at a very small scale to much larger models! Much more in the paper, including supercollapse: collapse between curves less than the noise floor of per-model loss curves across seeds. 2/3
08.07.2025 14:45 β π 2 π 0 π¬ 1 π 0
Our new ICML paper discovers scaling collapse: through a simple affine transformation, whole training loss curves across model sizes with optimally scaled hypers collapse to a single universal curve! We explain the collapse, providing a diagnostic for model scaling.
arxiv.org/abs/2507.02119
1/3
08.07.2025 14:45 β π 32 π 5 π¬ 3 π 0
I love havarti!
02.07.2025 15:52 β π 2 π 0 π¬ 0 π 0
It sometimes feels like papers are written this way. <Make claim that may or may not be true but aligns with the paper's narrative> <find arbitrary reference that supposedly supports that claim, but may be making a different point entirely>. I guess grammarly is giving the people what they want?
26.06.2025 10:23 β π 1 π 0 π¬ 0 π 0
Excited about our new ICML paper, showing how algebraic structure can be exploited for massive computational gains in population genetics.
25.06.2025 14:06 β π 4 π 2 π¬ 0 π 0
Machine learning is perhaps the only discipline that has become less mature over time. A reverse metamorphosis, from butterfly to caterpillar.
24.06.2025 22:11 β π 22 π 3 π¬ 1 π 0
AI this, AI that, the implications of AI for X... can we just never talk about AI again?
17.06.2025 20:27 β π 9 π 0 π¬ 1 π 0
Really excited about our new paper, "Why Masking Diffusion Works: Condition on the Jump Schedule for Improved Discrete Diffusion". We explain the mysterious success of masking diffusion to propose new diffusion models that work well in a variety settings, including proteins, images, and text!
16.06.2025 14:29 β π 6 π 0 π¬ 0 π 0
What's irrational is the idea that some group of authors writing a paper about something foundational also should be the team of people to put it into production in the real world and demonstrate its impact, all in one paper. That happens over years, and involves different interests and skills.
16.06.2025 14:26 β π 4 π 1 π¬ 1 π 0
I find that this position is often more emotionally rooted than rational. It makes no sense to expect a paper on foundations to demonstrate significant real-world impact. As you say, it's a cumulative process carried out by different people, over time.
16.06.2025 09:43 β π 3 π 0 π¬ 1 π 0
Sorry to miss it! Currently in Cambridge UK for a Newton Institute programme on uncertainty representation.
16.06.2025 09:14 β π 1 π 0 π¬ 1 π 0
Great topic for a workshop!
15.06.2025 21:35 β π 3 π 0 π¬ 1 π 0
YouTube video by Lex Fridman
Terence Tao: Hardest Problems in Mathematics, Physics & the Future of AI | Lex Fridman Podcast #472
A really outstanding interview of Terence Tao, providing an introduction to many topics, including the math of general relativity (youtube.com/watch?v=HUkB...). I love relativity, and in a recent(ish) paper we also consider the wave maps equation (section 5, arxiv.org/abs/2304.14994).
15.06.2025 20:25 β π 14 π 2 π¬ 0 π 1
AI benchmarking culture is completely out of control. Tables with dozens of methods, datasets, and bold numbers, trying to answer a question that perhaps no one should be asking anymore.
30.05.2025 21:55 β π 19 π 5 π¬ 1 π 1
We have a strong bias to overestimate the speed of technological innovation and impact. See past claims about autonomous driving, AI curing diseases... or the timeline in every sci-fi book ever written. Where is my flying car?
05.05.2025 22:28 β π 9 π 1 π¬ 1 π 0
Mad not to have Hey Jude or Yesterday.
25.04.2025 01:42 β π 1 π 0 π¬ 0 π 0
Still not sure I totally understand what you're saying. Why is it unfortunate that it doesn't fit with academia or industry?
17.04.2025 03:27 β π 2 π 0 π¬ 0 π 0
That said, if you're saying that there are a space of projects that have value but that value is not easily recognized with traditional structures (academia, industry), I certainly agree.
16.04.2025 22:28 β π 2 π 0 π¬ 1 π 0
It all depends. GPyTorch, BoTorch, and CoLA have not only facilitated but actually suggested to us some of our best research ideas. Google and Meta also must feel they benefit from Tensorflow, JAX, and PyTorch being open source.
16.04.2025 22:21 β π 2 π 0 π¬ 2 π 0
Can you give some examples of these sorts of systems?
15.04.2025 16:13 β π 1 π 0 π¬ 3 π 0
Did you invoke "number of atoms in the universe"? :)
13.03.2025 14:06 β π 2 π 0 π¬ 1 π 0
The textbooks don't need to be rewritten -- they just needed to pay attention to what was already known about generalization, decades ago! I've had thoughts about this for 12 years, and people always ask for the paper -- so I finally wrote it. Thankful to many for feedback! 12/12
05.03.2025 15:37 β π 9 π 0 π¬ 1 π 0
We also further consider overparametrization and double descent. Deep learning of course _is_ different, and much is not well understood. To this end, we particularly highlight representation learning, mode connectivity, and universal learning. Much more in the paper! 11/12
05.03.2025 15:37 β π 8 π 0 π¬ 1 π 0
Computing the bounds is also very simple: (i) train a model to find hypothesis h*, using any optimizer; (ii) measure the empirical risk R(h*) (e.g., training loss); (iii) measure the filesize of the stored model. 10/12
05.03.2025 15:37 β π 5 π 0 π¬ 1 π 0
Official account of the NYU Center for Data Science, the home of the Undergraduate, Masterβs, and Ph.D. programs in data science. cds.nyu.edu
Astrophysicist turned Climate Scientist | AI for Science | Professor of Physics at @cuny.bsky.social; happy camper at @leapstc.bsky.social | Author, ML for Physics and Astronomy: https://shorturl.at/VJ9k9 | Lover of puns, data, good people, and unfinished
Co-founder/CEO Windscape AI, Co-founder/Pres @ml-collective.bsky.socialβ¬.
Prev: Uber AI Labs founding team, adviser @RecursionPharma, Cornell, Montreal, Caltech π»
Bioinformatics Scientist / Next Generation Sequencing, Single Cell and Spatial Biology, Next Generation Proteomics, Liquid Biopsy, SynBio, Compute Acceleration in biotech // http://albertvilella.substack.com
Discover the Languages of Biology
Build computational models to (help) solve biology? Join us! https://www.deboramarkslab.com
DM or mail me!
head, Comp Systems Neurosci Lab @wigner centre. neuro + ML
Science and Shenanigans. Heart and Hope.
https://bunsenbernerbmd.com
Prof at the University of British Columbia. Research in statistics, ML, and AI for science. Views are my own. https://charlesm93.github.io./
Co-founder & Chief Scientist at Yutori. Prev: Senior Director leading FAIR Embodied AI at Meta, and Professor at Georgia Tech.
Chief Scientist & CEO QIM. UVa School of Data Science (Founder/Funder). Embrace Simplicity, Beware Complexity. Coding at CODE. Honesty. Sleep. Squash.
Senior Researcher@Criteo,
Rising Star in AI Ethics 2025 |
Fairness, Causality, GenAI
Co-founder and CEO, Mistral AI
AGI Cyberneticist @daiosai. Ex AI Safety Team lead @deepmind. PhD @CambridgeMLG. AI, Machine Learning, Information Theory, Causality, Decision Theory.
assistant prof at USC Data Sciences and Operations; phd Cornell ORIE. data-driven decision-making, operations research/management, causal inference, algorithmic fairness/equity
angelamzhou.github.io
Professor a NYU; Chief AI Scientist at Meta.
Researcher in AI, Machine Learning, Robotics, etc.
ACM Turing Award Laureate.
http://yann.lecun.com
Security and Privacy of Machine Learning at UofT, Vector Institute, and Google π¨π¦π«π·πͺπΊ Co-Director of Canadian AI Safety Institute (CAISI) Research Program at CIFAR. Opinions mine
Assistant Professor at UW and Staff Research Scientist at Google DeepMind. Social Reinforcement Learning in multi-agent and human-AI interactions. PhD from MIT. Check out https://socialrl.cs.washington.edu/ and https://natashajaques.ai/.
something new | Gemini RL+TTS @ Google DeepMind | Conversational AI @ Meta | RL Agents @ EA | ML+Information Theory @ MIT+Harvard+Duke | Georgia Tech PhD | Ψ²Ω Ψ²ΩΨ―Ϊ―Ϋ Ψ’Ψ²Ψ§Ψ―Ϋ
π{NYC, YYZ}
π https://beirami.github.io/
Machine learning researcher
@Google DeepMind. My opinions do not necessarily represent my employer. Prefer email over DMs.
https://scholar.google.com/citations?hl=e&user=ghbWy-0AAAAJ
https://www.cs.toronto.edu/~gdahl/