Lester Mackey's Avatar

Lester Mackey

@lestermackey.bsky.social

Machine learning researcher at Microsoft Research. Adjunct professor at Stanford.

153 Followers  |  17 Following  |  13 Posts  |  Joined: 20.11.2024  |  2.2249

Latest posts by lestermackey.bsky.social on Bluesky

Preview
Microsoft Research Lab - New York City - Microsoft Research Apply for a research position at Microsoft Research New York & collaborate with academia to advance economics research, prediction markets & ML.

Microsoft Research New York City (www.microsoft.com/en-us/resear...) is seeking applicants for multiple Postdoctoral Researcher positions in ML/AI!

These are positions for up to 2 years, starting in July 2026.

Application deadline: October 22, 2025

12.09.2025 14:57 โ€” ๐Ÿ‘ 4    ๐Ÿ” 1    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0
Preview
Microsoft Research Lab - New York City - Microsoft Research Apply for a research position at Microsoft Research New York & collaborate with academia to advance economics research, prediction markets & ML.

MSR NYC is hiring spring and summer interns in AI/ML/RL!

Apply here: jobs.careers.microsoft.com/global/en/jo...

02.10.2025 20:57 โ€” ๐Ÿ‘ 20    ๐Ÿ” 7    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0
Video thumbnail

Only in LA ๐Ÿข

11.10.2025 03:35 โ€” ๐Ÿ‘ 1    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0
Post image

Very excited to share our preprint: Self-Speculative Masked Diffusions

We speed up sampling of masked diffusion models by ~2x by using speculative sampling and a hybrid non-causal / causal transformer

arxiv.org/abs/2510.03929

w/ @vdebortoli.bsky.social, Jiaxin Shi, @arnauddoucet.bsky.social

07.10.2025 22:09 โ€” ๐Ÿ‘ 13    ๐Ÿ” 6    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0
Video thumbnail

Could this be our future? (Sound on)

01.10.2025 13:21 โ€” ๐Ÿ‘ 1    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0
Preview
LeaVS: Accelerating learning for biological AI โ€” JURA Bio, Inc. A fundamental lesson of modern AI is that scale is essential: training bigger models on bigger datasets unlocks new capabilities. A fundamental lesson of AI engineering is that scaling up isn't trivia...

You can read more in our post at www.jurabio.com/blog/leavs; preprint forthcoming.

@jura.bsky.social @eliweinstein.bsky.social @mgollub.bsky.social @highvariance.bsky.social

22.09.2025 12:15 โ€” ๐Ÿ‘ 4    ๐Ÿ” 5    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 1
Post image

The Microsoft Research Undergraduate Internship Program offers 12-week internships in our Redmond, NYC, or New England labs for rising juniors and seniors who are passionate about technology. Apply by October 6: msft.it/6015scgSJ

12.09.2025 15:00 โ€” ๐Ÿ‘ 6    ๐Ÿ” 2    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 3

If you're an undergraduate interested in interning with me or one of my amazing colleagues at Microsoft Research New England this summer, please apply here: msft.it/6015scgSJ

17.09.2025 03:00 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0
Post image

Tomorrow we're excited to host @sarahalamdari.bsky.social at Chalmers for the AI4Science seminar and hear about generative models for protein design! Talk at 3pm CEST. ๐Ÿคฉ

For more info, including details on how to join virtually, please see psolsson.github.io/AI4ScienceSe...

@smnlssn.bsky.social

10.09.2025 08:24 โ€” ๐Ÿ‘ 8    ๐Ÿ” 3    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0
Search Jobs | Microsoft Careers

We may have the chance to hire an outstanding researcher 3+ years post PhD to join Tarleton Gillespie, Mary Gray and me in Cambridge MA bringing critical sociotechnical perspectives to bear on new technologies.

jobs.careers.microsoft.com/global/en/jo...

28.07.2025 17:26 โ€” ๐Ÿ‘ 89    ๐Ÿ” 49    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 3
Post image Post image

In 1965, Margaret Dayhoff published the Atlas of Protein Sequence and Structure, which collated the 65 proteins whose amino acid sequences were then known.

Inspired by that Atlas, today we are releasing the Dayhoff Atlas of protein sequence data and protein language models.

25.07.2025 22:05 โ€” ๐Ÿ‘ 66    ๐Ÿ” 28    ๐Ÿ’ฌ 3    ๐Ÿ“Œ 3

So you want to skip our thinning proofsโ€”but youโ€™d still like our out-of-the-box attention speedups? Iโ€™ll be presenting the Thinformer at two ICML workshop posters tomorrow!

Catch me at Es-FoMo (1-2:30, East hall A) and at LCFM (10:45-11:30 & 3:30-4:30, West 202-204)

19.07.2025 07:04 โ€” ๐Ÿ‘ 5    ๐Ÿ” 4    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

Jikai Jin, Lester Mackey, Vasilis Syrgkanis: It's Hard to Be Normal: The Impact of Noise on Structure-agnostic Estimation https://arxiv.org/abs/2507.02275 https://arxiv.org/pdf/2507.02275 https://arxiv.org/html/2507.02275

04.07.2025 06:53 โ€” ๐Ÿ‘ 2    ๐Ÿ” 5    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0
Post image 15.07.2025 03:30 โ€” ๐Ÿ‘ 4    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0
Preview
Low-Rank Thinning The goal in thinning is to summarize a dataset using a small set of representative points. Remarkably, sub-Gaussian thinning algorithms like Kernel Halving and Compress can match the quality of unifor...

Off to ICML next week?

Check out my student Annabelleโ€™s paper in collaboration with @lestermackey.bsky.social and colleagues on low-rank thinning!

New theory, dataset compression, efficient attention and more:

arxiv.org/abs/2502.12063

12.07.2025 16:27 โ€” ๐Ÿ‘ 11    ๐Ÿ” 5    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 1

Your data is low-rank, so stop wasting compute! In our new paper on low-rank thinning, we share one weird trick to speed up Transformer inference, SGD training, and hypothesis testing at scale. Come by ICML poster W-1012 Tuesday at 4:30!

14.07.2025 18:29 โ€” ๐Ÿ‘ 25    ๐Ÿ” 7    ๐Ÿ’ฌ 2    ๐Ÿ“Œ 2

Jikai Jin, Lester Mackey, Vasilis Syrgkanis
It's Hard to Be Normal: The Impact of Noise on Structure-agnostic Estimation
https://arxiv.org/abs/2507.02275

04.07.2025 04:40 โ€” ๐Ÿ‘ 2    ๐Ÿ” 1    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0
2025 Call For Ethics Reviewers If you are able and willing to participate in the review process, please sign up at this form. Feel free to share this call with your colleagues.

NeurIPS is seeking additional ethics reviewers this year. If you are able and willing to participate in the review process, please sign up at the form in the link:
neurips.cc/Conferences/...
Please share this call with your colleagues!

02.07.2025 16:06 โ€” ๐Ÿ‘ 11    ๐Ÿ” 8    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0
Preview
็”˜ๅˆฉ ไฟŠไธ€ ๆ „่ช‰็ ”็ฉถๅ“กใŒใ€Œไบฌ้ƒฝ่ณžใ€ใ‚’ๅ—่ณž ็”˜ๅˆฉ ไฟŠไธ€ๆ „่ช‰็ ”็ฉถๅ“ก๏ผˆๆœฌๅ‹™๏ผšๅธไบฌๅคงๅญฆ ๅ…ˆ็ซฏ็ทๅˆ็ ”็ฉถๆฉŸๆง‹ ็‰นไปปๆ•™ๆŽˆ๏ผ‰ใฏใ€ไบบๅทฅใƒ‹ใƒฅใƒผใƒฉใƒซใƒใƒƒใƒˆใƒฏใƒผใ‚ฏใ€ๆฉŸๆขฐๅญฆ็ฟ’ใ€ๆƒ…ๅ ฑๅนพไฝ•ๅญฆๅˆ†้‡Žใงใฎๅ…ˆ้ง†็š„ใช็ ”็ฉถใŒ่ฉ•ไพกใ•ใ‚Œใ€็ฌฌ40ๅ›ž๏ผˆ2025๏ผ‰ไบฌ้ƒฝ่ณž๏ผˆๅ…ˆ็ซฏๆŠ€่ก“้ƒจ้–€ใ€€ๅ—่ณžๅฏพ่ฑกๅˆ†้‡Ž๏ผšๆƒ…ๅ ฑ็ง‘ๅญฆ๏ผ‰ใ‚’ๅ—่ณžใ—ใพใ—ใŸใ€‚

Shunichi Amari has been awarded the 40th (2025) Kyoto Prize in recognition of his pioneering research in the fields of artificial neural networks, machine learning, and information geometry

www.riken.jp/pr/news/2025...

20.06.2025 13:26 โ€” ๐Ÿ‘ 35    ๐Ÿ” 12    ๐Ÿ’ฌ 2    ๐Ÿ“Œ 0

If youโ€™d like to expand your analysis to support equal weighted kernel thinning coresets, have a look at Low-Rank Thinning (arxiv.org/pdf/2502.12063); that bounds kernel thinning MMD directly in terms of eigenvalue decay

01.05.2025 18:07 โ€” ๐Ÿ‘ 1    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

Very cool!

01.05.2025 17:56 โ€” ๐Ÿ‘ 1    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0
Post image

๐Ÿ† I'm delighted to share that I've won a 2025 COPSS Emerging Leader Award! ๐Ÿ˜ƒ And congratulations to my fellow winners! ๐Ÿ™Œ๐Ÿฝ

Check out how each of us is improving and advancing the profession of #statistics and #datascience here: tinyurl.com/copss-emerging-leader-award

11.03.2025 06:47 โ€” ๐Ÿ‘ 13    ๐Ÿ” 4    ๐Ÿ’ฌ 3    ๐Ÿ“Œ 2

Congratulations to the 2025 #COPSS Awardees, @ericjdaza.com, @lucystats.bsky.social, @lestermackey.bsky.social, and all of you. I hope to congratulate you at #JSM2025 in Nashville with @amstatnews.bsky.social. God I hope to go. #rstats #statssky

14.03.2025 21:56 โ€” ๐Ÿ‘ 12    ๐Ÿ” 3    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0
Preview
Self-nomination for reviewing at NeurIPS 2025 โ€“ NeurIPS Blog Communications Chairs 2025 2025 Conference

NeurIPS 2025 is soliciting self-nominations for reviewers and ACs. Please read our blog post for details on eligibility criteria, and process to self-nominate:

10.03.2025 14:07 โ€” ๐Ÿ‘ 16    ๐Ÿ” 9    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

Thanks Arthur!

08.03.2025 13:21 โ€” ๐Ÿ‘ 1    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

Thanks Michele!

08.03.2025 13:20 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0
Post image

Congratulations to @lestermackey.bsky.social for receiving the 2025 COPSS Award! ๐ŸŽ‰๐Ÿ‘

Lester is currently the Chair of the Section on Bayesian Statistical Sciences (SBSS) of the American Statistical Association.

08.03.2025 05:14 โ€” ๐Ÿ‘ 18    ๐Ÿ” 4    ๐Ÿ’ฌ 2    ๐Ÿ“Œ 3
Post image

Off to #AAAI25! We're presenting #SatCLIP (w/ @marccoru.bsky.social, @estherrolf.bsky.social, @calebrob6.bsky.social & @lestermackey.bsky.social) at the 12.30-2.30pm poster session on Feb 28! Let me know if you're around & want to chat #GeoAI!๐Ÿ›ฐ๏ธ

Paper: tinyurl.com/5eejz5kw
Code: tinyurl.com/2zm64967

26.02.2025 14:28 โ€” ๐Ÿ‘ 8    ๐Ÿ” 5    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0
Quality of T2T-ViT attention approximations on ImageNet. Their method is faster than everything but Performer, while losing less accuracy than any other approximation (including Performer, which is too lossy to really show up in prod).

Quality of T2T-ViT attention approximations on ImageNet. Their method is faster than everything but Performer, while losing less accuracy than any other approximation (including Performer, which is too lossy to really show up in prod).

Time-power trade-off curves for detecting Higgs bosons with deep kernel MMD tests. Their method is way past pareto optimal.

Time-power trade-off curves for detecting Higgs bosons with deep kernel MMD tests. Their method is way past pareto optimal.

idk dude it's 30 pages of proofs and then they drop this

20.02.2025 15:53 โ€” ๐Ÿ‘ 8    ๐Ÿ” 1    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

I've been shocked that a theory-driven method yields practical results this good, especially on attention approximation. I proposed my best new optimizer design originally as a dumb baseline; the fact that you can get these efficiency gains with a principled approach makes me a lil insecure.

20.02.2025 15:43 โ€” ๐Ÿ‘ 19    ๐Ÿ” 1    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

@lestermackey is following 17 prominent accounts