Max Ilse's Avatar

Max Ilse

@maxilse.bsky.social

Working at microsoft research health futures. Interested in causal representation learning and generative modelling applied to medical data.

858 Followers  |  798 Following  |  18 Posts  |  Joined: 19.11.2024  |  2.4814

Latest posts by maxilse.bsky.social on Bluesky

The result is a fair, end‑to‑end comparison that isolates what actually drives performance for radiology foundation models.

#AI #MedicalImaging #FoundationModels #ScalingLaws #Radiology

23.09.2025 08:34 β€” πŸ‘ 2    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

including not just findings but also lines & tubes classification/segmentation and report generation. We also test the effect of adding structured labels alongside reports during CLIP‑style pretraining, and study scaling laws under these controlled conditions.

23.09.2025 08:34 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

That makes it hard to tell whether wins come from the model design or just from more data/compute or favorable benchmarks. We fix this by holding the pretraining dataset and compute constant and standardizing evaluation across tasks,

23.09.2025 08:34 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

Why this matters: Prior comparisons of radiology encoders have often been apples‑to‑oranges: models trained on different datasets, with different compute budgets, and evaluated mostly on small datasets of finding‑only tasks.

23.09.2025 08:34 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

βœ… Pretrained on 3.5M CXRs to study scaling laws for radiology models
βœ… Compared MedImageInsight (CLIP-based) vs RAD-DINO (DINOv2-based)
βœ… Found that structured labels + text can significantly boost performance
βœ… Showed that as little as 30k in-domain samples can outperform public foundation models

23.09.2025 08:34 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Preview
Data Scaling Laws for Radiology Foundation Models Foundation vision encoders such as CLIP and DINOv2, trained on web-scale data, exhibit strong transfer performance across tasks and datasets. However, medical imaging foundation models remain constrai...

🩻Excited to share our latest preprint: β€œData Scaling Laws for Radiology Foundation Models”
Foundation vision encoders like CLIP and DINOv2 have transformed general computer vision, but what happens when we scale them for medical imaging?

πŸ“„ Read the full preprint here: arxiv.org/abs/2509.12818

23.09.2025 08:34 β€” πŸ‘ 5    πŸ” 2    πŸ’¬ 1    πŸ“Œ 0

What a damning abstract

30.04.2025 08:29 β€” πŸ‘ 5    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Post image

I want to reshare @brandfonbrener.bsky.social's @NeurIPSConf 2024 paper on CoLoR-Filter: A simple yet powerful method for selecting high-quality data for language model pre-training!

With @hlzhang109.bsky.social @schwarzjn.bsky.social @shamkakade.bsky.social

05.04.2025 12:04 β€” πŸ‘ 18    πŸ” 8    πŸ’¬ 2    πŸ“Œ 1
Screenshot of 'SHADES: Towards a Multilingual Assessment of Stereotypes in Large Language Models.'
SHADES is in multiple grey colors (shades).

Screenshot of 'SHADES: Towards a Multilingual Assessment of Stereotypes in Large Language Models.' SHADES is in multiple grey colors (shades).

⚫βšͺ It's coming...SHADES. βšͺ⚫
The first ever resource of multilingual, multicultural, and multigeographical stereotypes, built to support nuanced LLM evaluation and bias mitigation. We have been working on this around the world for almost **4 years** and I am thrilled to share it with you all soon.

10.02.2025 08:28 β€” πŸ‘ 128    πŸ” 23    πŸ’¬ 6    πŸ“Œ 3
Preview
Econometrics and Data Science

We’re looking for a motivated researcher to apply for a Marie SkΕ‚odowska-Curie postdoc with our Econometrics & Data Science group at SDU!

Focus: Causal Inference, Machine Learning, Big Data
Full support for promising projects

More info & apply:
www.sdu.dk/en/om-sdu/in...

30.01.2025 07:43 β€” πŸ‘ 18    πŸ” 12    πŸ’¬ 0    πŸ“Œ 1
Preview
a cartoon pikachu says join us in white letters ALT: a cartoon pikachu says join us in white letters

Apply!
Assistant Professor (L/SL) in AI, including computer vision [DL 5 Mar] @BristolUni - awarded AI University of the Year in 2024.
DM to myself or @_SethBullock_ for inquiries (don't us send CV pls!, apply directly)
www.bristol.ac.uk/jobs/find/de...

28.01.2025 18:53 β€” πŸ‘ 20    πŸ” 9    πŸ’¬ 0    πŸ“Œ 0
Post image

πŸŽ“ πŸ’« We are opening post-doc positions at the intersection of AI, data science, and medicine:
β€’ Large Language Models for French medical texts
β€’ Evaluating digital medical devices: statistics and causal inference

29.01.2025 08:19 β€” πŸ‘ 27    πŸ” 16    πŸ’¬ 1    πŸ“Œ 0
Preview
When A.I. Passes This Test, Look Out The creators of a new test called β€œHumanity’s Last Exam” argue we may soon lose the ability to create tests hard enough for A.I. models.

If you name your AI benchmark "Humanity's Last Exam" and get Kevin Roose to gush about it, you work in advertising, not in computer science.

www.nytimes.com/2025/01/23/t...

26.01.2025 00:09 β€” πŸ‘ 283    πŸ” 28    πŸ’¬ 10    πŸ“Œ 2

Happy to announce a collaboration with the Mayo Clinic to advance our research in radiology report generation!
newsnetwork.mayoclinic.org/discussion/m...
Tagging some of the core team: @valesalvatelli.bsky.social @fepegar.com @maxilse.bsky.social @sambondtaylor.bsky.social @anton-sc.bsky.social

15.01.2025 10:56 β€” πŸ‘ 7    πŸ” 2    πŸ’¬ 1    πŸ“Œ 0

The video of our talk "From Augustus to #Trump – Why #Disinformation Remains a Problem and What We Can Do About It Anyway" at #38c3, Europe's largest Hacker conference, was published, including the German original and English and Spanish translations:

media.ccc.de/v/38c3-von-a...

10.01.2025 08:49 β€” πŸ‘ 8    πŸ” 2    πŸ’¬ 1    πŸ“Œ 0

Internship in our group at Mila in reinforcement learning + graphs for reducing energy use in buildings.

More info and submit an application by Jan 13 here:
forms.gle/TCChXnvSAHqz...

Questions? Email donna.vakalis@mila.quebec with [intern!] in the subject line.

10.01.2025 01:14 β€” πŸ‘ 13    πŸ” 4    πŸ’¬ 0    πŸ“Œ 0
Screenshot of Table of Contents (Part 1)

Contents
1 Introduction 217
2 Positionality 221
3 Overview of Risks and Harms Associated with Computer
Vision Systems and Proposed Mitigation Strategies 223
3.1 Representational Harms . . . . . . . . . . . . . . . . . . . 223
3.2 Quality-of-Service and Allocative Harms . . . . . . . . . . 229
3.3 Interpersonal Harms . . . . . . . . . . . . . . . . . . . . . 237
3.4 Societal Harms: System Destabilization and Exacerbating
Inequalities . . . . . . . . . . . . . . . . . . . . . . . . . . 245
4 Frameworks and Principles for Computer Vision
Researchers 266
4.1 Guidelines for Responsible Data and Model Development . 267
4.2 Measurement Modeling . . . . . . . . . . . . . . . . . . . 271
4.3 Reflexivity . . . . . . . . . . . . . . . . . . . . . . . . . . 273
5 Reorientations of Computer Vision Research 276
5.1 Grounded in Historical Context and Considering
Power Dynamics . . . . . . . . . . . . . . . . . . . . . . . 276
5.2 Small, Task Specific . . . . . . . . . . . . . . . . . . . . . 279
5.3 Community-Rooted . . . . . . . . . . . . . . . . . . . . . 280

Screenshot of Table of Contents (Part 1) Contents 1 Introduction 217 2 Positionality 221 3 Overview of Risks and Harms Associated with Computer Vision Systems and Proposed Mitigation Strategies 223 3.1 Representational Harms . . . . . . . . . . . . . . . . . . . 223 3.2 Quality-of-Service and Allocative Harms . . . . . . . . . . 229 3.3 Interpersonal Harms . . . . . . . . . . . . . . . . . . . . . 237 3.4 Societal Harms: System Destabilization and Exacerbating Inequalities . . . . . . . . . . . . . . . . . . . . . . . . . . 245 4 Frameworks and Principles for Computer Vision Researchers 266 4.1 Guidelines for Responsible Data and Model Development . 267 4.2 Measurement Modeling . . . . . . . . . . . . . . . . . . . 271 4.3 Reflexivity . . . . . . . . . . . . . . . . . . . . . . . . . . 273 5 Reorientations of Computer Vision Research 276 5.1 Grounded in Historical Context and Considering Power Dynamics . . . . . . . . . . . . . . . . . . . . . . . 276 5.2 Small, Task Specific . . . . . . . . . . . . . . . . . . . . . 279 5.3 Community-Rooted . . . . . . . . . . . . . . . . . . . . . 280

Screenshot of Table of Contents (Part 2)

6 Systemic Change 285
6.1 Collective Action and Whistleblowing . . . . . . . . . . . . 285
6.2 Refusal/The Right not to Build Something . . . . . . . . . 287
6.3 Independent Funding Outside of Military and Multinational
Corporations . . . . . . . . . . . . . . . . . . . . . . . . . 289
7 Conclusion 291
References 293

Screenshot of Table of Contents (Part 2) 6 Systemic Change 285 6.1 Collective Action and Whistleblowing . . . . . . . . . . . . 285 6.2 Refusal/The Right not to Build Something . . . . . . . . . 287 6.3 Independent Funding Outside of Military and Multinational Corporations . . . . . . . . . . . . . . . . . . . . . . . . . 289 7 Conclusion 291 References 293

Dear computer vision researchers, students & practitionersπŸ”‡πŸ”‡πŸ”‡

Remi Denton & I have written what I consider to be a comprehensive paper on the harms of computer vision systems reported to date & how people have proposed addressing them, from different angles.

PDF: cdn.sanity.io/files/wc2kmx...

16.12.2024 16:52 β€” πŸ‘ 388    πŸ” 166    πŸ’¬ 9    πŸ“Œ 10
Posts, barbed wire fences and the main gate of the former Auschwitz II-Birkenau camp.

Posts, barbed wire fences and the main gate of the former Auschwitz II-Birkenau camp.

Help us commemorate victims, preserve memory & educate the world. Amplify our voice.

Your interaction here is more than just a click. It is an act of remembrance against forgetting. Like, share, or quote our posts.

Let people know that @auschwitzmemorial.bsky.social is present here.

29.11.2024 16:41 β€” πŸ‘ 1717    πŸ” 762    πŸ’¬ 49    πŸ“Œ 62
Video thumbnail

New timeline, same problems, same solution

13.12.2024 14:22 β€” πŸ‘ 251    πŸ” 49    πŸ’¬ 10    πŸ“Œ 3
Post image

Kudos to @blackhc.bsky.social for calling out this neurips oral: openreview.net/forum?id=0NM... for not giving the rho-loss paper (arxiv.org/abs/2206.07137) the recognition it deserves!

12.12.2024 10:14 β€” πŸ‘ 8    πŸ” 0    πŸ’¬ 1    πŸ“Œ 1

In all of the reposts I see of articles criticising generative AI, I still don't see enough mention of work like Dr. Birhane's, which shows the biases against disadvantaged groups in the training datasets.
This is very good.

10.12.2024 19:20 β€” πŸ‘ 50    πŸ” 18    πŸ’¬ 1    πŸ“Œ 0
Preview
Probabilistic weather forecasting with machine learning - Nature GenCast, a probabilistic weather model using artificial intelligence for weather forecasting, has greater skill and speed than the top operational medium-range weather forecast in the world and provid...

The next generation of probabilistic machine learning for weather called GenCast is published in @natureportfolio.bsky.social today πŸ₯³. Amazing to see the collective progress in ML for weather as a field over the last 5 years. πŸ–οΈ www.nature.com/articles/s41...

04.12.2024 19:30 β€” πŸ‘ 157    πŸ” 29    πŸ’¬ 3    πŸ“Œ 6
Post image

One of the reasons the university sector has come so spectacularly off the rails is the fact it's so unfriendly to family life, people with caring responsibilities and parents. The attitude is often: 'Not working 24/7? You're not fully committed!'
www.science.org/content/arti...

29.11.2024 09:10 β€” πŸ‘ 95    πŸ” 12    πŸ’¬ 8    πŸ“Œ 2

This is just sad

27.11.2024 19:38 β€” πŸ‘ 3    πŸ” 1    πŸ’¬ 0    πŸ“Œ 0
Preview
Medical Adaptation of Large Language and Vision-Language Models: Are We Making Progress? Several recent works seek to develop foundation models specifically for medical applications, adapting general-purpose large language models (LLMs) and vision-language models (VLMs) via continued pret...

Medically adapted foundation models (think Med-*) turn out to be more hot air than hot stuff. Correcting for fatal flaws in evaluation, the current crop are no better on balance than generic foundation models, even on the very tasks for which benefits are claimed.
arxiv.org/abs/2411.04118

26.11.2024 18:12 β€” πŸ‘ 259    πŸ” 57    πŸ’¬ 8    πŸ“Œ 13

What a surprise (not!). Yet again ... poor evaluations of specialized medical LLMs result in overhyped claims relative to the base LLMs. #bioMLeval

27.11.2024 02:16 β€” πŸ‘ 76    πŸ” 14    πŸ’¬ 1    πŸ“Œ 1

Not sure about IRM thought. They early stopped their experiments to get the colored MNIST results.

22.11.2024 22:54 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

Good repost would have completely missed this :)

22.11.2024 14:12 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

Thanks for the pack! Can you please add me :)

21.11.2024 13:22 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Screenshot of the paper.

Screenshot of the paper.

Even as an interpretable ML researcher, I wasn't sure what to make of Mechanistic Interpretability, which seemed to come out of nowhere not too long ago.

But then I found the paper "Mechanistic?" by
@nsaphra.bsky.social and @sarah-nlp.bsky.social, which clarified things.

20.11.2024 08:00 β€” πŸ‘ 232    πŸ” 28    πŸ’¬ 7    πŸ“Œ 2

@maxilse is following 20 prominent accounts