work with
Diantha Schipaanboord, Floor B.H. van der Zalm, RenΓ© van Es, Melle Vessies, Rutger R. van de Leur, Klaske R. Siegersma, Pim van der Harst, Hester M. den Ruijter, N. Charlotte Onland-Moret, on behalf of the IMPRESS consortium
@vanamsterdam.bsky.social
machine learning, causal inference, healthcare - assistant professor in dep. of Data Science Methods, Julius Center, of University Medical Center Utrecht, the Netherlands; wvanamsterdam.com
work with
Diantha Schipaanboord, Floor B.H. van der Zalm, RenΓ© van Es, Melle Vessies, Rutger R. van de Leur, Klaske R. Siegersma, Pim van der Harst, Hester M. den Ruijter, N. Charlotte Onland-Moret, on behalf of the IMPRESS consortium
Conclusion: The convolutional neural networks in this study demonstrated resilience to simulated sex-imbalance in training ECG data.
pre-print: doi.org/10.1101/2025...
Discrimination remained stable across sexes; only calibration shifted in extreme scenarios when prevalence differed by sex, with similar patterns for women and men.
04.09.2025 12:03 β π 0 π 0 π¬ 1 π 0Using ~165k ECGs, we simulated sex-imbalances in representation (women-to-men ratio), outcome prevalence, and misclassification in the training data for LBBB, long QT syndrome, LVH, and physician-labeled βabnormalβ ECGs.
04.09.2025 12:03 β π 0 π 0 π¬ 1 π 0
Pre-print alert:
Many ECG-AI models have been developed to predict a wide range of cardiovascular outcomes. But, underrepresentation of women in cardiovascular studies raises the question: Are ECG-AI models equally predictive for women and men with sex-imbalanced training data?
New paper in @annalsofim.bsky.social
"50 ways to misinterpret clinical prediction models for treatment decisionsβ
--> Published version: www.acpjournals.org/doi/10.7326/...
--> Open access version: arxiv.org/pdf/2402.17366
BMS-ANed Spring Meeting on Thursday, June 19
Time: 13:00β18:00 (CEST)
Location: Vredenburg 19, 3511 BB, Utrecht
Details and registration: vvsor.nl/biometrics/e...
Still some spots available in our summer school on all things causal inference, 7-11 July in Utrecht! Discounts for those working in universities and non-profits, and affordable accommodation offered by @utrechtuniversity.bsky.social summer school!
28.04.2025 08:00 β π 7 π 6 π¬ 1 π 0Even if you model a physical system, e.g. avg yearly temperature depending on height, and assume that temp given height is the same everywhere. If you invert it into predicting presence of mountain given temp, youβll find varying discrimination in diff countries. Example from scholkopfβs talks
25.04.2025 14:58 β π 0 π 0 π¬ 0 π 0Youβve modeled a system with no meaningful variation across environments. The model may be reliable in the tested environments but you havenβt shown robustness against variation in distributions as you havenβt observed any
25.04.2025 14:56 β π 0 π 0 π¬ 2 π 0
A question that remains is how these differences in environments may come about and what to do with this in practice? On this, I wrote a paper titled, available here: arxiv.org/abs/2409.01444
fin!
if the distribution of outcome given features remains the same (Y|X), calibration is preserved. If both are the same, the environments were not meaningfully different to begin with!
a more lengthy explanation is in this blog post: wvanamsterdam.com/posts/250425...
as promised (so all of you can breathe normally again), here's my TLDR answer:
Environments must differ with respect to something. If the distribution of features given outcome remains the same (X|Y), discrimination is preserved;
tagging some prediction modelers / statisticians, @maartenvsmeden.bsky.social @benvancalster.bsky.social @gelovennan.bsky.social @f2harrell.bsky.social @lucystats.bsky.social @miguelhernan.org @gscollins.bsky.social
(I will answer tomorrow)
Which is stronger evidence for robustness?
When evaluating predictive performance of one model in several different environments (e.g. regions / hospitals):
A. stable discrimination (AUC) and calibration in all environments
B. stable discrimination, varying calibration
vote with π=A; β€οΈ=B
ask chatGPT o3 this before submitting your next paper to, I got ~10 usable comments out of it:
you're a reviewer for <journal>; review the attached paper when you're either:
what are the exceptions?
11.04.2025 06:12 β π 0 π 0 π¬ 1 π 0
2. an external reproduction of the PROTECT method from Manchester University with Charlie Cuniffe, Matt Sperrin and Gareth Price (www.nature.com/articles/s41...)
3. a 'causal' meta-analysis method using only aggregate data, exciting work with Qingyang Shi from Groningen University
Very excited for my first (belated) visit to #EuroCIM2025!
I'm here with 3 bits of work:
1. a poster on a causal understanding of prediction model performance under shifts in 'case-mix' (or covariate / outcome drift); I show how discrimination and calibration respond differently
bit.ly/ccm-arxiv
this seems pretty cool: an overview of llms for statisticians
arxiv.org/abs/2502.17814
Vacancy for a postdoc position.
Improve the transparency of decision support algorithms by figuring out how we can quantify and communicate uncertainty in individual causal predictions.
With Marleen Kunneman, Daniala Weir and me.
Three more days to apply π
www.lumc.nl/en/about-lum...
Building in the physics is one way to potentially get the right causal mechanisms
In sofar as the model is trained on real world patient data, you'll still have to ensure no biases e.g. related to confounding creep in
Digital twins are useful insofar as they reflect causal mechanisms
Don't think a generative model ('digital twin') can inform treatment decisions just because it procudes different outputs when you give it different inputs. Doesn't matter if it's 'AI' or not.
saliency maps are the new table 2 fallacy
17.12.2024 13:21 β π 0 π 0 π¬ 0 π 0
Not sure about overfitting, results seemed robust to 5-site cross validation.
It just learns correlations, what's wrong with that? The words 'confounders' and 'bias' make it sound they expected the model to yield some causal understanding. Maybe these heatmaps are the new table 2 fallacy
Awesome, congrats!
16.12.2024 19:00 β π 1 π 0 π¬ 1 π 0Liking this interaction with @mmbronstein.bsky.social and Denis Danilov so much I'm reposting it here
06.12.2024 16:13 β π 38 π 4 π¬ 3 π 0
Interested in how to use non-experimental data to answer causal research questions? Mystified by DAGs and counterfactuals? Want to learn what Target Trial Emulation is all about?
Sign up now for the 2nd edition of our summer school, 7-11 July in Utrecht, with @vanamsterdam.bsky.social & BPdeVries
Probably more like "the average of an infinite sequence of throws hits the bulls eye"
27.11.2024 14:22 β π 1 π 0 π¬ 1 π 0@oisinryan.bsky.social and I are developing a julia package for target trial emulation with a student, happy to be added to the list
17.11.2024 19:04 β π 3 π 0 π¬ 0 π 0