Huge thanks to my co-authors and mentors - Douglas Wightman, Christiaan de Leeuw, and @daniposthu.bsky.social โ for their guidance and collaboration, and to the REALMENT consortium for supporting this work.
๐งต 6/6
Huge thanks to my co-authors and mentors - Douglas Wightman, Christiaan de Leeuw, and @daniposthu.bsky.social โ for their guidance and collaboration, and to the REALMENT consortium for supporting this work.
๐งต 6/6
Take home:
Additive PGSs remain the most robust default for most complex traits.
ML/DL can help when traits are:
โข highly heritable
โข low in polygenicity
โข driven by strong dominance deviations
Full paper + code: github.com/nybell/non-a...
๐งต 5/6
In the UK Biobank (10 traits), ML/DL models outperformed additive PGSs for traits known to show dominance - including lipoprotein(a), alkaline phosphatase, and ApoB - but not for height (no dominance).
๐งต 4/6
Across most scenarios, additive PGSs were remarkably robust - even when up to 20% of SNP-hยฒ came from dominance SNPs.
Performance dropped mainly for traits with:
โข high SNP-hยฒ
โข low polygenicity
โข strong dominance deviations
๐งต 3/6
Most PGS methods assume additivity - each allele contributes linearly to risk - but real traits can show dominance deviations.
We simulated phenotypes varying in:
โข SNP heritability (SNP hยฒ)
โข % heritability from dominance
โข polygenicity
โข dominance deviation strength
๐งต 2/6
When do machine learning models actually outperform standard polygenic scores? ๐ค
In our new preprint, we benchmark how non-additive genetic effects (i.e, dominance deviations) shape polygenic prediction across simulated and UK Biobank traits.
๐ www.medrxiv.org/content/10.1...
๐งต 1/6