Gentle reminder that a correlation coefficient isnβt a particularly great way to quantify the effect of a dichotomous treatment. See also
www.the100.ci/2025/07/28/w...
@linstonwin.bsky.social
senior lecturer in statistics, penn NYC & Philadelphia https://www.stat.berkeley.edu/~winston
Gentle reminder that a correlation coefficient isnβt a particularly great way to quantify the effect of a dichotomous treatment. See also
www.the100.ci/2025/07/28/w...
Excellent new editorial and guideline on interpreting p values and interval estimates
bjsm.bmj.com/content/earl...
from Amrhein, Greenland, & McShane ("Retire statistical significance," Nature, 2019) "For example, the authors above could have written: βLike a previous study, our results suggest a 20% increase in risk of new-onset atrial fibrillation in patients given the anti-inflammatory drugs. Nonetheless, a risk difference ranging from a 3% decrease, a small negative association, to a 48% increase, a substantial positive association, is also reasonably compatible with our data, given our assumptions.β "
from Amrhein, Greenland, & McShane ("Retire statistical significance," Nature, 2019) "Whatever the statistics show, it is fine to suggest reasons for your results, but discuss a range of potential explanations, not just favoured ones. Inferences should be scientific, and that goes far beyond the merely statistical. Factors such as background evidence, study design, data quality and understanding of underlying mechanisms are often more important than statistical measures such as P values or intervals."
I like this from @vamrhein.bsky.social et al. I assigned it to my class last semester and tried to explain that p-values measure how compatible (vs. surprising) the data are with the null, given our assumptions. But yeah, tests & CIs are hard to understand!
www.blakemcshane.com/Papers/natur...
Orley Ashenfelter's papers often have good introductions. Here's Ashenfelter & Plant
www.journals.uchicago.edu/doi/abs/10.1...
Koenker & Geling is one of my favorites
www.jstor.org/stable/2670284
Are you or one of your students considering doing a Ph.D. in a social science? I've spent a lot of time talking about this w/ students & finally wrote something up.
IMO, there are only 3 good reasons to do it. One of them needs to be true--otherwise, don't.
medium.com/the-quantast...
This is written for economists, but I think itβs very helpful more generally
www.aeaweb.org/articles?id=...
A nice recent article on why you should abandon hazard ratios.
#statssky #episky
See our No-Spin report on a widely-covered NBER study of Medicaid expansion. In brief: Despite the abstract's claims that expansion reduced adult mortality 2.5%, the study found much smaller effects that fell short of statistical significance in its main preregistered analysis.π§΅
11.06.2025 17:38 β π 1 π 2 π¬ 1 π 0Starting to look like I might not be able to work at Harvard anymore due to recent funding cuts. If you know of any open statistical consulting positions that support remote work or are NYC-based, please reach out! π
04.06.2025 19:02 β π 153 π 96 π¬ 11 π 7In case this is of interest, even ANCOVA I is consistent and asymptotically normal in completely randomized experiments (though II is asymptotically more efficient in imbalanced or multiarm designs)
05.05.2025 19:31 β π 3 π 0 π¬ 1 π 0Issues with interpreting p-values haunts even AI, which is prone to same biases as human researchers. ChatGPT, Gemini & Claude all fall prey to "dichotomania" - treating p=0.049 & p=0.051 as categorically different, and paying too much attention to significance. www.cambridge.org/core/journal...
21.04.2025 16:51 β π 41 π 6 π¬ 2 π 0NEW: CONSORT 2025 now published!
Some notable changes:
-items on analysis populations, missing data methods, and sensitivity analyses
-reporting of non-adherence and concomitant care
-reporting of changes to any study methods, not just outcomes
-and lots of other things
www.bmj.com/content/389/...
How to write a response to reviewers. www.sciencedirect.com/science/arti...
14.04.2025 12:22 β π 9 π 6 π¬ 1 π 0Very nice explainer by @economictricks.bsky.social
www.econometrics.blog/post/why-eco...
today we will all read imbens 2021 on statistical significance and p values, which is a strong contender for having the best opening paragraph of any stats paper
pubs.aeaweb.org/doi/pdf/10.1...
Email to Stata explaining that Welch (1949) "welched" on Welch (1947), but refraining from asking if their co-founder Finis Welch was related to B. L. Welch. :) Continued in next screenshot
Second half of email, suggesting that the "unequal" option be described as using "the formula of Satterthwaite (1946) and Welch (1949)" (because what many people call the Welch test is what Stata's "unequal" option does, not what their "welch" option does)
Btw here's an email I sent Stata in 2012, suggesting a clarification to their descriptions of the "unequal" and "welch" ttest options. Got a polite reply but I don't think they changed it :)
06.04.2025 16:59 β π 2 π 0 π¬ 0 π 0Imbens & Kolesar give a nice overview and simulations
www.jstor.org/stable/24917...
Here's some older, related stuff (from me) aimed at political scientists.
Related paper #1
"Arguing for a Negligible Effect"
Journal: onlinelibrary.wiley....
PDF: www.carlislerainey.c...
"The Need for Equivalence Testing in Economics"
from Jack Fitzgerald (@jackfitzgerald.bsky.social)
Preprint: osf.io/preprints/met...
We know that "not significant" does not imply evidence for "no effect," but I still see papers make this leap.
Good to see more work making this point forcefully!
Isnβt the Hausman approach likely to lead to undercoverage, similar to what @jondr44.bsky.social wrote about in a different context?
www.aeaweb.org/articles?id=...
For RCTs, another reference with simulation evidence on the robustness of OLS is Judkins & Porter (2016). But average marginal effects from logit are also robust
doi.org/10.1002/sim....
In the '90s when I worked at Abt and MDRC, I wrote an email that initially had the subject header "OLS without apology". I shared a later version with Freedman, who cited it as "Lim (1999)" in his "Randomization does not justify logistic regression"
08.03.2025 20:39 β π 10 π 1 π¬ 0 π 0Clarification: my paper doesnβt advocate a specific estimator. Thatβs one of the meanings of βagnosticβ in the title :)
08.03.2025 17:16 β π 2 π 0 π¬ 0 π 0A warning and a plea - As fields start to use more advanced quantitative / "cause" methods, there is a desire to help consumers of the research (four journal reviewers, editor) to easily assess the study quality and validity (e.g. JAMA causality language) ... - We need to push against this - need ways to help people understand and assess the (inherently untestable) assumptions in many studies.
An important plea from @lizstuart.bsky.social in today's SCI-OCIS Special Webinar Series:
19.02.2025 17:43 β π 9 π 2 π¬ 2 π 0I don't wanna put words in Rosenbaum's mouth ("spectrum" is just a word that came to my mind for a quick Bluesky reply) and I'd really encourage anyone interested to read his papers and the Stat Sci discussion in full, and then critique them. :) But here's a screenshot from his reply to Manski
10.02.2025 17:25 β π 2 π 0 π¬ 0 π 0competing theories. He has an interesting debate with Manski on external validity in the comments on the Stat Sci paper (I'll send you some excellent responses that my undergrad students at Yale wrote).
10.02.2025 15:01 β π 1 π 0 π¬ 1 π 0can lead to badly misleading literatures; (3) we can sometimes learn from collections of studies with different designs & weaknesses (I think he's partly influenced by the literature on smoking & lung cancer, which he cites elsewhere); (4) we should try to falsify or corroborate predictions of 2/
10.02.2025 15:01 β π 2 π 0 π¬ 1 π 0Thanks, Alex! I think that's a small part of his message. It's hard for me to do justice to these papers in a short thread, but I think he's also saying (1) credibility is on a spectrum and we should try to learn from all sorts of designs; (2) repeating the same design with the same weaknesses 1/
10.02.2025 15:01 β π 2 π 0 π¬ 1 π 1