Multicolinearity— they all look the same
Heteroscedasticity— the variation varies
Attenuation— being too modest
Overfitting— too good to be true
Confounding— nothing is what it seems
P-value— it’s complicated
Sensitivity analysis— tried a bunch of stuff
Post-hoc— main analysis not sexy enough
Multivariate— oops, meant to say multivariable
Normality— a very rare shape for data
Dichotomized— data was tortured
Extrapolation— just guessing
Linear regression— line through data points
t-test— linear regression
correlation— linear regression
ANOVA— linear regression
ANCOVA— linear regression
Chi-square test— logistic regression
Deep learning— bunch of regressions
Advanced stuff:
Non-convergence— computer says no
Heywood case— science fiction
Bootstrap standard errors— could not do the math
Robust standard errors— pretending to be cautious
Shrinkage— regularization
Regularization— couldn’t get large enough dataset
Validated— we did a test (it might have failed)
Interaction— the variation varies
Heterogeneity— the variation varies
Risk factor— observed a correlation
Meta-analysis— calculated a weighted average
Collider— mass murderer of interpretable statistical analyses
Power calculation— effect size that matched budget
Exploratory— playing around
Replicated— did it again
Missing data— holes in the dataset
Measurement error— observe A, make conclusions about B
Stepwise regression— no idea what I am doing
• • •
Missing some Tweet in this thread? You can try to
force a refresh
And if you really want to know about the effects of measurement error on results (conditional or marginal effect), this fabulous tool by @lindanab1 will help you lindanab.shinyapps.io/SensitivityAna…
The main problem with badly designed medical prediction models is not research waste or hampering scientific progress. It is the risk that someone takes the model seriously and use it to inform medical decision which can ruin someone’s life
There is no “hypothesis generating” or “too small, but maybe useful for a meta-analysis”. It’s building a tool, probably more that it is a scientific endeavor
Perhaps many of the published prediction models are not developed with the actual intention to eventually be used in (clinical) practice. We can and should do a lot better in flagging not-for-actual use clinical prediction models. And perhaps not publishing the fancy R-shiny app?
This is my *top 10* favorite methods papers of 2021
Disclaimer: this top 10 is just personal opinion. I’m biased towards explanatory methods and statistics articles relevant to health research, particularly those relating to prediction models.
Shameless plugs alert. Two papers I co-authored (but did not lead) made the top 10
#1: (non-)collapsibility is one of these unintuitive phenomena that can confuse you for the rest of your career. This paper does an excellent job explaining