SGEM Xtra: The 12 Days of Christmas the SGEM Gave to Me

SGEM Xtra: The 12 Days of Christmas the SGEM Gave to Me

Released Saturday, 21st December 2024
Good episode? Give it some love!
SGEM Xtra: The 12 Days of Christmas the SGEM Gave to Me

SGEM Xtra: The 12 Days of Christmas the SGEM Gave to Me

SGEM Xtra: The 12 Days of Christmas the SGEM Gave to Me

SGEM Xtra: The 12 Days of Christmas the SGEM Gave to Me

Saturday, 21st December 2024
Good episode? Give it some love!
Rate Episode

Date: December 15, 2024

Guest Skeptics: Dr. Chris Carpenter, Vice Chair of Emergency Medicine at Mayo Clinic.Today, we’re sleighing through the holiday season with a special episode filled with statistical cheer, a dash of skepticism, and a hint of eggnog-flavoured nerdiness.This is an SGEM Xtra like the one we did on What I Learned from Top Gun. It’s fun to mix it up and not do a structured critical appraisal of a recent publication and have a more philosophical chat. 

The inspiration for the SGEM Xtra episode is the BMJ 2022 holiday article called "The 12 Days of Christmas, the Statistician Gave to Me,". The BMJ statistical editors gifted us a publication highlighting common statistical faux pas, laid out in true holiday spirit. We came up with our own SGEM 12 nerdy days.

The 12 Days of Christmas the SGEM Gave to Me

Day 1: A P Value in a Pear Tree

Ah, the P value of frequentists statistics —often misunderstood and frequently abused. The key here is remembering that a P value isn’t proof of the “truth”. We define truth as the best point estimate of an observed effect size with a confidence interval around the point estimate. We should not hang our clinical decisions on a single P value. When do we ever base our care on one data point?

Day 2: Two Confidence Intervals

Confidence intervals (CIs) tell us the range of plausible values for our estimate. If they’re too wide, it’s like a holiday sweater that’s too loose—unflattering and not very useful. Reconsider that sample size or effect size to get clinically impactful intervals you would want to share with Santa Claus. And remember, CIs don’t mean certainty!

Day 3: Three Missing Values

Missing data is the Grinch of research. Ignoring it or using improper methods to handle it can bias your results in unpredictable directions. Using methods like multiple imputation or sensitivity analysis can salvage your data without sacrificing rigour. We often forget that missing data extends beyond interventional studies. For example, in diagnostic research like emergency ultrasound indeterminate results are often swept under the rug and either excluded or not reported – even when methods like 3x2 tables exist to improve the transparency of these indeterminate results that reflect real-life when trying to deploy these diagnostics that require technical expertise and subjective interpretation.

Day 4: Four Overfit Models

Overfitting is like over-decorating your Christmas tree—too many ornaments and it collapses. Overfit models describe the noise, not the signal, making them poor predictors in new datasets. Keeping your models simple and robust can pay downstream dividends for widespread usability and replicability.

Day 5: Five Golden Rules

Here are the five statistical golden rules for a randomized control trial:

Power Calculation: Do an a priori power calculation that defines the expected effect size, choose a significance level (α, commonly 0.05 and two-sided), specify the desired power (commonly 80-90%) and account for anticipated dropouts.Randomization: Use proper randomization techniques (simple, stratified, or block randomization). Ensure allocation concealment to prevent prediction of group assignment.Outcomes: Clearly defined primary and secondary endpoints. Pick patient-oriented outcomes (POO) rather than disease-oriented outcomes (DOOs), surrogate-oriented outcomes (SOOs) or monitor-oriented outcomes (MOOs)Statistical Analysis Plan: Provides a roadmap for how data will be analyzed and reduces the risk of data dredging or p-hacking. Specify which tests will be used, pre-plan subgroup and sensitivity analyses and define how missing data will be handled.Control of Bias and Confounding: Bias and confounding can distort study results and lead to incorrect conclusions. Use blinding (single, double, or triple) to reduce performance and detection biases. Collect baseline characteristics to assess the balance between gro...

Show More
Rate

Join Podchaser to...

  • Rate podcasts and episodes
  • Follow podcasts and creators
  • Create podcast and episode lists
  • & much more

Episode Tags

Do you host or manage this podcast?
Claim and edit this page to your liking.
,

Unlock more with Podchaser Pro

  • Audience Insights
  • Contact Information
  • Demographics
  • Charts
  • Sponsor History
  • and More!
Pro Features