Bored Analyst

Experiment Analytics

Analytics for experiments: A/B testing, statistical significance, and causal inference. Core for product and growth teams.

Key concepts

  • Null hypothesis — No difference between control and treatment
  • Statistical significance — p-value, confidence intervals
  • Power — Probability of detecting a real effect; sample size
  • MDE — Minimum detectable effect for a given sample size

Experiment design

  • Randomization — Unit of assignment (user, session, page)
  • Sample size — Power analysis, duration, traffic allocation
  • Guardrail metrics — Metrics to monitor for unintended harm
  • Primacy and novelty — Early effects that may not persist

Analysis

  • Two-sample tests — t-test, z-test for means and proportions
  • CUPED — Variance reduction for faster, more precise experiments
  • Segmentation — Heterogeneous treatment effects by segment
  • Multiple comparisons — Bonferroni, FDR when testing many variants

Causal inference

  • When A/B isn't possible — Diff-in-diff, synthetic control, propensity scoring
  • Instrumental variables — When treatment isn't randomly assigned
  • Regression discontinuity — Natural experiments at a threshold

Common analyses

  • Experiment results with confidence intervals and p-values
  • Power and sample size calculations
  • CUPED-adjusted analysis for faster ship decisions
  • Meta-analysis across experiments

On this page