Sample Size Planning Framework
Compact checklist to choose or justify sample size with statistical validity, resource efficiency, and ethical balance.
1. Core Formulas (Most Frequent)
Survey proportion (worstācase p=0.5): n = Z²·0.25 / E²
Proportion (general): n = Z²·p(1āp)/E²
Mean (Ļ known approx): n = (ZĀ·Ļ/E)²
Two means (equal n): n = 2(Z_{α/2}+Z_β)² ϲ / Γ²
Two proportions: n = (Z_{α/2}+Z_β)² [pā(1āpā)+pā(1āpā)] / (pāāpā)²
Finite pop adjust: n_adj = n / (1+(nā1)/N)
Design effect: n_eff = nĀ·DEFF = n[1+(mā1)ICC]
Proportion (general): n = Z²·p(1āp)/E²
Mean (Ļ known approx): n = (ZĀ·Ļ/E)²
Two means (equal n): n = 2(Z_{α/2}+Z_β)² ϲ / Γ²
Two proportions: n = (Z_{α/2}+Z_β)² [pā(1āpā)+pā(1āpā)] / (pāāpā)²
Finite pop adjust: n_adj = n / (1+(nā1)/N)
Design effect: n_eff = nĀ·DEFF = n[1+(mā1)ICC]
2. Minimal Workflow
- Clarify objective (estimate / detect difference / correlation).
- Select α (commonly 0.05) and desired power (ā„0.80).
- Specify target effect (Ī“, p difference, r) using prior data or conservative assumption.
- Choose appropriate formula or power method (software) for design.
- Adjust for design: clustering, stratification, unequal allocation, finite N.
- Inflate for expected attrition/nonāresponse.
- Round up; document every assumption source.
3. Parameter Selection Heuristics
- Unknown p ā use 0.5 (max variance).
- Lacking effect size ā pilot or literature metaāanalysis; else pick minimal clinically important difference.
- Attrition inflation: n_final = n_calc / (1 ā drop%)
- Unequal allocation k = nā/nā ā inflate by (1+k)²/(4k).
4. Design Adjustments
- Cluster trials: multiply by DEFF = 1 + (mā1)ICC.
- Strata with disproportionate sampling: compute per stratum then sum.
- Sequential / adaptive: add interim looks penalty (alpha spending).
- Multiple endpoints/tests: control FWER (Bonferroni) or FDR ā raises n if strict.
5. Quick Z & Power Reference
Z(90%)=1.645 · Z(95%)=1.96 · Z(97.5%)=2.24 · Z(99%)=2.576 | Typical power Z_β (80%)=0.84, (90%)=1.28
6. Quality / Sanity Checks
- Result smaller than pilot sample? Reāexamine variance/effect assumptions.
- Computed n impractically large ā consider redefining effect or alternative design.
- Effect size inflation risk: compare assumed Ī“ to historical distribution.
7. Common Pitfalls
- Postāhoc āretrospective powerā misuse (report CI instead).
- Ignoring correlation in repeated measures (overestimates n).
- Using tātest formula when outcome is proportion / count.
- Not adjusting for high anticipated nonāresponse.
- Overly optimistic Ī“ leading to underpowered study.
8. FAQ
- Why p=0.5 default? Maximizes p(1āp) giving largest required n (safe upper bound).
- Power vs significance? Power controls false negatives; α controls false positivesāthey balance study risk.
- When use FPC? When sample > ~5ā10% of finite population.
- Cluster vs individual n? First compute individual n then multiply by design effect.
9. Action Tip
Create a reproducible āassumptions tableā (α, power, effect, variance, design effects, attrition) and store with protocolāprevents silent later changes and supports transparency.