Skip to calculator
Free tool

Sample Size & Power Calculator

Calculate the number of participants your study needs. Supports t-tests, ANOVA, chi-square, correlation, regression, two proportions, and animal studies. No install, no account required.

Loading calculator...

Power Analysis Resources

How to Choose an Effect Size Recommended reading

The effect size you enter is the single most consequential input in a power analysis. It determines your required sample size more than any other parameter. Choosing "medium" by default is tempting but rarely defensible.

Where to find your effect size (in preference order)

  1. Published meta-analyses — The gold standard. Search for meta-analyses in your field that report effect sizes for your outcome. A meta-analytic estimate pools data across multiple studies, giving you the most stable estimate of the true effect.
  2. Prior studies in your domain — If no meta-analysis exists, look at individual studies that measured a similar outcome with a similar population. Extract their reported effect sizes and use the median or a conservative estimate.
  3. Smallest effect size of interest (SESOI) — Ask: "What is the smallest effect that would be practically meaningful?" This approach, gaining traction in methodological literature, grounds your power analysis in clinical or practical relevance rather than statistical convention (Lakens, 2022).
  4. Pilot data — Use cautiously. Pilot studies typically have small samples, producing noisy effect size estimates that can be misleadingly large. If using pilot data, consider reducing the estimate by 20-30%.
  5. Cohen's benchmarks — A last resort. Cohen (1988) proposed small/medium/large benchmarks (e.g., d = 0.2/0.5/0.8 for t-tests), but these are arbitrary and do not correspond to equivalent magnitudes across different effect size metrics. Using them without justification weakens your methods section.

Effect size benchmarks by test type

TestMetricSmallMediumLarge
t-testsCohen's d0.20.50.8
ANOVACohen's f0.100.250.40
Chi-squareCohen's w0.100.300.50
Correlationr0.100.300.50
Regressionf20.020.150.35

Remember: these benchmarks describe what Cohen observed across the behavioral sciences in the 1960s. Your field may be different. An effect of d = 0.2 might be negligible in one context and transformative in another.

References: Cohen, J. (1988). Statistical Power Analysis for the Behavioral Sciences (2nd ed.). Lakens, D. (2022). Sample size justification. Collabra: Psychology, 8(1).

Need help choosing? Read our guide to understanding effect sizes for worked examples across different study designs.

Post-Hoc Power: Why You Don't Need It

Post-hoc power (also called observed power or retrospective power) computes "the power your study had" using the observed effect size after data collection. It sounds useful but is mathematically uninformative.

Why post-hoc power is circular

Post-hoc power is a one-to-one function of the p-value. When your p-value is non-significant (p > 0.05), post-hoc power is always less than 50%. When your p-value is significant, post-hoc power is always greater than 50%. It cannot tell you anything the p-value doesn't already tell you.

As Hoenig and Heisey (2001) demonstrated, observed power calculated from the study's own data is essentially a transformation of the test statistic. Gelman (2019) was more blunt, calling post-hoc power calculations uninformative for the same reason.

What to do instead

If your study produced a non-significant result and you want to understand whether you lacked power:

  • Sensitivity analysis — Given your actual sample size and alpha, what is the minimum effect size you could have detected with 80% power? This is informative because it uses a fixed (assumed) effect size, not the observed one.
  • Confidence interval for the effect size — Report the 95% CI around your observed effect. A wide CI that includes meaningful effect sizes suggests inadequate precision, which is a more direct and informative framing than "low power."
  • Prospective power for future studies — Use effect size estimates from your study (with appropriate caution about overestimation) to plan the next study.

This calculator supports sensitivity analysis mode — switch to it using the analysis mode toggle to determine the minimum detectable effect for your sample size.

References: Hoenig, J. M., & Heisey, D. M. (2001). The abuse of power: The pervasive fallacy of power calculations for data analysis. The American Statistician, 55(1), 19–24. Gelman, A. (2019). Don't calculate post-hoc power using observed estimate of effect size. Annals of Surgery, 269(1).

How to Report Power Analysis in Your Methods Section

Grant applications, ethics reviews, and journal submissions expect a clear power analysis in your methods section. Here is what to include and how to format it.

Required elements

  1. The statistical test you plan to use
  2. Alpha level (typically .05, two-tailed)
  3. Target power (typically .80 or .90)
  4. Expected effect size with justification for the value chosen
  5. Resulting sample size (per group and total)
  6. Attrition adjustment if applicable
  7. Software used for the calculation

APA-style template

For a two-group comparison:

"An a priori power analysis was conducted using [software] to determine the required sample size for an independent-samples t-test. With an expected effect size of d = [value] (based on [justification]), significance level of α = .05 (two-tailed), and power = .80, the required sample size was n = [value] per group (N = [total]). Accounting for an expected attrition rate of [X]%, we plan to recruit [adjusted N] participants."

For ANOVA:

"A power analysis using [software] indicated that to detect a medium effect (f = [value]) in a one-way ANOVA with [k] groups, α = .05, and power = .80, a minimum of n = [value] per group (N = [total]) would be required."

Common reviewer comments to avoid

  • "The power analysis uses Cohen's medium benchmark without justification" — Always cite the source of your effect size
  • "Post-hoc power is not meaningful" — Only report a priori power analyses
  • "It is unclear whether N refers to total or per-group" — Always specify both

Use the Copy APA snippet button in the results to get a pre-formatted methods paragraph you can paste directly into your manuscript. For help reporting your ANOVA results, see our guide to reporting ANOVA results in APA format.

Power Analysis for Animal Studies

Animal research presents unique challenges for power analysis. Ethics committees (IACUCs) require justification that you are using the minimum number of animals necessary to achieve scientifically valid results while minimizing animal use — the "reduction" principle of the 3Rs (Replace, Reduce, Refine).

The resource equation method

When effect size estimates are unavailable or unreliable (common in early-stage animal research), the resource equation method provides a practical alternative to traditional power analysis. It uses the error degrees of freedom (E) of the planned ANOVA design:

  • E = N − k, where N is total animals and k is number of groups
  • Target: 10 ≤ E ≤ 20
  • Below 10: insufficient power to detect meaningful effects
  • Above 20: likely using more animals than necessary

For example, with 4 treatment groups, you need 14–24 total animals (3.5–6 per group, typically rounded to 4–6).

When to use each approach

MethodBest forRequirements
Traditional power analysisStudies with prior effect size estimatesEffect size, alpha, power
Resource equationExploratory studies, novel endpoints, no prior dataNumber of groups only
Pilot studyGenerating effect size estimates for the definitive studyMinimum viable sample

This calculator supports the resource equation method — select "Animal study (resource equation)" in the test selection flowchart to calculate the recommended range of animals per group.

Reference: Mead, R. (1988). The Design of Experiments. Cambridge University Press. Festing, M. F. W., & Altman, D. G. (2002). Guidelines for the design and statistical analysis of experiments using laboratory animals. ILAR Journal, 43(4), 244–258.

Statistical Tests Supported

This calculator supports sample size and power calculations for nine statistical methods. Each method addresses a different research question.

TestUse whenEffect size metric
Independent t-testComparing the means of two separate groups (e.g., treatment vs. control)Cohen's d
Paired t-testComparing two measurements from the same subjects (e.g., before vs. after)Cohen's d
One-sample t-testComparing a sample mean to a known population valueCohen's d
One-way ANOVAComparing means across three or more groupsCohen's f
Chi-square testTesting association between two categorical variablesCohen's w
Pearson correlationTesting whether two continuous variables are linearly relatedr
Multiple regressionTesting whether a set of predictors explains variation in an outcomef2
Two proportionsComparing success/failure rates between two groupsArcsine-transformed difference
Animal study (resource equation)Planning animal experiments when effect size is unknownError degrees of freedom (E)

Not sure which test to use? Our guide to choosing a statistical test walks through the decision based on your research question and data type.