Understanding effect sizes: Cohen's d, eta-squared, and R-squared

Updated March 2026

A p-value tells you whether an effect exists. An effect size tells you how large the effect is. APA 7th edition requires effect sizes for all statistical tests, and an increasing number of journals will reject manuscripts that report only p-values.

This guide covers the most common effect size measures, when to use each, and how to interpret them.

Why effect sizes matter

Consider two studies that both find p < .05:

Both are statistically significant. But Study A found a small effect in a large sample, while Study B found a very large effect in a small sample. Without the effect size, you'd treat these as equivalent — they're not.

Effect sizes are essential for:

Cohen's d: comparing two groups

Cohen's d measures the standardized difference between two group means. It expresses the difference in standard deviation units.

Formula

d = (M1M2) / SDpooled

Interpretation benchmarks (Cohen, 1988)

Magnitude Cohen's d What it means
Small 0.20 The groups overlap substantially; the difference is subtle
Medium 0.50 The difference is noticeable and often practically meaningful
Large 0.80 The groups are clearly different; minimal overlap in distributions

When to use Cohen's d

APA example

t(48) = 3.45, p = .001, d = 0.97, 95% CI [0.38, 1.56]

Eta-squared and partial eta-squared: ANOVA effect sizes

Eta-squared (η2)

The proportion of total variance in the outcome explained by the grouping variable. Used with one-way ANOVA.

Formula

η2 = SSbetween / SStotal

Partial eta-squared (η2p)

The proportion of variance explained by one factor after removing variance explained by other factors. Used with factorial ANOVA and repeated measures ANOVA.

Formula

η2p = SSeffect / (SSeffect + SSerror)

Interpretation benchmarks

Magnitude η2 / η2p What it means
Small .01 The factor explains about 1% of variance
Medium .06 The factor explains about 6% of variance
Large .14 The factor explains about 14% or more of variance

Important: Partial eta-squared is always larger than eta-squared for the same data (because the denominator is smaller). Don't compare partial eta-squared from one study to eta-squared from another. Always note which measure you're using.

R-squared: regression effect sizes

R2 (the coefficient of determination) measures the proportion of variance in the outcome that is explained by the predictor(s) in a regression model.

For multiple regression, report adjusted R2, which penalizes for the number of predictors. This prevents overfitting from inflating the apparent effect size.

APA example

The model was significant, F(3, 96) = 8.42, p < .001, R2 = .21, adjusted R2 = .18.

Other effect sizes you may encounter

Effect size Used with Small / Medium / Large
Cramér's V Chi-square test .10 / .30 / .50
Rank-biserial r Mann-Whitney U, Wilcoxon .10 / .30 / .50
Odds ratio (OR) Logistic regression 1.5 / 2.5 / 4.3 (Rosenthal, 1996)
Pearson r Correlation .10 / .30 / .50
Hazard ratio (HR) Cox regression Context-dependent; no universal benchmarks

Context matters more than benchmarks

Cohen himself called his benchmarks "small, medium, and large" with the caveat that they were starting points, not rules. In practice:

Always interpret effect sizes in the context of your field, your intervention, and the practical consequences of the effect.

Reporting checklist

  1. Choose the effect size appropriate to your test (Cohen's d for t-tests, η2 for ANOVA, R2 for regression, etc.)
  2. Report the effect size alongside the test statistic and p-value — in the same sentence
  3. Include a confidence interval for the effect size when possible
  4. Note the magnitude label (small/medium/large) but interpret in context
  5. Use the same effect size measure consistently throughout a paper for the same type of comparison

Join the beta to try this in GraphHelix — every statistical test automatically reports the appropriate effect size with a magnitude label and 95% confidence interval.

Join the beta