F-Test Calculator

Compute the F-statistic for testing equality of two variances, ANOVA model fit, or regression significance. Includes degrees of freedom, critical values, and variance ratio confidence interval.

F-Statistic
df₁ (numerator)
df₂ (denominator)
Interpretation (α=0.05)
Extended More scenarios, charts & detailed breakdown
F-Statistic
df₁
df₂
Variance Ratio (σ₁²/σ₂²)
Decision
Professional Full parameters & maximum detail

F-Test Result

F-Statistic
df₁
df₂

Variance Details

Variance 1 (s₁²)
Variance 2 (s₂²)
95% CI for Variance Ratio

Notes

Robustness Note

How to Use This Calculator

  1. Enter the two sample variances and sample sizes for an equality-of-variances test.
  2. The larger variance goes in Variance 1 (calculator will auto-detect).
  3. Use F in ANOVA tab if you have MS between/within from an ANOVA table.
  4. Use F in Regression tab to test overall model significance from SS regression and residual.
  5. Professional tab computes 95% CI for the variance ratio and flags robustness concerns.

Formula

Variance equality: F = s₁² / s₂² (larger over smaller), df₁ = n₁−1, df₂ = n₂−1

ANOVA: F = MS_between / MS_within

Regression: F = MS_regression / MS_residual

Example

s₁² = 36 (n₁=15), s₂² = 20 (n₂=12): F = 36/20 = 1.80, df₁=14, df₂=11. F_crit ≈ 3.09 (α=0.05). Not significant — variances are equal.

Frequently Asked Questions

  • The F-test is a statistical hypothesis test that uses the F-distribution to compare two variances or assess overall model fit. In its simplest form — the variance equality F-test — it tests whether two population variances are equal by computing the ratio of the larger sample variance to the smaller: F = s₁² / s₂². Under the null hypothesis that both populations have equal variance, this ratio follows an F-distribution with (n₁−1) and (n₂−1) degrees of freedom. The F-distribution is always non-negative and right-skewed. Large F-values (far from 1.0) suggest the variances differ significantly. In ANOVA, the F-test compares mean square between groups to mean square within groups, testing whether any group means differ. In linear regression, the F-test assesses whether the model as a whole explains a statistically significant portion of the variance in the outcome. The F-statistic is named after R.A. Fisher, though George Snedecor formalized the F-distribution in 1934.
  • The t-test and F-test address fundamentally different hypotheses. A t-test compares means: 'Are the average values in these groups the same?' An F-test for variance equality compares spread: 'Is the variability in these groups the same?' These are separate questions and require separate tests. For example, two student cohorts might have the same average exam score (t-test not significant) but very different variability — one cohort scores consistently around 75 while the other ranges from 40 to 100 (F-test significant). There is a mathematical relationship: a t-statistic squared equals an F-statistic with df₁=1, which is why an F-test with one degree of freedom in the numerator is equivalent to a two-tailed t-test. In ANOVA, the F-test is the natural generalization of the t-test to more than two groups. But when comparing exactly two group means with equal variances, the t-test and F-test give identical p-values. In regression, the F-test assesses overall model significance while t-tests assess individual predictor significance.
  • These are two distinct applications of the F-distribution. The variance equality F-test is a preprocessing step or standalone test: use it before a two-sample t-test to decide whether to apply Student's (equal variance) or Welch's (unequal variance) version. You're testing a structural assumption about the data-generating process. The ANOVA F-test is the main inferential test: it determines whether group means differ. It uses MS_between / MS_within rather than a direct variance ratio. In ANOVA, you assume variances ARE equal (homoscedasticity) — that's an assumption you check with Levene's test, not with the F-test for variances. Conflating these two uses is a common error. The variance equality F-test produces a ratio of two sample variances. The ANOVA F-test produces a ratio of mean squares representing different sources of variance. Both follow F-distributions but with different degrees of freedom and different null hypotheses.
  • The F-test for variance equality is notoriously sensitive to departures from normality — far more so than the t-test. This is because the ratio of two chi-squared random variables follows an F-distribution only when the underlying data is normally distributed. Even mild skewness or kurtosis in the data can cause the F-test to produce highly inflated Type I error rates (false rejections of H₀). A Monte Carlo simulation by Box in 1953 showed that F-tests for variance equality with non-normal data can have actual Type I error rates of 20–40% even when the nominal α is 0.05. For this reason, statisticians now generally recommend Levene's test (which tests whether absolute deviations from the group median differ) or the Brown-Forsythe test (using medians instead of means) as more robust alternatives. Bartlett's test, while more powerful under normality, is even more sensitive to non-normality than the classical F-test. In practice, the ANOVA F-test for means is much more robust to non-normality than the variance equality F-test.
  • There is no single universal critical F-value — it depends on the degrees of freedom in the numerator and denominator, and the chosen significance level α. For the variance equality F-test with n₁=15, n₂=12 (df₁=14, df₂=11) at α=0.05 (two-tailed, so α/2=0.025 per tail), the critical value is approximately 3.09. For ANOVA with k=3 groups and N=30 total observations (df₁=2, df₂=27) at α=0.05, the critical F is approximately 3.35. For a regression model with 3 predictors and 46 error degrees of freedom at α=0.05, critical F ≈ 2.80. As degrees of freedom increase, critical F values decrease toward 1.0. F=1.0 means between-group and within-group variances are equal — exactly what the null hypothesis predicts. Rules of thumb: F > 4 is often significant for small studies; F > 2 may be significant for large studies with many degrees of freedom. Always compute the exact p-value rather than relying on a fixed threshold.

Related Calculators

Sources & References (5)
  1. Snedecor 1934 — Calculation and Interpretation of Analysis of Variance (F-distribution) — Iowa State College Press
  2. Fisher — Statistical Methods for Research Workers — Oliver and Boyd
  3. NIST/SEMATECH — F-Test for Equality of Two Variances — NIST
  4. OpenStax Statistics — F Distribution and ANOVA — OpenStax
  5. Stanford Statistics Resources — Hypothesis Testing — Stanford University