ANOVA Calculator
Perform one-way ANOVA (Analysis of Variance) to test differences between group means. Calculate F-statistic, p-value, and determine statistical significance.
Enter data for each group (comma or space separated):
F-Statistic
40.4086
P-Value
< 0.0001
Critical Value
3.8853
Decision
Reject H₀
Conclusion
There is a statistically significant difference between at least two group means.
ANOVA Table
| Source | SS | df | MS | F | P-value |
|---|---|---|---|---|---|
| Between Groups | 250.53 | 2 | 125.27 | 40.41 | < 0.0001 |
| Within Groups | 37.2 | 12 | 3.1 | - | - |
| Total | 287.73 | 14 | - | - | - |
Effect Size
η² (Eta-squared) = 0.8707(Large)
ω² (Omega-squared) = 0.8401
87.1% of variance explained by group membership
Group Summaries
Grand Mean: 24.87
Total N: 15
Group 1: n=5, M=24.6, SD=2.07
Group 2: n=5, M=30, SD=1.58
Group 3: n=5, M=20, SD=1.58
ANOVA Assumptions
- • Independence: Observations must be independent
- • Normality: Each group should be approximately normally distributed
- • Homogeneity of Variance: Groups should have similar variances
- • Note: ANOVA is robust to moderate violations with equal sample sizes
Related Calculators
About This Calculator
ANOVA (Analysis of Variance) is a statistical method for testing whether there are significant differences between the means of three or more groups. It extends the t-test concept to multiple groups by analyzing variance components. This calculator performs one-way ANOVA with complete statistical output.
What is ANOVA? ANOVA tests the null hypothesis that all group means are equal against the alternative that at least one differs. Instead of comparing means directly, it compares the variance between groups to the variance within groups. A large ratio (F-statistic) suggests group means differ significantly.
Why Use ANOVA?
- Compare multiple groups simultaneously (avoids multiple t-tests)
- Controls overall Type I error rate
- Provides a single test for overall differences
- Foundation for more complex experimental designs
Key Concepts:
- Between-group variance: How much group means differ from the grand mean
- Within-group variance: How much individual values vary within each group
- F-statistic: Ratio of between to within variance
- Effect size (η²): Proportion of variance explained by group membership
This calculator handles one-way ANOVA with 2-5 groups. For comparing just two groups, see our T-Test Calculator. For related statistical analysis, see our Chi-Square Calculator.
How to Use the ANOVA Calculator
- 1Select the number of groups you want to compare (2-5).
- 2Choose your significance level (α), typically 0.05.
- 3Enter data for each group, separated by commas or spaces.
- 4Ensure each group has at least 2 values.
- 5Review the F-statistic and p-value.
- 6Check the ANOVA table for detailed breakdown.
- 7Examine effect size (η²) for practical significance.
- 8Read the conclusion about group differences.
- 9If significant, consider post-hoc tests to identify which groups differ.
- 10Verify assumptions are reasonably met.
Understanding One-Way ANOVA
One-way ANOVA tests for differences among groups based on one factor.
The Hypotheses
Null Hypothesis (H₀): μ₁ = μ₂ = μ₃ = ... = μₖ All group means are equal.
Alternative (H₁): At least one μᵢ differs At least two groups have different means.
The F-Statistic
F = MSB / MSW
Where:
- MSB = Mean Square Between (variance between group means)
- MSW = Mean Square Within (variance within groups)
Interpretation
- Large F: Between-group variance >> Within-group variance → Groups likely differ
- Small F: Between-group variance ≈ Within-group variance → No evidence of difference
- F ≈ 1: Expected when H₀ is true
Example
Testing if teaching method affects test scores:
- Group 1 (Traditional): M = 75
- Group 2 (Online): M = 72
- Group 3 (Hybrid): M = 80
ANOVA tests whether these differences are statistically significant or could be due to chance.
The ANOVA Table Explained
Understanding each component of the ANOVA output.
Sum of Squares (SS)
SST (Total): Total variation in all data SST = Σ(xᵢⱼ - x̄)²
SSB (Between): Variation between group means SSB = Σnⱼ(x̄ⱼ - x̄)²
SSW (Within): Variation within groups SSW = ΣΣ(xᵢⱼ - x̄ⱼ)²
Relationship: SST = SSB + SSW
Degrees of Freedom (df)
- dfBetween: k - 1 (number of groups minus 1)
- dfWithin: N - k (total observations minus groups)
- dfTotal: N - 1
Mean Squares (MS)
MSB = SSB / dfBetween MSW = SSW / dfWithin
Complete ANOVA Table
| Source | SS | df | MS | F | p |
|---|---|---|---|---|---|
| Between | SSB | k-1 | MSB | MSB/MSW | P(F > f) |
| Within | SSW | N-k | MSW | - | - |
| Total | SST | N-1 | - | - | - |
Effect Size and Practical Significance
Statistical significance doesn't always mean practical importance.
Eta-Squared (η²)
η² = SSB / SST
Proportion of total variance explained by group membership.
| η² Value | Interpretation |
|---|---|
| < 0.01 | Negligible |
| 0.01 - 0.06 | Small |
| 0.06 - 0.14 | Medium |
| ≥ 0.14 | Large |
Omega-Squared (ω²)
ω² = (SSB - dfB × MSW) / (SST + MSW)
Less biased estimate, especially for small samples. Always smaller than η².
Why Effect Size Matters
- Large samples: Can detect tiny, meaningless differences
- Small samples: May miss important differences
- Practical decisions: Need to know if effect is meaningful
Example
With N = 1000 and p = 0.001:
- Statistically significant? Yes
- But if η² = 0.01 (only 1% variance explained)
- Practically significant? Maybe not
Always report: F(df1, df2) = value, p = value, η² = value
Assumptions of ANOVA
ANOVA results are valid when certain conditions are met.
1. Independence
Observations must be independent of each other.
- Random sampling
- No repeated measures (that requires repeated-measures ANOVA)
- No clustering effects
2. Normality
Each group should be approximately normally distributed.
Checking:
- Histograms and Q-Q plots
- Shapiro-Wilk test
Robustness:
- ANOVA is robust with n ≥ 30 per group
- Central Limit Theorem helps with larger samples
3. Homogeneity of Variance
Groups should have similar variances.
Checking:
- Levene's test
- Rule of thumb: largest variance < 3× smallest variance
If violated:
- Use Welch's ANOVA
- Transform data
- Use non-parametric Kruskal-Wallis test
When Assumptions Fail
| Violation | Solution |
|---|---|
| Non-normality | Kruskal-Wallis test |
| Unequal variances | Welch's ANOVA |
| Non-independence | Mixed models |
| All of above | Bootstrap methods |
Post-Hoc Tests
When ANOVA is significant, post-hoc tests identify which groups differ.
Why Needed?
ANOVA only tells us "at least one group differs" - not which ones.
With k groups, there are k(k-1)/2 pairwise comparisons:
- 3 groups: 3 comparisons
- 4 groups: 6 comparisons
- 5 groups: 10 comparisons
Common Post-Hoc Tests
Tukey's HSD (Honestly Significant Difference)
- Most common choice
- Controls family-wise error rate
- Good for equal sample sizes
Bonferroni
- Divides α by number of comparisons
- Conservative (low power)
- Good for few planned comparisons
Scheffé
- Most conservative
- Good for unplanned comparisons
- Controls for ALL possible contrasts
Games-Howell
- Use when variances are unequal
- Does not assume equal variances
Choosing a Test
| Situation | Recommended Test |
|---|---|
| Equal n, equal variance | Tukey HSD |
| Unequal n, equal variance | Tukey-Kramer |
| Unequal variance | Games-Howell |
| Few planned comparisons | Bonferroni |
| Exploratory | Scheffé |
Comparing ANOVA to Other Tests
Choosing the right statistical test for your data.
ANOVA vs. Multiple T-Tests
Problem with multiple t-tests:
- 3 groups = 3 t-tests
- At α = 0.05, Type I error rate becomes 1 - (0.95)³ = 14.3%
- More comparisons = higher error rate
ANOVA advantage:
- Single test controls overall α
- F-test provides omnibus result
ANOVA vs. T-Test
| Situation | Test |
|---|---|
| 2 groups | t-test (or ANOVA - equivalent) |
| 3+ groups | ANOVA |
For 2 groups: F = t², and p-values are identical.
One-Way vs. Two-Way ANOVA
One-Way: One factor (independent variable)
- Example: Effect of drug type on blood pressure
Two-Way: Two factors
- Example: Effect of drug type AND dosage on blood pressure
- Can test for interaction effects
ANOVA vs. Kruskal-Wallis
| ANOVA | Kruskal-Wallis |
|---|---|
| Parametric | Non-parametric |
| Assumes normality | No distribution assumption |
| Compares means | Compares medians/ranks |
| More powerful when assumptions met | More robust |
Pro Tips
- 💡Check assumptions before running ANOVA: independence, normality, equal variances.
- 💡Use equal sample sizes when possible - increases robustness to violations.
- 💡Always report effect size (η²) alongside p-value.
- 💡If ANOVA is significant, follow up with post-hoc tests.
- 💡Consider practical significance, not just statistical significance.
- 💡For 2 groups, t-test and ANOVA give identical results.
- 💡With unequal variances, use Welch's ANOVA.
- 💡Multiple t-tests inflate Type I error - use ANOVA instead.
- 💡Non-parametric alternative: Kruskal-Wallis test.
- 💡ANOVA tests if ANY groups differ, not which ones.
- 💡Larger samples make ANOVA more robust to assumption violations.
- 💡Report: F(df1, df2) = value, p = value, η² = value.
Frequently Asked Questions
A significant result (p < α) means at least one group mean differs significantly from at least one other. It does NOT tell you which groups differ - you need post-hoc tests for that. It also doesn't mean all groups differ from each other.

