/ Statistical Tools
All Tools
Non-Parametric · Verified Results · Doctoral Level

Friedman Test
Calculator

A non-parametric test for detecting differences across three or more related groups or repeated measurements. Produces a verified chi-square statistic, Kendall's W effect size, Bonferroni-corrected post-hoc comparisons, and complete APA 7th, Chicago, and MLA narratives.

Q Statistic (Friedman) Ties Correction Kendall's W Post-hoc Pairwise Bonferroni Adjustment APA 7th Reporting Rows = Subjects, Columns = Conditions

The Friedman Test: Conceptual Foundation

The Friedman test is a non-parametric statistical procedure for evaluating whether three or more related groups or repeated measurement conditions differ significantly from one another. It was introduced by Milton Friedman in 1937 in the Journal of the American Statistical Association as a method that avoids the normality assumption inherent in parametric analysis of variance.

The procedure works by ranking observations within each block (subject or matched group) from lowest to highest. If no true differences exist among conditions, the ranks across blocks would distribute randomly, producing approximately equal column rank totals. Systematic differences, in contrast, produce consistently high or low ranks in certain columns, generating a large test statistic.

Non-Parametric Alternative

The Friedman test is the direct non-parametric counterpart of the one-way repeated measures ANOVA. It is appropriate when the dependent variable is measured at the ordinal level, or when interval or ratio data violate the normality assumption required by parametric tests.

Statistical Formula

The Friedman test statistic Q is computed from rank sums across conditions. For N subjects and k conditions, each row of data is ranked from 1 to k. The column rank totals Rj are then used in the following sequence of calculations.

Step 1 — Friedman Q Statistic (Uncorrected)
Q =  12  /  [N × k × (k + 1)]  ×  ΣRj²  −  3N(k + 1)
NNumber of subjects (rows / blocks)
kNumber of conditions (columns)
RjSum of ranks in condition j
ΣRj²Sum of squared column rank totals
Step 2 — Ties Correction Factor (CF)
CF  =  1 − Σ(ti³ − ti)  /  [N × k × (k² − 1)]
Applied when two or more values in the same row are equal. When no ties are present, CF = 1 and the corrected statistic is identical to the uncorrected statistic.
tiNumber of tied observations in tie group i within any row
Step 3 — Corrected Q Statistic
Qcorrected  =  Q  /  CF
This corrected value is used to obtain the p-value from the chi-square distribution with df = k − 1. When CF = 1 (no ties), this step is omitted.
Step 4 — Degrees of Freedom
df  =  k − 1
Q is evaluated against the chi-square distribution with k − 1 degrees of freedom. For three conditions, df = 2.
Effect Size — Kendall's Coefficient of Concordance (W)
W  =  Q  /  [N × (k − 1)]
W ranges from 0 (no agreement) to 1 (perfect concordance). It quantifies the degree to which all subjects rank the conditions in the same order. Interpreted using Landis & Koch (1977) benchmarks.
Post-hoc Standard Error (SE)
SE  =  √[N × k × (k + 1)  /  6]
Used to compute pairwise z-statistics after a significant omnibus result. The z-statistic for each pair is (Ra − Rb) / SE, evaluated two-tailed with Bonferroni correction.

When to Use the Friedman Test

ConditionRequirementExample
Number of groupsThree or more related conditionsPre-test, Post-test 1, Post-test 2
Measurement levelOrdinal, interval, or ratioLikert ratings, pain scores, reaction times
Sample structureRepeated measures on the same subjects, or randomised blocks of matched subjectsSame participants rated under three conditions
NormalityNot assumed; violations are acceptableSkewed distributions, small samples
Independence of blocksEach row (subject) is independent of other rowsNo cluster effects across subjects

Assumptions

The Friedman test requires four conditions to hold. The dependent variable must be measured at the ordinal level or higher. The k groups must represent related samples — either repeated observations from the same individuals, or matched subjects in randomised blocks. There must be a minimum of three conditions; the test is not applicable for two-group comparisons, for which the Wilcoxon signed-rank test is appropriate. Finally, observations within each block are assumed to be mutually independent, meaning no carryover or order effects systematically influence the data.

Effect Size: Kendall's Coefficient of Concordance (W)

Kendall's W quantifies the degree of agreement among blocks (subjects) regarding the ordering of conditions. It is the effect size measure for the Friedman test and ranges from 0 (no agreement) to 1 (perfect agreement). It is computed as W = Q / [N(k − 1)].

W Range Interpretation Visual Reference
< 0.10 Negligible concordance Landis & Koch (1977)
0.10 – 0.29 Weak concordance Landis & Koch (1977)
0.30 – 0.49 Moderate concordance Landis & Koch (1977)
0.50 – 0.69 Strong concordance Legendre (2005)
0.70 – 1.00 Very strong concordance Legendre (2005)

Friedman Test Calculator

Configure the study design, build the data table, enter your observed scores, and run the test. Results include the Friedman Q statistic, p-value, Kendall's W effect size, Bonferroni-corrected post-hoc comparisons, and four reporting narratives.

Data Orientation — Read Before Entering Data

Each row represents one subject. Each column represents one condition. Enter the score that subject received under that condition at the intersection of their row and column. This orientation is the same used by SPSS and matches how most published Friedman test tutorials present data. If you enter data column-first (all subjects for Condition 1, then all subjects for Condition 2), the calculator will rank within the wrong groupings and produce incorrect results.

Example for 3 subjects and 3 conditions: Row 1 contains Subject 1's scores under Condition A, Condition B, and Condition C. Row 2 contains Subject 2's scores. Row 3 contains Subject 3's scores.

Study Design Configuration
Minimum 3 subjects
Minimum 3 conditions
Two-tailed
Data Entry — Enter Observed Scores
Complete Statistical Output
Column Rank Totals and Mean Ranks
Rank Assignment Matrix

Ranks assigned within each row (subject). Tied values receive the average of their tied ranks. The Friedman test statistic is computed from these rank assignments.

Post-hoc Pairwise Comparisons (Bonferroni Correction)

Reporting Narratives

Frequently Asked Questions

Both are non-parametric alternatives to ANOVA, but they address different study designs. The Kruskal-Wallis test applies to independent groups — different participants in each condition. The Friedman test applies to related groups — the same participants measured under each condition, or matched subjects in randomised blocks. Choosing the wrong test when groups are related inflates error variance and reduces statistical power.

A non-significant result means the data do not provide sufficient evidence to conclude that any condition differs from the others at the chosen significance level. Post-hoc comparisons are not warranted following a non-significant omnibus test, and individual pairwise comparisons should not be conducted or reported. The interpretation should state that no statistically significant differences were observed across conditions, acknowledge the possibility of a Type II error, and note that the conclusion is limited to the observed sample.

When two or more values within the same row are equal, they receive the average of the ranks they would have occupied had they been distinct. The uncorrected Friedman statistic is then divided by a correction factor that accounts for the degree of tying present in the data. The correction factor equals one minus the sum of all tie correction terms divided by the product of N, k, and k squared minus one. In the absence of ties, the correction factor equals one and the corrected statistic is identical to the uncorrected statistic. This calculator applies the ties correction automatically.

The minimum is three subjects (N = 3) with three conditions (k = 3), but the chi-square approximation becomes increasingly accurate as sample size grows. For three conditions, N of at least 10 is generally recommended for the chi-square approximation to be reliable. For larger k, the approximation holds at smaller N. Exact critical values from Friedman's original tables are available for small samples (N < 10 with k = 3, or N < 6 with k = 4) and should be consulted when sample sizes are very small. This calculator uses the chi-square approximation, which is standard practice in published research.

For k conditions, there are m = k(k − 1)/2 possible pairwise comparisons. The Bonferroni correction adjusts the significance threshold to α/m and multiplies each raw p-value by m (capping at 1.000) to produce an adjusted p-value. This controls the familywise error rate — the probability of making at least one Type I error across all comparisons — at the chosen α level. The pairwise z-statistic for each comparison is derived from the difference in column rank totals divided by the standard error, which equals the square root of N times k times k plus one divided by six.