PEP 6305 Measurement in
Health & Physical Education
This Topic has 2 Sections.
Vincent & Weir, Statistics in Kinesiology, 4th ed. Chapter
12 “Analysis of Variance with Repeated Measures”
To demonstrate the comparison of measures repeatedly collected
from the same group of subjects using analysis of variance (ANOVA).
Measures: Within Subjects Factors
Repeated measures research, in which the same measure is collected
several times in a single group of subjects, is very common in studies of
kinesiology and health.
Pretest-posttest designs are one example, in which a measure
collected before a treatment is compared to the same measure collected after the
Repeated measures designs are also called within subjects
The repeated factor is called a “within” subjects factor because
comparisons are made multiple times ("repeated") “within” the same subject rather than across ("between")
requires that the means being compared are
independent of one another.
When the subjects are in groups that are independent from one another, then
group is a between subjects factor.
Since repeated measures are collected on the same subjects, the
means of those measures study are dependent.
A particular subject’s scores will be more alike than scores
collected from multiple subjects, meaning that there is less variability from
measure to measure than observed from person to person in simple ANOVA.
Repeated measures ANOVA separates the
sources of variance:
persons. This separation of the sources of variance decreases MSE,
the random variation (sampling error) component, because there are now two
sources of known variation (subjects and measures) instead of just one
(subjects) as in simple ANOVA. The variation in scores due to differences between subjects
is separated from variation due to differences from measure to measure within
a subject. (A smaller MSE increases
Instead of comparing treatment effects to a group of different
subjects, treatment effects are compared across multiple measures in
the same subjects. Each subject provides their own "control" value for the comparison.
Consequently, this type of design is more sensitive to differences
(i.e., requires smaller differences in the dependent variable to reject
the null hypothesis) than are between subjects designs.
Repeated measures ANOVA requires different computation
than simple ANOVA.
The repeated measures ANOVA null hypothesis is that the
means of the measures all have the same value. If the number of
repeated measures = k , the null hypothesis is:
or the differences between the means of each repeated measure is equal to 0.
The statistic used in repeated measures ANOVA is F, the
same statistic as in simple ANOVA, but now computed using the ratio of the
variation “within” the subjects to the “error” variation.
The observed between measures variance is an estimate of
the variation between measures that would be expected in the population under
the conditions of the study.
The observed error variance is an estimate of the variation
that would be expected to occur as a result of sampling error alone.
If the observed (computed) value for F is
higher than the value expected by sampling variation alone, then the variance
between groups is larger than would be expected by sampling error alone.
In other words, at least one mean differs from the others enough
to cause large variation between the measures.
The assumptions of repeated measures ANOVA are similar to simple
ANOVA, except that independence
is not required and an assumption about the relations among the repeated
measures (sphericity) is added.
The dependent variable is
distributed in the population being sampled.
Normality of the dependent variable can be evaluated using a
and kurtosis statistics.
Homogeneity of variance
If there are separate groups of subjects in addition to the
repeated measure (within subjects) factor, then the variance of the dependent
variable in each group is equal (in the population).
As in simple ANOVA,
variance can be evaluated using a variety of statistical tests, but the most
straightforward method is to compare the within-group variances; one or more
variances twice as large as other variances may be a problem.
This means that the variances of the repeated measures are all equal,
and the correlations among the repeated measures are all equal.
This assumption is needed to allow for comparing the variances
among the repeated measures (within subjects).
When evidence suggests sphericity is violated, the analysis is
adjusted. We’ll discuss this briefly
the next section.
“Within Subjects” Variance and “Error”
Means are computed for each measure (typically the
measures are in the columns of the data file) by summing across the
subjects and dividing by the number of subjects:
where Xij is the score of a person (i) for measure j
and N is the number of subjects.
Means are computed for each subject (typically the
subjects are the rows of the data file) by summing across the
measures and dividing by the number of measures:
where J is the total number of measures in the study.
The grand mean is computed by summing across all measures and all
subjects and dividing by the total number of scores (N × j):
The between measures variance is the variation of the
mean of each measure from the grand mean.
The between subjects variance is the variation of the
mean of each subject from the grand mean.
The total variance is the variation of each score
(each measure for each subject) from the grand mean.
The error variance is any variation not accounted for by
the variation among the subjects and the variation among the measures.
This "residual" variance = total variance – between
subjects variance – between measures variance.
Sum of Squares
Each component (Between Measures, Between Subjects,
Error/Residual, and Total) has a SS.
Each of these SS is a measure of
Repeated measures ANOVA compares the between measures
variability to the error variability using the F statistic.
Both SS measures are standardized by their respective
degrees of freedom,
creating Mean Squares.
Compute the MS by dividing the SS by its respective
degrees of freedom (df).
df = (the number of elements being summed in the SS) – (the
number of means subtracted in the SS)
For each SS equation, compute the difference between the number
of elements ahead of the subtraction sign from the number of elements
behind the subtraction sign.
dfW : How many elements are being summed in the SSM
equation above? The J measures’ means. How many means are subtracted from
these J means? 1—the grand mean is subtracted from the J means. So
dfM = J – 1.
dfB : How many elements are being summed in the SSS
equation above? The N subjects' means. How many means are subtracted from
these N means? The grand mean is subtracted from the n means (in
each group). So dfS = N – 1.
dfT : What is the df for SST? N × J
elements (all subjects and all measures) are summed, and then the grand mean is
subtracted, so dfT = NJ – 1.
dfE : What is the df for SSE? dfE
= [dfT – dfM – dfS] = [(NJ – 1) – (J
– 1) – (N – 1)] = [NJ – 1 – J + 1 – N + 1] = [NJ – J – N + 1] = (N – 1)(J – 1).
Each component (Between Measures, Between Subjects, and Error) in
ANOVA has a MS : divide each SS by its df.
Compute F :
MSE represents the variation expected as a result of sampling
As in simple ANOVA, we compare the observed sample F value
to the F distribution to determine the likelihood that the value is due
to sampling error alone.
If MSM is several times larger than MSE,
we conclude that the variation between measures is larger than sampling error.
In other words, the error probability (p-value) will be low, and we infer that
the measures are significantly different.
The key difference between simple ANOVA and repeated measures
ANOVA is that the variation from measure to measure and the variation
from subject to subject have been removed from the
SSE in repeated measures, thus adjusting the SSE
(and MSE) for the within-subjects dependency of the measures.
to go to the next section (Section 10.2)