Linear Models and Analysis of
Variance:
CONCEPTS, MODELS, AND APPLICATIONS
Volume II
First Edition
David W. Stockburger
Southwest Missouri State
University
@Copyright 1993
TABLE OF CONTENTS
Title Page
EXPERIMENTAL DESIGNS.............................................................................. 132
Notation.................................................................................................... 133
Kinds of Factors........................................................................................ 134
Treatment...................................................................................... 134
Group Factors............................................................................... 135
Trials Factors................................................................................ 135
Blocking........................................................................................ 135
Unit Factors.................................................................................. 137
Error Factors................................................................................. 137
Fixed and Random Factors........................................................................ 137
Fixed Factors................................................................................ 137
Random Factors............................................................................ 138
Relationships Between Factors................................................................... 139
Crossed........................................................................................ 139
Nested.......................................................................................... 140
An Example Design................................................................................... 142
A Second Example Design......................................................................... 144
A Third Example Design............................................................................ 146
Determining the Number of Subjects and
Measures per Subject................. 148
Setting up the Data Matrix......................................................................... 148
A Note of Caution..................................................................................... 149
One Between Group ANOVA.............................................................................. 150
Why Multiple Comparisons Using t-tests is NOT
the Analysis of Choice..... 150
The Bottom Line - Results and Interpretation
of ANOVA.......................... 151
HYPOTHESIS TESTING THEORY UNDERLYING ANOVA.............. 153
The Sampling Distribution Reviewed.............................................. 153
Two Ways of Estimating the Population
Parameter σX˛................... 154
The F-ratio and F-distribution.................................................................... 157
Non-significant and Significant F-ratios....................................................... 159
Similarity of ANOVA and t-test................................................................. 162
EXAMPLE OF A NON-SIGNIFICANT ONE-WAY ANOVA.............. 164
EXAMPLE OF A SIGNIFICANT ONE-WAY ANOVA........................ 164
USING MANOVA.................................................................................. 164
The Data....................................................................................... 165
Example Output............................................................................. 166
Dot Notation............................................................................................. 168
POST-HOC Tests of Significance.............................................................. 170
Example SPSS Program using MANOVA................................................. 173
Interpretation of Output............................................................................ 174
Graphs of Means....................................................................................... 175
The ANOVA Summary Table................................................................... 176
Main Effects.................................................................................. 177
Simple Main Effects....................................................................... 178
Interaction Effects.......................................................................... 178
Example Data Sets, Means, and Summary Tables...................................... 180
No Significant Effects..................................................................... 180
Main Effect of A............................................................................ 181
Main Effect of B............................................................................ 182
AB Interaction............................................................................... 183
Main Effects of A and B................................................................ 184
Main effect of A, AB Interaction ................................................... 185
Main Effect of B, AB Interaction.................................................... 186
Main Effects of A and B, AB Interaction........................................ 187
No Significant Effects..................................................................... 188
Dot Notation Revisited.............................................................................. 189
Nested Two Factor Between Groups Designs B(A)............................................... 192
The Design................................................................................................ 192
The Data................................................................................................... 192
SPSS commands....................................................................................... 193
The Analysis.............................................................................................. 193
The Table of Means....................................................................... 194
Graphs.......................................................................................... 194
The ANOVA Table....................................................................... 195
Interpretation of Output............................................................................. 195
Similarities to the A X B Analysis............................................................... 195
Contrasts, Special and Otherwise........................................................................... 197
Definition................................................................................................... 197
Sets of Contrasts....................................................................................... 198
Orthogonal Contrasts..................................................................... 198
Non-orthogonal Contrasts............................................................. 199
Sets of Orthogonal Contrasts......................................................... 199
Finding Sets of Orthogonal Contrasts............................................ 199
The Data................................................................................................... 201
SPSS commands....................................................................................... 202
The Analysis.............................................................................................. 202
The Table of Means....................................................................... 202
The ANOVA table........................................................................ 203
Interpretation of Output............................................................................. 203
Constants.................................................................................................. 203
Contrasts, Designs, and Effects.................................................................. 205
Non-Orthogonal Contrasts........................................................................ 208
Smaller than Total Sum of Squares................................................. 208
Larger than Total Sum of Squares.................................................. 209
Standard Types of Orthogonal Contrasts.................................................... 209
DIFFERENCE.............................................................................. 210
SIMPLE....................................................................................... 210
POLYNOMIAL........................................................................... 210
Conclusion................................................................................................ 213
ANOVA and Multiple Regression.......................................................................... 214
ONE FACTOR ANOVA......................................................................... 214
ANOVA and Multiple Regression.................................................. 214
Example........................................................................................ 215
Example Using Contrasts............................................................... 216
Dummy Coding............................................................................. 216
ANOVA, Revisited....................................................................... 218
TWO FACTOR ANOVA........................................................................ 219
Example........................................................................................ 220
Example Using Contrasts............................................................... 220
Regression Analysis using Dummy Coding...................................... 221
Conclusion................................................................................................ 223
Unequal Cell Frequencies...................................................................................... 225
Equal Cell Frequency - Independence of
Effects......................................... 225
Unequal Cell Frequency - Dependent Effects............................................. 226
Solutions for Dealing with Dependent Effects.............................................. 229
UNEQUAL CELL SIZES FROM A MULTIPLE REGRESSION
VIEWPOINT 231
REGRESSION ANALYSIS OF UNEQUAL N ANOVA........................ 234
RECOMMENDATIONS......................................................................... 237
Subjects Crossed With Treatments S X A............................................................. 238
The Design................................................................................................ 238
The Data................................................................................................... 239
SPSS commands....................................................................................... 239
The Correlation Matrix.................................................................. 240
The Table of Means....................................................................... 240
Graphs.......................................................................................... 241
The ANOVA Table....................................................................... 241
Interpretation of Output................................................................. 242
Additional Assumptions for Univariate S X A
Designs................................ 244
SS, MS, and Expected Mean Squares (EMS)................................ 245
Subjects Crossed With Two Treatments -S X A X
B............................................. 249
The Design................................................................................................ 249
The Data................................................................................................... 249
SPSS commands....................................................................................... 250
The Correlation Matrix.................................................................. 251
The Table of Means....................................................................... 251
Graphs.......................................................................................... 253
The ANOVA Table....................................................................... 253
Interpretation of Output................................................................. 253
Additional Assumptions for Univariate S X A X
B Designs......................... 258
SS, MS, and Expected Mean Squares (EMS)................................ 258
Mixed Designs - S ( A ) X B................................................................................. 261
The Design................................................................................................ 261
The Data................................................................................................... 261
SPSS commands....................................................................................... 262
The Table of Means....................................................................... 263
Graphs.......................................................................................... 264
Interpretation of Output................................................................. 264
Expected Mean Squares (EMS).................................................... 267
Three Factor ANOVA.......................................................................................... 268
Effects....................................................................................................... 268
Main Effects.................................................................................. 269
Two-Way Interactions................................................................... 269
Three-Way Interaction................................................................... 271
Additional Examples.................................................................................. 272
All Effects Significant..................................................................... 272
Example 3 - B, AC, and BC......................................................... 274
Two More Examples..................................................................... 275
Expected Mean Squares............................................................................ 276
Tests of Significance.................................................................................. 279
Error Terms................................................................................... 279
SPSS Output................................................................................. 279
Examples................................................................................................... 279
S ( A X B X C)............................................................................. 279
S ( A X B ) X C............................................................................ 280
S ( A ) X B X C............................................................................ 280
BIBLIOGRAPHY................................................................................................. 287
INDEX................................................................................................................. 289
Chapter
8
EXPERIMENTAL
DESIGNS
Experimental design refers to the manner in which the
experiment was set up. Experimental design includes the way the treatments were
administered to subjects, how subjects were grouped for analysis, how the
treatments and grouping were combined.
In ANOVA there is a single dependent variable or score. In Psychology the dependent measure is
usually some measure of behavior. If
more than one measure of behavior is taken, multivariate analysis of variance,
or MANOVA, may be the appropriate analysis.
Because the ANOVA model breaks the score into component parts, or
effects, which sum the total score, the one must assume the interval property
of measurement for this variable. Since
in real life the interval property is never really met, one must be satisfied
that at least an approximation of an interval scale exists for the dependent
variable. To the extent that this
assumption is unwarranted, the ANOVA hypothesis testing procedure will not
work.
In ANOVA there is at least one independent variable or factor. There are different kinds of factors; treatment, trial, blocking, and group. Each will be discussed in the following
section. All factors, however, have
some finite number of different levels.
Each level is the same in either some quality or quantity. The only restriction on the number of
levels is that there are fewer levels than scores, although in practice one
seldom sees more than ten levels in a factor unless the data set is very large. It is not necessary that the independent
variables or factors be measured on an interval scale. If the factors are measured on an
(approximate) interval scale, then some flexibility in analysis is gained. The continued popularity of ANOVA can
partially be explained by the lack of the necessity of the interval assumption
for the factors.
Notation
Every writer of an introductory, intermediate, or advanced
statistics text has his or her own pet notational system. I have taught using a number of different
systems and have unabashedly borrowed the one to be described below from Lee
(1975). In my opinion it is the easiest
for students to grasp.
The dependent variable or score will be symbolized by the
letter X. Subscripts (usually multiple)
will be tagged on this letter to differentiate the different scores. For example, to designate a single score
from a group of scores a single subscript would be necessary and the symbol Xs
could be used. In this case X1
would indicate the first subject, X2 the second, X3 the
third, and so forth.
When it is desired to indicate a single score belonging to a
given combination of factors, multiple subscripts must be used. For example, Xabs would describe
a given score for a combination of a and b.
Thus, X236 would describe the sixth score when a=2 and
b=3. Another example, X413,
would describe the third score when a=4 and b=1.
Bolded capital letters will be used to symbolize
factors. Example factors are A, B,
C, ..., Z. Some factor
names are reserved for special factors.
For example, S will always refer to the subject factor, E will always be the error factor, and
G will be the group factor.
Small letters with a numerical subscript are used to
indicate specific levels of a factor.
For example c1 will indicate the first level of factor C,
while cc will indicate a specific level of factor C, but the
level is unspecified. The number of
levels of a factor are given by the unbolded capital letter of that
factor. For example there are 1, 2,
..., C levels of factor C.
In an example experiment, let X, the score, be the dollar
amount after playing WindowsTM Solitaire for an hour. In this experiment the independent variable
(factor) is the amount of practice, called factor A. Let nine subjects each participate in one of
four (A=4) levels of training. The
first level, a1, consists on no practice, a2 = one hour
of practice, a3 = five hours of practice, and a4 = twenty
hours of practice. A given score
(dollar amount) would be symbolized by Xas, where X35
would be the fifth subject in the group that received five hours of practice.
Kinds
of Factors
Treatment
Treatments will be defined as quantitatively or
qualitatively different levels of experience.
For example, in an experiment on the effects of caffeine, the treatment
levels might be exposure to different amounts of caffeine, from none to .0375
milligrams. In a very simple experiment
there are two levels of treatment, none, called the control condition, and
some, called the experimental condition.
Treatment factors are usually the main focus of the
experiment. A treatment factor is
characterized by the following two attributes (Lee, 1975):
1. An investigator could assign any of his
experimental subjects to any one of the levels of the factor.
2. The different levels of the factor
consist of explicitly distinguishable stimuli or situations in the environment
of the subject.
In the solitaire example, practice time would be a treatment
factor if the experimenter controlled the amount of time that the
subject practiced. If subject's came to
the experiment having already practiced a given amount, then the experimenter
could not arbitrarily or randomly assign that subject to a given practice
level. In that case the factor would no
longer be considered a treatment factor.
In an experiment where subjects are run in groups, it
sometimes is valuable to treat each group as a separate level of a factor. There might be, for example, an obnoxious
subject who affects the scores of all other subjects in that group. In this case the second attribute would not
hold and the factor would be called a group factor.
Group Factors
As described above, a group factor is one in which the
subjects are arbitrarily assigned to a given group which differs from other
groups only in that different subjects are assigned to it. If each group had some type of
distinguishing feature, other than the subjects assigned to it, then it would
no longer be considered as a group factor.
If a group factor exists in an experimental design, it will be
symbolized by G.
Trials Factors
If each subject is scored more than once under the same
condition and the separate scores are included in the analysis, then a trials
factor exists. If the different scores
for a subject are found under different levels of a treatment, then the factor would
be called a treatment factor rather than a trials factor. Trials factors will be denoted by T.
Trials factors are useful in examining practice or fatigue
effects. Any change in scores over time
may be attributed to having previously experienced similar conditions.
Blocking
If subjects are grouped according to some pre-existing subject similarity, then that grouping is
called a blocking factor. The
experimenter has no choice but to assign the subject to one or the other of the
levels of a blocking factor. For
example, gender (sex) is often used as a blocking factor. A subject enters the experiment as either a
male or female and the experimenter may not arbitrarily (randomly) assign that
individual to one gender or the other.
Because the experimenter has no control over the assignment
of subjects to a blocking factor, causal inference is made much more
difficult. For example, if in the
solitaire experiment, the practice factor was based on a pre-existing
condition, then any differences between the groups may be due either to
practice or to the fact that some subjects liked to play solitaire, were
better at the game and thus practiced more.
Since the subjects are self-selected, it is not possible to attribute
the differences between groups to practice, enjoyment of the game, natural skill
in playing the game, or some other reason.
It is possible, however, to say that the groups differed.
Even though causal inference is not possible, blocking
factor can be useful. A factor
which accounts for differences in the
scores adds power to the experiment.
That is, a blocking factor which explains some of the differences
between scores may make it more likely to find treatment effects. For example, if males and females performed significantly different in the
solitaire experiment, it might be useful to include sex as a blocking factor
because differences due to gender would be included in the error variance
otherwise.
In other cases blocking factors are interesting in their own
right. It may be interesting to know
that freshmen, sophomores, juniors, and seniors differ in attitude toward
university authority, even though causal inferences may not be made.
In some cases the pre-existing condition is quantitative, as
in an IQ score or weigh. In these cases
it is possible to use a median split where the scores above the median are placed
in one group and the scores below the median are placed in another. Variations of this procedure divide the
scores into three, four, or more approximately equal sized groups. Such procedures are not recommended as there
are better ways of handling such data (Edwards, 1985).
Unit Factors
The unit factor is the entity from which a score is
taken. In experimental psychology, the
unit factor is usually a subject (human or animal), although classrooms,
dormitories, or other units may serve the same function. In this text, the unit factor will be
designated as S, with the understanding that it might be some other type
of unit than subject.
Error Factors
The error factor, designated as E, is not a factor in
the sense of the previous factors and is not included in the experimental
design. It is necessary for future
theoretical development.
Fixed
and Random Factors
Each factor in the design must be classified as either a
fixed or random factor. This is
necessary in order to find the correct error term for each effect. The MANOVA program in SPSS does not require
that the user designate the type for each factor. If the user is willing to accept the program defaults, which are
correct in most cases, no problem is encountered. There are situations, however, where the program defaults are
incorrect and additional coding is necessary to do the correct hypothesis
tests.
Fixed Factors
A factor is fixed if
(Lee, 1975)
1. The results of the factor generalize only
to the levels that were included in the experimental design. The experimenter may wish to generalize to
other levels not included in the factor, but it is done at his or her own
peril.
2. Any procedure is allowable to select
the levels of the factor.
3. If the experiment were replicated, the
same levels of that factor would be included in the new experiment.
Random Factors
A factor is random if
1. The results of the factor generalize to
both levels that were included in the factor and levels which were not. The experimenter wishes to generalize to a
larger population of possible factor levels.
2. The levels of the factor used in the
experiment were selected by a random procedure.
3. If the experiment were replicated,
different levels of that factor would be included in the new experiment.
In many cases an exact determination of whether a factor is
fixed or random is not possible. In
general, the subjects (S) and groups (G) factors will always be a
random factor and all other factors will be considered fixed. The default designation of MANOVA will set
the subjects factor as random and all other factors as fixed.
Some reflection on the assumption of a random selection of
subjects may cause the experimenter to question whether it is in fact a random
factor. Suppose, as often happens,
subjects volunteered to participate in the experiment. In this case the assumptions underlying the
ANOVA are violated, but the procedure is used anyway. Seldom, if ever, will all the assumptions necessary to do an
ANOVA be completely satisfied. The
experimenter must examine how badly the assumptions were violated and then make
a decision as to whether or not the ANOVA is useful.
In general, when in doubt as to whether a factor is fixed or
random, consider it fixed. One should
never have so much doubt, however, as to consider the subjects factor as a
fixed factor.
Relationships
Between Factors
The following two relationships between factors describe a
large number of useful designs. Not all
possible experimental designs fit neatly into categories described by the
following two relationships, but most do.
Crossed
When two factors are crossed, each level of each factor
appears with each level of the other factor.
A crossing relationship is indicated by an "X".
For example, consider two factors, A and B,
were A is gender (a1 = Females, a2 = Males) and B
is practice (b1 = none, b2 = one hour, b3 =
five hours, and b4 = twenty hours).
If gender was crossed with practice, A X B, then both
males and females would participate in all four levels of practice. There would be eight groups of subjects
including: ab11, females who
had no practice, ab12, females who had one hour of practice, and so
forth to ab24, males who practiced twenty hours. An additional factor may be added to the
design, say handedness (C), where c1 = right handed and c2
= left handed. If the design of the
experiment was A X B X C, then there would be sixteen groups, including abc231,
left-handed males who practiced five hours.
If subjects (S) are crossed with treatments (A),
S X A, each subject sees each level of the treatment conditions. In a very simple experiment such as the
effects of caffeine on alertness (A), each subject would be exposed to
both a caffeine condition (a1) and a no caffeine condition (a2). For example, using the members of a
statistics class as subjects, the experiment might be conducted as
follows. On the first day of the
experiment the class is divided in half with one half of the class getting
coffee with caffeine and the other half getting coffee without caffeine. A measure of alertness is taken for each
individual, such as the number of yawns during the class period. On the second day the conditions are
reversed, that is, the individuals who received coffee with caffeine are now
given coffee without and vice-versa.
The distinguishing feature of crossing subjects with
treatments is that each subject will have more than one score. This feature is sometimes used in referring
to this class of designs as repeated measures designs. The effect also occurs within each subject,
thus these designs are sometimes referred to as within subjects
designs.
Crossing subjects with treatments has two advantages. One, they generally require fewer subjects,
because each subject is used a number of times in the experiment. Two, they are more likely to result in a
significant effect, given the effects are real. This is because the effects of individual differences between
subjects is partitioned out of the error term.
Crossing subjects with treatments also has
disadvantages. One, the experimenter
must be concerned about carry-over effects.
For example, individuals not used to caffeine may still feel the effects
of caffeine on the second day, when they did not receive the drug. Two, the first measurements taken may
influence the second. For example, if
the measurement of interest was score on a statistics test, taking the test
once may influence performance the second time the test is taken. Three, the assumptions necessary when more
than two treatment levels are employed in a crossing subjects with
treatments may be restrictive.
When a factor is a blocking factor, it is not possible to
cross that factor with subjects. It is
difficult to find subjects for a S X A design where A is
gender. I generally will take points
off if a student attempts such a design.
Nested
Factor B is said to be nested within factor if each
meaningful level of factor B occurs in conjunction with only one level
of A. This relationship is
symbolized a B(A), and is read as "B nested within
A". Note that B(A)
is considerably different from A(B). In the latter, each meaningful level of A would occur in
one and only one level of B.
These types of designs are also
designated as hierarchical designs in some textbooks.
A B(A) design occurs, for example, when the
first three levels of factor B (b1 ,b3, and b3)
appear only under level a1 of factor A and the next three levels of
B (b4 ,b5, and b6) appear only under level a2 of
factor A. Depending upon the labelling
scheme, b4 ,b5, and b6 may also be called b1
,b3, and b3, respectively. It is understood by the design designation that the b1
occurring under a1 is different from the b1 occurring
under a2.
Nested or hierarchical designs can appear because many
aspects of society are organized
hierarchically. For example
within the university, classes (sections) are nested within courses, courses
are nested within departments, departments within colleges, and colleges within
the university..
In experimental research it is also possible to nest
treatment conditions within other treatment conditions. For example, suppose a researcher was
interested in the effect of diet on health in hamsters. One factor (A) might be a high
cholesterol (a1) or low cholesterol (a2) diet. A second factor (B) might be type of
food, peanut butter (b1), cheese (b2), red meat (b3),
chicken (b4), fish (b5), or vegetables(b6). Because type of food may be categorized as
being either high or low in cholesterol, a B(A) experimental
design would result. Chicken, fish, and
vegetables would be relabelled as b1 ,b3, and b3,
respectively, but it would be clear from the experimental design specification
that peanut butter and chicken, cheese and fish, and red meat and vegetables,
were qualitatively different, even though they all share the same label.
While any factor may possibly be nested within any other
factor, the critical nesting relationship is with respect to subjects. If S is nested within some
combination of other factors, then each subjects appear under one, and only
one, combination of factors within which they are nested. These effects are often called the Between
Subjects effects. If S is
crossed with come combination of other factors, then each subject see all
combinations of factors with which they are crossed. These effects are referred to as Within Subjects effects.
As mentioned earlier subjects are necessarily nested within
blocking factors. Subjects are
necessarily nested within the effects of gender and current religious
preference, for example.
Treatment factors, however, may be nested or crossed with
subjects. The effect of caffeine on
alertness could be studied by dividing the subjects into two groups, with one
receiving a beverage with caffeine and one group not. This design would nest subjects with caffeine and be specified as
S(A), or simply A, as the S is often dropped when
the design is completely between subjects.
If subjects appeared under both caffeine conditions,
receiving caffeine on one day and no caffeine on the other, then subjects would
be crossed with caffeine. The design
would be specified as S X A.
In this case the S would remain in the design.
An Example
Design
A psychologist (McGuire, 1993) was interested in studying
adults' memory for medical information presented by a videotape. She included one-hundred and four
participants in which sixty-seven
ranged in age from 18 to 44 years and thirty seven ranged in age from 60
to 82 years. Participants were randomly
assigned to one of two conditions, either an organized presentation condition
or an unorganized presentation
condition. Following observation of the
videotape, each participant completed an initial recall sequence consisting of
free-recall and probed recall retrieval tasks. A probed recall is like a multiple-choice test and a free-recall
is like an essay test. Following a
one-week interval, participants completed the recall sequence again.
This experimental design provides four factors in addition
to subjects (S). The age factor
(A) has two level a1=young and a2=old and would
necessarily be a blocking factor. The
type of videotape factor (B) would be a treatment factor and would
consist of two levels b1=organized and b2=unorganized. The recall method factor (C) would be
a form of trials factor and would have two levels c1=free-recall and
c2=probed recall. The forth
factor (D) would be another trials factor where d1=immediate
and d2=one week delay.
Each level of B appears with each level of A,
thus A is crossed with B.
Since each subject appears in one and only one combination of A
and B, subjects are nested within A X B. That is, each subject is either young or old
and sees either an organized or unorganized videotape. The design notation thus far would be S
( A X B ).
Each type of recall (C) was done by each subject at
both immediate and delayed intervals (D). Thus subjects would be crossed with recall method and
interval. The complete design
specification would be S ( A X B ) X C X D. In words this design would be subjects
nested within A and B and crossed with C and D.
In preparation for entering the data into a data file, the
design could be viewed in a different perspective. Listing each subject as a row and each measure as a column, the
design would appear as follows:
Immediate Week Later
Age Videotape Subject Free Probed Free Probed
S1
Organized S2
Young ...
Unorganized S1
...
S1
Organized S2
Old ...
Unorganized S1
...
In this design,
two variables would be needed. One to classify each subjects as either young or
old, and one to document which type of videotape the subject saw. In addition to the classification variables,
each subject would require four variables to record the two types of measures
taken at the two different times.
A score taken
from the design presented above could be represented as Xabscd. For example, the immediate probed test score
taken from the third subject in the old group who viewed an organized videotape
would be X21312.
A
Second Example Design
The Lombard
effect is a phenomenon in which a speaker or singer involuntarily raises his or
her vocal intensity in the presence of high levels of sound. In a study of the Lombard effect in choral
singing (modified from Tonkinson, 1990), twenty-seven subjects, some
experienced choral singers and some not,
were asked to sing the national anthem along with a choir heard through
headphones. The performances were
recorded and vocal intensity readings from three selected places in the song
were obtained from a graphic level recorder chart. Each subject sang the song four times: with a none, or a soft, medium, or loud choir accompaniment. After some brief instructions to resist
increasing vocal intensity as the choir increased, each subject again sang the
national anthem four times with the four different accompaniments. The order of accompaniments was
counterbalanced over subjects.
In this design,
there would be four factors in addition to subjects. Subjects would be nested within experience level (A), with
a1=inexperienced and a2=experienced choral singers. This factor would be a blocking
factor. Subjects would be crossed with
instructions (B), where b1=no instructions and b2=resist
Lombard effect. In addition, subjects
would be crossed with accompaniment (C) and place in song (D).
The accompaniment factor would include four levels c1=soft, c2=medium,
c3=loud, and c4=none.
This factor would be considered a treatment factor. The place in song factor could be considered
a trial factor and would have three levels.
The
experimental design could be written as S ( A ) X B X C
X D. In words, subjects were
nested within experience level and crossed with instructions, accompaniment,
and place in song. In this design, one
variable would be needed for the classification of each subject and twenty-four
variables would be needed for each subject, one for each combination of
instructions, accompaniment, and place in song. The design could be written:
No
Instructions Resist
Lombard Effect
Soft Medium Loud
None Soft Medium
Loud None
Exp S 1 2 3 1 2 3
1 2 3 1 2 3 1 2 3
1 2 3 1 2 3 1 2 3
1 1
1 2
...
2 1
2 2
...
A
Third Example Design
From the
Springfield News-Leader, March 1, 1993:
Images of
beauty such at those shown by Sports Illustrated's annual swimsuit issue, are
harmful to the self-esteem of all women and contribute to the number of eating
disorder cases in the U. S., says a St. Louis professor who researches women's
health issues.
In a recent
study at Washington University, two groups of women - one with bulimia and one
without - watched videotapes of SI models in swimsuits.
Afterwards,
both groups reported a more negative self-image than they did before watching
the tape, describing themselves as "feeling fat and flabby" and
"feeling a great need to diet."
The experiment
described above has a number of inadequacies, the lack of control conditions
being the most obvious. The original
authors, unnamed in the article, may have designed a much better experiment
than is described in the popular press.
In any case, this experiment will now be expanded to illustrate a
complex experimental design.
The dependent
measure, apparently a rating of "feeling fat and flabby" and
"feeling a great need to diet", will be retained. In addition, two neutral questions will be
added, say "feeling anxious" and "feeling good about the
environment." These four
statements will be rated by all subjects, thus subjects will be crossed with
ratings. The first two statements deal
with body image and diet and the last two do not, thus they will form a factor
in the design (called D). Since
the statements within each of body image factor share no similarity across
levels of D, these statements (A) are nested within D. For
example, the rating of "feeling a great need to diet" and
"feeling good about the environment" share no qualitative
relationship. At this point the design
may be specified as S X A(D).
Suppose the
researcher runs the subjects in groups of six to conserve time and effort, thus
creating a groups (G) factor.
In addition to the two groups, with bulimia and without (B),
suppose the subjects viewed one of the following videotapes (V): SI models, Rosanne Barr, or a show about the
seals of the great northwest. Assuming
that all the subjects in each level of group either had bulimia or did not,
then the design could be specified as
S(G(B
X V)).
The factor B
is crossed with V because each level of B appears with each level
of V. That is, subjects with and
without bulimia viewed all three videotapes. Because each group viewed only a single videotape and was composed
of subjects either with bulimia or without, the groups factor is nested within
the cross of B and V.
Because subjects appeared in only one group, subjects are nested within
groups.
Combining the
between subjects effects, S(G(B X V)), and the
within subjects effects, A(D), yields the complete design
specification
S(G(B
X V)) X A(D).
Determining
the Number of Subjects and Measures per Subject
It is important
to be able to determine the number of subjects and the number of measures per
subject for practical reasons, namely, is the experiment feasible? After listening to a student propose an
experiment and a little figuring, I remarked "according to my
calculations, you should be able to complete the experiment sometime near the
middle of the next century." If an
experimenter is limited in the time a subject is available, then the number of
measures per subject is another important consideration.
To determine
the number of subjects, multiply the number of levels of the between subjects
factor together. In the previous
example, S = 6 because the subjects were run in groups of six. Let G=4, or there be four groups of six each
of combinations of bulimia and videotape.
Since there were two levels of bulimia, B=2, and three levels of
videotape, V=3. Since S(G(B
X V)), then the total number of subjects needed would be S * G * B * V
or 6*4*2*3 or 144. Since half of the
subjects must have bulimia, the question of whether or not 72 subjects with
bulimia are available must be asked before the experiment proceeds.
To find the
number of measures per subject, multiply the number of levels of the within
subjects factors together. In the
previous example A(D), where A=2 and D=2, there would be A * D or
2 * 2 or 4 measures per subject.
Setting
up the Data Matrix
Columns |
1
2 ... C 1 Rows 2 . R |
A few rules
simplify setting up the data matrix.
First, each subject appears on a single row of the data matrix. Second, each measure or combination of
within subjects factors appears in a column of data. Third, each subject must be identified as to the combination of
between subjects factors which he or she appears.
1 1 1 3 5 4 3 |
1 1 1 2 5 5 3 1 1 1 5 5 5 4 1 1 1 3 2 1 3 1 1 1 2 5 3 1 1 1 1 3 5 4 3 2 1 1 5 4 5 5 ... 4 2 3 3 5 5 4 |
In the previous
example, since there would be 144 subjects in the experiment, there would be
144 rows of data. Each subject would be
identified as to the level of G,
B, and V to which she belonged.
For example, a subject who appeared under g3 of b1
and v4 would be labelled as 3 1 4.
Since there are four measures per subject, these would appear as columns
in addition to the identifiers. An
example data matrix might appear as follows.
In this example, the level of G is in the first column, B in the second,
and V in the third. The four
combinations of within subjects factors appear next as ad11 ad12
ad12 ad22.
A Note
of Caution
It is fairly
easy to design complex experiments.
Running the experiments and interpreting the results are a different
matter. Many complex experiments are
never completed because of such difficulties.
This is from personal experience.
Chapter
9
One Between
Group ANOVA
Why Multiple Comparisons Using
t-tests is NOT the Analysis of Choice
Group
Therapy Method _ SX SX˛ |
1
Reality 20.53 3.45
11.9025 2
Behavior 16.32 2.98
8.8804 3
Psychoanalysis 10.39 5.89
35.7604 4
Gestalt 24.65 7.56
57.1536 5
Control 10.56 5.75
33.0625 |
Suppose a researcher has performed a study of
the effectiveness of various methods of individual therapy. The methods used were: Reality Therapy, Behavior Therapy,
Psychoanalysis, Gestalt Therapy, and, of course, a control group. Twenty patients were randomly assigned to
each group. At the conclusion of the
study, changes in self-concept were found for each patient. The purpose of the study was to determine if
one method was more or less effective than the other methods.
At the conclusion of the experiment the
researcher organizes the collected data in the following manner:
The researcher wishes to compare the means of
the groups with each other to decide about the effectiveness of the
therapy.
One method of performing this analysis is by
doing all possible t-tests, called multiple t-tests. That is, Reality Therapy is first compared with Behavior Therapy,
then Psychoanalysis, then Gestalt Therapy, then the Control Group. Behavior Therapy is then individually
compared with the last three groups, and so on. Using this procedure there would be ten different t-tests
performed. Therein lies the difficulty
with multiple t-tests.
First, because the number of t-tests
increases geometrically as a function of the number of groups, analysis becomes
cognitively difficult somewhere in the neighborhood of seven different
tests. An analysis of variance
organizes and directs the analysis, allowing easier interpretation of results.
Secondly, by doing a greater number of
analyses the probability of committing at least one type I error somewhere in
the analysis greatly increases. The
probability of committing at least one type I error in an analysis is called
the experiment-wise error rate. The researcher may desire to perform a fewer
number of hypothesis tests in order to reduce the experiment-wise error
rate. The ANOVA procedure performs this
function.
The Bottom Line - Results and
Interpretation of ANOVA
Results of an ANOVA are usually presented in
an ANOVA table. This table
contains columns labelled
"Source", "SS or Sum of Squares", "df - for degrees of
freedom", "MS - for mean square", "F or F-ratio", and
"p, prob, probability, sig., or
sig. of F". The only columns that
are critical for interpretation are the first and the last, the others are used
mainly for intermediate computational purposes.
Source SS df MS F sig of F |
BETWEEN
5212.960 4 1303.240 4.354 .0108 WITHIN
5986.400 20 299.320 TOTAL
11199.360 24 |
An example of an ANOVA table appears below:
The row labelled "BETWEEN" under
"Source", having a probability value associated with it, is the only
one of any any great importance at this time.
The other rows are used mainly for computational purposes. The researcher then would most probably
first look at the value ".0108" located under the "sig of
F" column.
Of all the information presented in the ANOVA
table, the major interest of the researcher will most likely be focused on the
value located in the "sig of F." column. If the number (or numbers) found in this column is (are) less
than the critical value (α) set by the experimenter, then the effect is said to be
significant. Since this value is
usually set at .05, any value less than this will result in significant
effects, while any value greater than this value will result in nonsignificant
effects.
If the effects are found to be significant
using the above procedure, it implies that the means differ more than would be
expected by chance alone. In terms of
the above experiment, it would mean that the treatments were not equally
effective. This table does not tell the
researcher anything about what the effects were, just that there most likely
were real effects.
If the effects are found to be
nonsignificant, then the differences between the means are not great enough to
allow the researcher to say that they are different. In that case no further interpretation is attempted.
When the effects are significant, the means
must then be examined in order to determine the nature of the effects. There are procedures called "post-hoc
tests" to assist the researcher in this task, but often the analysis is
fairly evident simply by looking at the size of the various means. For example, in the preceding analysis
Gestalt and Reality Therapy were the most effective in terms of mean
improvement.
|
In the case of significant effects, a
graphical presentation of the means can sometimes assist in analysis. For example, in the preceding analysis, the
graph of mean values would appear as follows:
HYPOTHESIS TESTING THEORY UNDERLYING
ANOVA
The Sampling
Distribution Reviewed
In order to explain why the above procedure
may be used to simultaneously analyze a number of means, the following presents
the theory on ANOVA in relation to the hypothesis testing approach discussed in
earlier chapters.
First, a review of the sampling distribution
is necessary. If you have difficulty
with this summary, please go back and read the more detailed chapter on the sampling
distribution.
A sample is a finite number (N) of
scores. Sample statistics are numbers
which describe the sample. Example
statistics are the mean (_), mode (Mo), median (Md), and
standard deviation (sX).
Probability models exist in a theoretical
world where complete information is unavailable. As such, they can never be known except in the mind of the
mathematical statistician. If an
infinite number of infinitely precise scores were taken, the resulting distribution
would be a probability model of the population. Population models are characterized by parameters. Two common parameters are µX and σX.
Sample statistics are used as estimators of
the corresponding parameters in the population model. For example, the mean and standard deviation of the sample are
used as estimates of the corresponding population parameters µX and σX.
The sampling distribution is a distribution
of a sample statistic. It is a model of
a distribution of scores, like the population distribution, except that the
scores are not raw scores, but statistics.
It is a thought experiment; "what would the world be like if a
person repeatedly took samples of size N from the population distribution and
computed a particular statistic each time?" The resulting distribution of statistics is called the sampling
distribution of that statistic.
The sampling distribution of the mean is a special case of a sampling
distribution. It is a distribution of
sample means, described with the parameters µ_ and σ_. These parameters are closely related to the
parameters of the population distribution, the relationship being described by
the CENTRAL LIMIT THEOREM. The
CENTRAL LIMIT THEOREM essentially states that the mean of the sampling
distribution of the mean (µ_) equals the mean of the population (µX)
and that the standard error of the mean (σ_) equals the
standard deviation of the population (σX) divided by the
square root of N. These relationships
may be summarized as follows:
|
Two Ways of Estimating the
Population Parameter σX˛
When the data have been collected from more
than one sample, there exists two independent methods of estimating the
population parameter σX˛, called respectively the between and the
within method. The collected data are
usually first described with sample statistics as demonstrated in the following
example:
Group Therapy
Method _ SX SX˛ |
1 Reality 20.53 3.45 11.9025 2 Behavior 16.32 2.98 8.8804 3 Psychoanalysis 10.39
5.89 35.7604 4 Gestalt 24.65 7.56 57.1536 5 Control 10.56 5.75 33.0625 Mean 16.49 29.3519 Variance 38.83 387.8340 |
THE WITHIN METHOD
Since each of the sample variances may be
considered an independent estimate of the parameter σX˛, finding the mean of the variances provides a method of combining the
separate estimates of σX˛ into a single value. The resulting statistic is called the MEAN
SQUARES WITHIN, often represented by MSW. It is called the within method because it computes the estimate
by combining the variances within each sample. In the above example, the Mean Squares Within would be equal to
29.3519.
THE BETWEEN METHOD
The parameter σX˛ may also be
estimated by comparing the means of the different samples, but the logic is
slightly less straightforward and employs both the concept of the sampling
distribution and the Central Limit Theorem.
Sampling Distribution Actual Data |
_ _ _ _ _ _ . _ . _ . Mean µ_ __ Variance σ_˛ s_˛ |
First, the standard error of the mean squared
(σ_˛) is the population variance of a distribution of sample means. In real life in the situation where there is
more than one sample, the variance of the sample means may be used as an
estimate of the standard error of the mean squared (σ_˛). This is analogous to the
situation where the variance of the sample (sX˛) is used as an
estimate of σ_˛. The relationship is
demonstrated below:
In this case the Sampling Distribution
consists of an infinite number of means and the real life data consists of A
(in this case 5) means. The computed
statistic s_˛ is thus an estimate of the theoretical parameter σ_˛.
The relationship expressed in the Central Limit Theorem may now be used
to obtain an estimate of σ˛.
|
Thus
the variance of the population may be found by multiplying the standard error
of the mean squared (σ_˛) by N, the size of each sample.
Since the variance of the means, s_˛,
is an estimate of the standard error of the mean squared, σ_˛, the variance of the population, σX˛, may be
estimated by multiplying the size of each sample, N, by the variance of the
means. This value is called the Mean
Squares Between and is often symbolized by MSB. The computational procedure for MSB
is presented below:
MSB = N * s_˛
MSB = N * s_˛ |
MSB
= 20 * 38.83 MSB
= 776.60 |
The expressed value is called the Mean
Squares Between because it uses the variance between the samples, that
is the sample means, to compute the estimate.
Using the above procedure on the example data yields:
At this point it has been established that
there are two methods of estimating σX˛, Mean Squares
Within and Mean Squares Between. It could
also be demonstrated that these estimates are independent. Because of this independence, when both are
computed using the same data, in almost all cases different values will result. For example, in the presented data MSW=29.3519
while MSB=776.60. This
difference provides the theoretical background for the F-ratio and ANOVA.
The F-ratio and F-distribution
A new statistic, called the F-ratio is computed by dividing the MSB
by MSW. This is illustrated
below:
Fobs = MSB / MSW |
Using the example data described earlier the
computed F-ratio becomes:
Fobs = MSB
/ MSW
Fobs = 776.60 /
29.3519
Fobs = 26.4582
The F-ratio can be thought of as a measure of
how different the means are relative to the variability within each
sample. The larger this value, the
greater the likelihood that the differences between the means are due to
something other than chance alone, namely real effects. The size of the F-ratio necessary to make a
decision about the reality of effects is the next topic of discussion.
If the difference between the means means is
due only to chance, that is, there are no real effects, then the expected value
of the F-ratio would be one (1.00).
This is true because both the numerator and the denominator of the
F-ratio are estimates of the same parameter, σX˛. Seldom will the F-ratio be exactly equal to
1.00, however, because the numerator and the denominator are estimates rather
than exact, known values. Therefore,
when there are no effects the F-ratio will sometimes be greater than one, and
other times less than one.
To review, the basic procedure used in
hypothesis testing is that a model is created in which the experiment is
repeated an infinite number of times when there are no effects. A sampling distribution of a statistic is
used as the model of what the world would look like if there were no
effects. The results of the experiment,
a statistic, is compared with what would be expected given the model of no
effects was true. If the computed
statistic is unlikely given the model, then the model is rejected along with
the hypothesis that there were no effects.
In an ANOVA the F-ratio is the statistic used
to test the hypothesis that the effects are real, in other words, that the
means are significantly different from one another. Before the details of the hypothesis test may be presented, the
sampling distribution of the F-ratio must be discussed.
If
the experiment were repeated an infinite number of times, each time
computing the F-ratio, and there were no effects, the resulting distribution
could be described by the F-distribution.
The F-distribution is a theoretical probability distribution
characterized by two parameters, df1 and df2, both of
which affect the shape of the distribution.
Since the F-ratio must always be positive, the F-distribution is
non-symmetrical, skewed in the positive direction.
Two examples of an F-distribution are
presented below; the first with df1=1 and df2=5, and the
second with df1=10 and df2=25.
|
|
The
F-distribution has a special relationship to the t-distribution described
earlier. When df1=1, the
F-distribution is equal to the t-distribution squared (F=t˛). Thus the t-test and the ANOVA will always
return the same decision when there are two groups. That is, the t-test is a special case of ANOVA.
Non-significant and Significant
F-ratios
Theoretically, when there are no real effects, the F-distribution is an
accurate model of the distribution of F-ratios. The F-distribution will have the parameters df1=a-1
(where a-1 is the number of different groups minus one) and df2=a(N-1),
where a is the number of groups and N is the number in each group. In this case
an assumption is made that sample size is equal for each group. For example, if
five groups of five subjects each were run in an experiment and there were no
effects, then the F-ratios would be distributed with df1=k-1=5-1=4
and df2=k(n-1)=5(5-1)=5*4=20.
A visual representation of the preceding appears as follows:
|
|
The F-ratio in the above which cuts off
various proportions of the distributions may be computed for different values
of α. These F-ratios are called Fcrit
values. In the above example the Fcrit
value for α=.25 is 1.46, for α=.10 results in a value of 2.25, for α=.05 the value is 2.87, and for α=.01 the value is 4.43. These values are illustrated in the figure
below:
When there are real effects, that is, the
means of the groups are different due to something other than chance, then the
F-distribution no longer describes the distribution of F-ratios. In almost all cases the observed F-ratio
will be larger than would be expected when there were no effects. The rationale for this situation is
presented below.
First, an assumption is made that any effects
are an additive transformation of the score.
That is, the scores for each group can be modelled as a constant ( aa
- the effect) plus error (eae).
The scores appear as follows:
Xae = aa + eae
where X is the score, aa is the
treatment effect, and eae is the error. The eea, or error, is different for each subject,
while aa is constant within a given group.
As described in the chapter on
transformations, an additive transformation changes the mean, but not the
standard deviation or the variance.
Because the variance of each group is not changed by the nature of the
effects, the Mean Square Within, as the mean of the variances, is not
affected. The Mean Square Between, as N
time the variance of the means, will in most cases become larger because the
variance of the means will most likely become larger.
Imagine three individuals taking a test. An instructor first finds the variance of
the three score. He or she then adds
five points to one random individual and subtracts five from another random
individual. In most cases the variance
of the three test score will increase, although it is possible that the
variance could decrease if the points were added to the individual with the
No effects Real Effects |
Group Mean Variance
Group Mean Variance 1 µ σ˛ 1 µ + a1
σ˛ 2 µ σ˛ 2 µ + a2
σ˛ 3 µ σ˛ 3 µ + a3
σ˛ 4 µ σ˛ 4 µ + a4
σ˛ 5 µ σ˛ 5 µ + a5
σ˛ Mean µ σ˛ µ σ˛ Variance σ˛/N >σ˛/N |
lowest score and subtracted from the
individual with the highest score. If
the constant added and subtracted was 30 rather than 5, then the variance would
almost certainly be increased. Thus,
the greater the size of the constant, the greater the likelihood of a larger
increase in the variance.
With respect to the sampling distribution, the model differs depending
upon whether or not there are effects.
The difference is presented below:
Since
the MSB usually increases and MSW remains the same, the
F-ratio (F=MSB/MSW) will most likely increase. Thus, if there are real effects, then the
F-ratio obtained from the experiment will most likely be larger than the
critical level from the F-distribution.
The greater the size of the effects, the larger the obtained F-ratio is
likely to become.
Thus, when there are no effects, the obtained
F-ratio will be distributed as an F-distribution which may be specified. If effects exist, then the obtained F-ratio
will most likely become larger. By
comparing the obtained F-ratio with that predicted by the model of no effects,
an hypothesis test may be performed to decide on the reality of effects. If the obtained F-ratio is greater than the
critical F-ratio, then the decision will be that the effects are real. If not, then no decision about the reality
of effects can be made.
Similarity of ANOVA and t-test
When the number of groups (A) equals two (2),
an ANOVA and t-test will give similar results, with tCRIT˛=FCRIT
and tOBS˛=FOBS.
This equality is demonstrated in the example below:
Given the following numbers for two
groups:
Mean
Variance
Group 1 - 12 23 14 21 19 23 26 11 16
18.33 28.50
Group 2 - 10 17 20 14 23 11 14 15 19
15.89 18.11
Computing the t-test
s_1-_2
= Ö (s1˛ + s2˛)/ 9 = Ö (28.50 + 18.11)/9 = Ö 5.18 =
2.28
tOBS
= ( _1-_2 ) / s_1-_2 = 18.33 - 15.89 / 2.28 =
1.07
t(df=16) = 2.12 for α=.05 and two-tailed test
Computing the ANOVA
MSBETWEEN = N * s_˛ = 9 * 2.9768 = 26.7912
MSWITHIN = Mean of
the Variances = ( 28.50 + 18.11 ) / 2 = 23.305
FOBS
= MSBETWEEN/MSWITHIN = 1.1495
F(1,16) = 4.41 for α=.05 - two-tailed test is assumed
Comparing the results
tOBS˛
= 1.1449 FOBS = 1.1449
t(16)˛ = 4.49 F(1,16) = 4.49
The differences between the predicted and
observed results can be attributed to rounding error (close enough for
government work).
Because the t-test is a special case of the
ANOVA and will always yield similar results, most researchers perform the ANOVA
because the technique is much more powerful in complex experimental designs.
EXAMPLE OF A NON-SIGNIFICANT ONE-WAY
ANOVA
MEAN VARIANCE |
7 7 5 4 2 7 5 4 1 7 5 6 7 6 3 5 2 5 1
4 4.65 4.03 6 9 3 6 9 4 9 8 9 3 4 4 7 2 2 7 7 7 9
3 5.90 6.52 5 5 2 5 6 2 3 3 6 8 2 1 1 2 5 7 9 6 5
7 4.50 5.63 4 1 4 8 9 5 2 8 6 8 2 9 6 6 7 8 4 3 1
4 5.25 6.93 3 6 1 2 3 5 8 4 1 5 4 5 6 9 4 2 4 8 9
3 4.60 6.04 Computing
the ANOVA MSBETWEEN = N * s_˛
= 20 * .351 = 7.015 MSWITHIN = Mean of the Variances = 5.83 FOBS = MSBETWEEN/MSWITHIN
= 1.20 F(4,95) = 2.53 for α=.05 - non-directional test is
assumed |
Given the following data for five groups, perform an ANOVA:
Since the FCRIT is greater than
the FOBS, the means are not significantly different and no effects
are said to be discovered.
EXAMPLE OF A SIGNIFICANT ONE-WAY
ANOVA
Given the following data for five groups,
perform an ANOVA. Note that the numbers
are similar to the previous example except that one has been subtracted from
all scores in Group 3 and one has been added to all scores in Group 4.
In this case the FOBS is greater
than FCRIT, thus the means are significantly different and we decide
that the effects are real.
1 23 |
1 31 1 25 1 29 1 30 1 28 1 31 1 31 1 33 2 32 2 28 2 36 2 34 2 41 2 35 2 32 2 28 2 31 |
USING MANOVA
MEAN VARIANCE |
7 7 5 4 2 7 5 4 1 7 5 6 7 6 3 5
2 5 1 4 4.65 4.03 6 9 3 6 9 4 9 8 9 3 4 4 7 2 2 7
7 7 9 3 5.90 6.52 4 4 1 4 5 1 2 2 5 7 1 0 0 1 4 6
8 5 4 6 3.50 5.63 5 2 5 9 10 6 3 9 7 9 3 10 7 7 8 9 5 4 2
4 6.25 6.93 3 6 1 2 3 5 8 4 1 5 4 5 6 9 4 2
4 8 9 3 4.60 6.04
Computing
the ANOVA MSBETWEEN = N * s_˛
= 20 * 1.226 = 24.515 MSWITHIN = Mean of the Variances = 5.83 FOBS = MSBETWEEN/MSWITHIN
= 4.20 F(4,95) = 2.53 for α=.05 - two-tailed test is assumed |
While an single factor between groups ANOVA
may be done using the MEANS command in SPSS, the MANOVA command is a general purpose
command which allows the statistician to do almost any type of multifactor
univariate or multivariate ANOVA.
The Data
The data is entered into a data file
containing two columns. One column
contains the level of the factor to which the observation belongs and the
second the score for the dependent variable.
A third column containing the observation number, in the example a
number from one to nine, is optional.
As in all SPSS data files, the number of rows in the data file
corresponds to the number of subjects and each variable is lined up neatly in
each row. In the example data file
presented to the right, there are two groups of nine each. The level of the independent variable is
given in the first column of the data file, a space is entered, and the
dependent variable is entered in columns 3 and 4.
RUN NAME EXAMPLE FOR ANOVA BOOK - DESIGN A. |
DATA LIST FILE='DESIGNA
DATA A' /1 A 1 X 3‑4. VALUE LABELS A 1 'BLUE BOOK' 2
'COMPUTER' LIST. MANOVA X by A(1,2) /PRINT CELLINFO(MEANS) /DESIGN. |
The RUN NAME command of the example program
gives a general description of the purpose of the program. The second command reads in the data
file. Note that the group factor is
called "A" and the dependent variable is called "X". The value label command then describes the
different levels of the group variable.
The LIST command gives a description of the data as the computer
understands it.
The MANOVA command is followed by the name of
the dependent variable, here X, and the keyword BY . The factor name "A" is then entered, followed by the
the beginning and ending levels of that factor. In this case there were only two levels, defined by a beginning
value of "1", and an ending value of "2". The second line on the command is preceded
by a slash "/'" and then the subcommand PRINT=CELLINFO(MEANS). This command will print the means of the
respective groups. The last subcommand,
"/DESIGN" , is optional at this point, but not including it will
generate a WARNING. Nothing is altered
with the WARNING, but it is not neat.
Example Output
The output produced by the example MANOVA
command is presented on the next page.
The default error term in MANOVA has been changed
from WITHIN CELLS to WITHIN+RESIDUAL.
Note that these are the same for all full factorial designs. |
* *
* * * * A n a l y s i s o f V a r i a n c e * * * * * 18 cases accepted. 0 cases rejected because of out‑of‑range
factor values. 0 cases rejected because of missing
data. 2 non‑empty cells. 1 design will be processed. Cell Means and Standard Deviations Variable .. X FACTOR CODE Mean Std. Dev. N A 1
29.000 3.202 9 A 2
33.000 4.093 9 For entire sample 31.000
4.116 18 * *
* * * * A n a l y s i s o f V a r i a n c e ‑‑
design Tests of Significance for X using UNIQUE
sums of squares Source of Variation SS DF MS F
Sig of F WITHIN CELLS 216.00
16 13.50 A 72.00
1 72.00 5.33
.035 (Model) 72.00
1 72.00 5.33
.035 (Total) 288.00
17 16.94 R‑Squared = .250 Adjusted R‑Squared = .203 |
Note that the F-ratio was significant with a
value of .03. The means for the
two groups were 29 and 33 respectively.
Dot Notation
In order to simplify the notational system
involving the summation sign, a notational system call dot notation has
been introduced. Dot notation places a period in place of subscript to mean
summation. For example:
|
The symbol X. means that the variable X has
been summed over whatever counter variable was used as a subscript. In a like manner, if a bar is placed over a
variable, it refers to a mean over the dotted counter variable(s). For example:
|
where _. means the same thing as _. (The dot
notation does become a bit tricky where real periods are involved.)
The real advantage is apparent when two or
more subscripts are used. For example
|
and
|
or if a=1 then
|
Using the same notational system, means may
be given. For example:
|
and
|
or if a=1 then
|
The difference between _a. and _1.
is that the second is a special case of the first when a=1.
For example, if A=3, B=4 and X11=5,
X12=8, X13=6, X14=9, X21=7, X22=10,
X23=5, X24=9, X31=6, X32=4, X33=7,
X34=3, then
X.. = 79,
_.. = 79/12 = 6.5833,
X1. = 28,
_1. = 28/4 = 7.
POST-HOC Tests of Significance
If the results of the ANOVA are significant,
it indicates that there are real effects between the means of the groups. The nature of the effects are not specified
by the ANOVA. For example, an effect
could be significant because the mean of group three was larger than the means
of the rest of the groups. In another
case, the means of groups one, four and five might be significantly smaller
than the means of groups two, three, and six.
Often the pattern of results can be determined by a close examination of
the means. In other instances, the
reason for the significant differences is not apparent. To assist the statistician in interpreting
effects in significant ANOVAs, post-hoc tests of significance were developed.
A post-hoc (after the fact) test of significance
is employed only if the results of the overall significance test are
significant. A post-hoc test is
basically a multiple t-test procedure with some attempt to control for the
increase in the experiment wide error rate when doing multiple significance
tests. A number of different procedures
are available to perform post-hoc tests differing in the means of control of
the increase in error rates . The
different procedures include Duncan’s Multiple Range test, the Newman-Keuls
procedure, and a procedure developed by Sheffe’. The interested reader is referred to Winer (1971) or Hays (1981)
for a thorough discussion of these methods.
My personal feeling is that post-hoc tests
are not all that useful. Most often,
the reason for the significant results is obvious from a close observation of
the means of the groups. Better
procedures are available using pre-planned contrasts to test patterns of
results. The use of pre-planned
contrasts requires that the statistician have the type of comparisons in mind
before doing the analysis. This is the
difference between data-driven (post-hoc) and theory-driven (pre-planned)
analysis. If a choice is possible, the
recommendation is for theory-driven analysis.
Chapter
10
TWO Between
Groups ANOVA (A x B)
The Design
In this design there are two independent
factors, A and B, crossed with each other. That is, each level of A appears in
combination with each level of B.
Subjects are nested within the combined levels of A and B
such that the full design would be written as S ( A X B ).
Because the Subjects (S) term is confounded with the error term,
it is dropped from the description of the design.
A B
X |
1 1
23 1 1
32 1 1
25 1 2
29 1 2
30 1 2
34 1 3
31 1 3
36 1 3
33 2 1
32 2 1
26 2 1
26 2 2
34 2 2
41 2 2
35 2 3
24 2 3
27 2 3
31 |
Suppose a statistics teacher gave an essay
final to his class. He randomly divides
the classes in half such that half the class writes the final with a blue-book
and half with notebook computers. In
addition the students are partitioned into three groups, no typing ability,
some typing ability, and highly skilled at typing. Answers written in blue-books will be transcribed to word
processors and scoring will be done blindly.
Not with a blindfold, but the instructor will not know the method or
skill level of the student when scoring the final. The dependent measure will be the score on the essay part of the
final exam.
The first factor (A) will be called Method and will have two levels, a1=blue-book
and a2 = computer. The second factor (B) will be
designated as Ability and will have three levels: b1=none, b2=some, and b3=lots. Because each level of A appears with
each level of B, A is said to be crossed with B (AXB). Since different subjects will appear in each
combination of A and B, subjects are nested within AXB. Each subject will be measured a single
time. Any effects discovered will
necessarily be between subjects or groups and hence the designation
"between groups" designs.
The Data
The data file for the A X B
design is similar to the data file for design A with the addition of the
second descriptive variable, B, for each subject. In the case of the example data, the A
factor has two levels while the B factor has three. The X variable is the score on the final
exam. The example data file appears in
the text box on the right.
Example SPSS Program using MANOVA
The SPSS commands necessary to do the
analysis for an A X B design are given in the text box below.
RUN NAME EXAMPLES FOR ANOVA BOOK ‑ A X B DESIGNS |
DATA LIST FILE='AXB
DATA A' / A 3 B 6 X 15‑19. VARIABLE LABELS A 'METHOD OF WRITING EXAM' B 'KEYBOARD EXPERIENCE' X 'SCORE ON FINAL EXAM'. VALUE LABELS A 1 'BLUE‑BOOK' 2 'COMPUTER'/ B 1 'NONE' 2 'SOME' 3 'LOTS'. LIST. MANOVA X BY A (1,2) B
(1,3) /PRINT CELLINFO(MEANS) /DESIGN. |
Note that the DATA LIST command must read in
a variable to code each factor. In this
case the variables were named A and B to correspond with the factor names,
although in most real-life situations more descriptive names will be used. The addition of the optional VALUE LABELS
command will label the output from the PRINT CELLINFO(MEANS) command, making
the output easier to interpret.
The MANOVA command is followed by the name of
the dependent variable, in this case X and then the variable names of the
factors. As in the previous MANOVA
commands, the factor names are each followed by the beginning and ending levels
enclosed in parentheses. In this case
the A factor has two levels beginning at level 1 and the B factor
has three. The PRINT CELLINFO(MEANS)
command is optional, but usually included because the means are the central
focus of the analysis. The DESIGN
command is optional, but excluding it will generate a warning when the program
is run, so it is usually included for the sake of neatness.
All of the analyses on the following pages
were generated from the SPSS program presented above. The program will not be included as part of the interpretation of
the output.
Interpretation of Output
Cell Means and Standard Deviations |
Variable .. X
SCORE ON FINAL EXAM
FACTOR CODE Mean Std. Dev. N 95 percent Conf. Interval
A BLUE‑BOO
B NONE 26.717 4.822 3 14.739 38.695
B SOME 30.923 2.418 3 24.916 36.930
B LOTS 33.277 2.157 3 27.919 38.634
A COMPUTER
B NONE 28.057 3.285 3 19.896 36.217
B SOME 36.627 3.413 3 28.148 45.106
B LOTS 27.080 3.673 3
17.955 36.205 For
entire sample 30.447 4.675 18
28.122 32.771 |
The interpretation of the output from the
MANOVA command will focus on two parts:
the table of means and the ANOVA summary table. The table of means is the primary focus of
the analysis while the summary table directs attention to the interesting or
statistically significant portions of the table of means.
A table of means generated using the example
data and the PRINT CELLINFO(MEANS) subcommand in MANOVA is presented below:
Often the means are organized and presented
in a slightly different manner than the form of the output from the MANOVA
command. The table of means may be
rearranged and presented as follows:
|
b1
b2 b3 ┌────────┬────────┬────────┐ a1 │ 26.72 │ 30.92 │ 33.28 │ 30.31 ├────────┼────────┼────────┤ a2 │ 28.06 │ 36.62 │ 27.08 │ 30.58 └────────┴────────┴────────┘ 27.39
33.78 30.18 30.447 |
The means inside the boxes are called cell
means, the means in the margins are called marginal means, and the
number on the bottom right-hand corner is called the grand mean. An analysis of these means reveals that there is very little difference between
the marginal means for the different levels of A across the levels of B
(30.31 vs. 30.58). The marginal means
of B over levels of A are different (27.39 vs. 33.78 vs. 30.18)
with the mean for b2 being the highest. The cell means show an increasing pattern for levels of B
at a1 (26.72 vs. 30.92 vs. 33.28) and a different pattern for levels
of B at a2 (28.06 vs. 36.62 vs. 27.08).
Graphs of Means
|
|
Graphs of means are often used to present
information in a manner which is easier to comprehend than the tables of
means. One factor is selected for
presentation as the X-axis and its levels are marked on that axis. Separate lines are drawn the height of the
mean for each level of the second factor.
In the following graph, the B, or keyboard experience, factor was
selected for the X-axis and the A, or method, factor was selected for
the different lines. Presenting the
information in an opposite fashion would be equally correct, although some
graphs are more easily understood than others, depending upon the values for
the means and the number of levels of each factor. The second possible graph is presented below. It is recommended that if there is any doubt
that both versions of the graphs be attempted and the one which best
illustrates the data be selected for inclusion into the statistical
report. It is hopefully obvious that
the graph with B on the X-axis is easier to understand than the one with A on
the X-axis.
Because the interpretation of the graph of
the interaction depends upon the results of the analysis, the ANOVA summary
table will now be presented. Following
this, the graph of the interaction will be reanalyzed.
The ANOVA Summary Table
The results of the A X B ANOVA
are presented in the ANOVA summary table by MANOVA. An example of this table is presented below:
Tests of Significance for X using UNIQUE sums of squares |
Source SS DF MS F
Sig of F WITHIN CELLS
139.36 12 11.61 A .36 1
.36 .03 .863 B 123.08 2
61.54 5.30 .022 A BY B 108.73 2 54.36 4.68 .031 |
The items of primary interest in this table
are the effects listed under the "Source" column and the values under
the "Sig of F" column. As in
the previous hypothesis test, if the value of "Sig of F" is less than
the value of α as set by the experimenter, then that effect
is significant. If α=.05, then the B main effect and the A BY B
interaction would be significant in this table.
Main Effects
Main
effects are differences in means over levels of one factor collapsed over
levels of the other factor. This is
actually much easier than it sounds.
For example, the main effect of A is simply the difference
between the means of final exam score for the two levels of Method, ignoring or
collapsing over experience. As seen in
the second method of presenting a table of means, the main effect of A
is whether the two marginal means associated with the A factor are
different. In the example case these means
were 30.31 and 30.58 and the differences between these means was not
statistically significant.
As can be seen from the summary table, the
main effect of B is significant.
This effect refers to the differences between the three marginal means
associated with factor B. In
this case the values for these means were 27.39, 33.78, and 30.18 and the
differences between them may be attributed to a real effect.
Simple Main Effects
A simple main effect is a main effect of one
factor at a given level of a second factor. In the example data it would be possible to talk about the simple
main effect of B at a1.
That effect would be the difference between the three cell means at
level a1 (26.72, 30.92, and 33.28).
One could also talk about the simple main effect of A at b3
(33.28 and 27.08). Simple main effects
are not directly tested in design A X B, however they are
necessary to understand an interaction.
Interaction Effects
An interaction effect is a change in the
simple main effect of one variable over levels of the second. An AB or A BY B interaction is a change
in the simple main effect of B over levels of A or the change in
the simple main effect of A over levels of B. In either case the cell means cannot be
modelled simply by knowing the size of the main effects. An additional set of parameters must be used
to explain the differences between the cell means. These parameters are collectively called an interaction.
|
The change in the simple main effect of one
variable over levels of the other is most easily seen in the graph of the
interaction. If the lines describing
the simple main effects are not parallel, then a possibility of an interaction
exists. As can be seen from the graph
of the example data, the possibility of a significant interaction exists
because the lines are not parallel. The
presence of an interaction was confirmed by the significant interaction in the
summary table.
The following graph overlays the main effect
of B on the graph of the interaction.
Two things can be observed from this presentation. The first is that the main effect of B is
possibly significant, because the means are different heights. Second, the interaction is possibly significant
because the simple main effects of B at a1 and a2 are
different from the main effect of B.
One method of understanding how main effects
and interactions work is to observe a wide variety of data and data
analysis. With three effects, A,
B, and AB, which may or may not be significant there are eight
possible combinations of effects. All
eight are presented on the following pages.
Example Data Sets, Means, and
Summary Tables
No Significant Effects
A B X
1 1 22
|
1 1 24
1 1 25
1 2 26
1 2 29
1 2 23
1 3 21
1 3 25
1 3 22
2 1 21
2 1 26
2 1 25
2 2 24
2 2 20
2 2 24
2 3 23
2 3 26
2 3 20
Cell Means and Standard Deviations
Variable .. X
SCORE ON FINAL EXAM
FACTOR
CODE Mean Std. Dev. N 95 percent
Conf. Interval
A BLUE‑BOO
B
NONE 23.667 1.528 3 19.872 27.461
B
SOME
26.000 3.000 3 18.548 33.452
B
LOTS
22.667 2.082 3 17.495 27.838
A COMPUTER
B
NONE
24.000 2.646 3 17.428 30.572
B
SOME
22.667 2.309 3 16.930 28.404
B
LOTS 23.000 3.000 3 15.548 30.452
For entire sample 23.667 2.401 18 22.473 24.861
Tests of Significance for X using UNIQUE sums of squares
Source of Variation
SS DF MS
F Sig of F
WITHIN CELLS
74.00 12 6.17
A
3.56 1 3.56 .58 .462
B
7.00 2 3.50 .57 .581
A BY B
13.44 2 6.72 1.09 .367
Main Effect of A
A B X
|
1 1 32
1 1 34
1 1 35
1 2 36
1 2 39
1 2 33
1 3 31
1 3 35
1 3 32
2 1 21
2 1 26
2 1 25
2 2 24
2 2 20
2 2 24
2 3 23
2 3 26
2 3 20
Cell Means and Standard
Deviations
Variable .. X
SCORE ON FINAL EXAM
FACTOR
CODE Mean Std. Dev. N 95 percent
Conf. Interval
A BLUE‑BOO
B
NONE
33.667 1.528 3 29.872 37.461
B
SOME
36.000 3.000 3 28.548 43.452
B
LOTS
32.667 2.082 3 27.495 37.838
A COMPUTER
B
NONE
24.000 2.646 3 17.428 30.572
B
SOME
22.667 2.309 3 16.930 28.404
B
LOTS
23.000 3.000 3 15.548 30.452
For entire sample 28.667 6.078 18
25.644 31.689
Tests of Significance for X using UNIQUE sums of squares
Source of Variation
SS DF MS F Sig of F
WITHIN CELLS
74.00 12 6.17
A
533.56 1 533.56
86.52 .000
B
7.00 2 3.50 .57 .581
A BY B
13.44 2 6.72 1.09 .367
Main Effect of B
|
A B X
1 1 42
1 1 44
1 1 45
1 2 36
1 2 39
1 2 33
1 3 21
1 3 25
1 3 22
2 1 41
2 1 46
2 1 45
2 2 34
2 2 30
2 2 34
2 3 23
2 3 26
2 3 20
Cell Means and Standard Deviations
Variable .. X
SCORE ON FINAL EXAM
FACTOR
CODE Mean Std. Dev. N 95 percent
Conf. Interval
A BLUE‑BOO
B
NONE
43.667 1.528 3 39.872 47.461
B
SOME
36.000 3.000 3 28.548 43.452
B
LOTS
22.667 2.082 3 17.495 27.838
A COMPUTER
B
NONE 44.000 2.646 3 37.428 50.572
B
SOME
32.667 2.309 3 26.930 38.404
B
LOTS
23.000 3.000 3 15.548 30.452
For entire sample 33.667 9.133 18 29.125 38.208
Tests of Significance for X using UNIQUE sums of squares
Source of Variation
SS DF MS F Sig of F
WITHIN CELLS
74.00 12
6.17
A
3.56 1 3.56 .58 .462
B
1327.00 2 663.50
107.59 .000
A BY B
13.44 2 6.72 1.09 .367
AB Interaction
|
A B X
1 1 42
1 1 44
1 1 45
1 2 36
1 2 39
1 2 33
1 3 21
1 3 25
1 3 22
2 1 21
2 1 26
2 1 25
2 2 34
2 2 30
2 2 34
2 3 43
2 3 46
2 3 40
Cell Means and Standard Deviations
Variable .. X
SCORE ON FINAL EXAM
FACTOR
CODE Mean Std. Dev. N 95 percent
Conf. Interval
A BLUE‑BOO
B
NONE
43.667 1.528 3
39.872 47.461
B
SOME
36.000 3.000 3 28.548 43.452
B
LOTS
22.667 2.082 3 17.495 27.838
A COMPUTER
B NONE 24.000
2.646 3 17.428 30.572
B
SOME
32.667 2.309 3 26.930 38.404
B
LOTS
43.000 3.000 3 35.548 50.452
For entire sample 33.667 8.738 18 29.321 38.012
Tests of Significance for X using UNIQUE sums of squares
Source of Variation
SS DF MS F Sig of F
WITHIN CELLS 74.00 12 6.17
A
3.56 1 3.56 .58 .462
B
7.00 2 3.50 .57 .581
A BY B
1213.44 2 606.72
98.39 .000
Main Effects of A and B
|
A B X
1 1 52
1 1 54
1 1 55
1 2 46
1 2 49
1 2 43
1 3 31
1 3 35
1 3 32
2 1 41
2 1 46
2 1 45
2 2 34
2 2 30
2 2 34
2 3 23
2 3 26
2 3 20
Cell Means and Standard Deviations
Variable .. X
SCORE ON FINAL EXAM
FACTOR
CODE Mean Std. Dev. N 95 percent
Conf. Interval
A BLUE‑BOO
B
NONE 53.667 1.528 3 49.872 57.461
B
SOME
46.000 3.000 3 38.548 53.452
B
LOTS
32.667 2.082 3 27.495 37.838
A COMPUTER
B
NONE
44.000 2.646 3 37.428 50.572
B
SOME
32.667 2.309 3 26.930 38.404
B
LOTS 23.000 3.000 3 15.548 30.452
For entire sample 38.667
10.705 18 33.343 43.990
Tests of Significance for X using UNIQUE sums of squares
Source of Variation
SS DF MS F Sig of F
WITHIN CELLS
74.00 12 6.17
A
533.56 1 533.56
86.52 .000
B
1327.00 2 663.50
107.59 .000
A BY B
13.44 2 6.72 1.09 .367
Main effect of A, AB Interaction
|
A B X
1 1 52
1 1 54
1 1 55
1 2 46
1 2 49
1 2 43
1 3 31
1 3 35
1 3 32
2 1 21
2 1 26
2 1 25
2 2 34
2 2 30
2 2 34
2 3 43
2 3 46
2 3 40
Cell Means and Standard Deviations
Variable .. X
SCORE ON FINAL EXAM
FACTOR
CODE Mean Std. Dev. N 95 percent
Conf. Interval
A BLUE‑BOO
B
NONE
53.667 1.528 3 49.872 57.461
B
SOME
46.000 3.000 3 38.548 53.452
B
LOTS
32.667 2.082 3 27.495 37.838
A COMPUTER
B
NONE
24.000 2.646 3 17.428 30.572
B
SOME
32.667 2.309 3 26.930 38.404
B
LOTS
43.000 3.000 3 35.548 50.452
For entire sample 38.667
10.370 18
33.510 43.823
Tests of Significance for X using UNIQUE sums of squares
Source of Variation
SS DF MS F Sig of F
WITHIN CELLS
74.00 12 6.17
A
533.56 1 533.56
86.52 .000
B
7.00 2 3.50 .57 .581
A BY B
1213.44 2 606.72
98.39 .000
Main Effect of B, AB Interaction
A B X
|
1 1 32
1 1 34
1 1 35
1 2 46
1 2 49
1 2 43
1 3 21
1 3 25
1 3 22
2 1 31
2 1 36
2 1 35
2 2 34
2 2 30
2 2 34
2 3 33
2 3 36
2 3 30
Cell Means and Standard Deviations
Variable .. X
SCORE ON FINAL EXAM
FACTOR
CODE Mean Std. Dev. N 95 percent
Conf. Interval
A BLUE‑BOO
B
NONE
33.667 1.528 3
29.872 37.461
B
SOME
46.000 3.000 3 38.548 53.452
B
LOTS
22.667 2.082 3 17.495 27.838
A COMPUTER
B NONE 34.000
2.646 3 27.428 40.572
B
SOME
32.667 2.309 3 26.930 38.404
B
LOTS
33.000 3.000 3 25.548 40.452
For entire sample 33.667 7.268 18 30.052 37.281
Tests of Significance for X using UNIQUE sums of squares
Source of Variation
SS DF MS F Sig of F
WITHIN CELLS 74.00 12 6.17
A
3.56 1 3.56 .58 .462
B
397.00 2 198.50
32.19 .000
A BY B
423.44 2 211.72
34.33 .000
Main Effects of A and B, AB Interaction
A B X
|
1 1 22
1 1 24
1 1 25
1 2 36
1 2 39
1 2 33
1 3 41
1 3 45
1 3 42
2 1 41
2 1 46
2 1 45
2 2 44
2 2 40
2 2 44
2 3 43
2 3 46
2 3 40
Cell Means and Standard Deviations
Variable .. X
SCORE ON FINAL EXAM
FACTOR
CODE Mean Std. Dev. N 95 percent
Conf. Interval
A BLUE‑BOO
B NONE 23.667
1.528 3 19.872 27.461
B
SOME
36.000 3.000 3 28.548 43.452
B
LOTS
42.667 2.082 3 37.495 47.838
A COMPUTER
B
NONE
44.000 2.646 3 37.428 50.572
B
SOME
42.667 2.309 3 36.930 48.404
B LOTS 43.000
3.000 3 35.548 50.452
For entire sample 38.667 7.700 18 34.837 42.496
Tests of Significance for X using UNIQUE sums of squares
Source of Variation
SS DF MS F Sig of F
WITHIN CELLS
74.00 12 6.17
A
373.56 1 373.56
60.58 .000
B
247.00 2 123.50
20.03 .000
A BY B
313.44 2 156.72
25.41 .000
No Significant Effects
|
A B X
1 1 32
1 1 24
1 1 15
1 2 46
1 2 39
1 2 23
1 3 31
1 3 45
1 3 52
2 1 31
2 1 46
2 1 55
2 2 34
2 2 40
2 2 54
2 3 33
2 3 46
2 3 50
Cell Means and Standard Deviations
Variable .. X
SCORE ON FINAL EXAM
FACTOR
CODE Mean Std. Dev. N 95 percent
Conf. Interval
A BLUE‑BOO
B
NONE
23.667 8.505 3 2.539 44.794
B
SOME
36.000 11.790 3 6.712 65.288
B
LOTS 42.667 10.693 3 16.104 69.229
A COMPUTER
B
NONE
44.000 12.124 3 13.881 74.119
B
SOME
42.667 10.263 3
17.171 68.162
B
LOTS
43.000 8.888 3 20.920 65.080
For entire sample 38.667
11.499 18 32.948 44.385
Tests of Significance for X using UNIQUE sums of squares
Source of Variation
SS DF MS F Sig of F
WITHIN CELLS
1314.00 12 109.50
A
373.56 1 373.56 3.41 .090
B
247.00 2 123.50 1.13 .356
A BY B
313.44 2 156.72 1.43 .277
Note that the means and graphs of the last
two example data sets were identical.
The ANOVA table, however, provided a quite different analysis of each
data set. The data in this final set
was constructed such that there was a large standard deviation within each
cell. In this case the marginal and
cell means were not different enough to warrant rejecting the hypothesis of no
effects, thus no significant effects were observed.
Dot Notation Revisited
|
The reader may recall from the previous
chapter that placing dots instead of subscripts is a shorthand notation for
summation. For example
When two subscripts are involved the notation
becomes somewhat more complicated (and powerful). For example
|
and
|
or if a=1 then
|
When three subscripts are involved as is
necessary in an A X B design the notation involves even more
summation signs. For example
|
and, for example
|
where one sums over the subscript containing
the dot.
Using the dot notation with means rather than
sums is a relatively simple extension of the dot notation. The mean is found by
dividing the sum by the number of scores which were included in the sum. For example, the grand mean can be found as
follows:
|
and the cell means
|
All this is most easily understood in the
context of a data table.
b1 b2 b3
X111=10 X11.=33 X121=22 X12.=72 X131=14 X13.=45 X1..=150
a1 X112=11
_11.=11 X122=24 _12.=24 X132=15 _13.=15 _1..=16.67
X113=12
X123=26 X133=16
X211=20 X21.=63 X221=21 X22.=66 X231=18 X23.=57 X2..=186
a2 X212=21
_21.=21 X222=22 _22.=22 X232=19 _23.=19 _2..=20.67
X213=22
X223=23 X233=20
X.1..=96 X.2..=138 X.3..=102 X...=336
_.1..=16 _.2..=23 _.3..=17 _...=18.67
Chapter
11
Nested Two
Factor Between Groups Designs B(A)
The
Design
Factor B is
said to be nested within factor A when a given level of B appears under
a single level of A. This occurs, for
example, when the first three levels of factor B (b1 ,b3,
and b3) appear only under level a1 of factor A and the
next three levels of B (b4
,b5, and b6)
appear only under level a2 of factor A. These types of designs are also designated as hierarchical designs in
some textbooks.
A B X |
1 1
1 1 1
2 1 1
3 1 2
5 1 2
6 1 2
7 1 3
9 1 3 10 1 3 11 2 1
1 2 1
2 2 1
3 2 2
1 2 2
2 2 2
3 2 3
1 2 3
2 2 3
3 |
Nested or
hierarchical designs can appear because many aspects of society are
organized hierarchically. For example within the university, classes
(sections) are nested within courses, courses are nested within instructors,
and instructors are nested within departments.
In experimental
research it is also possible to nest treatment conditions within other
treatment conditions. For example in
studying the addictive potential of drugs, drug type could be nested within
drug function, i. e. Librium, Valium,
and Xanax are drugs used to treat anxiety, Prozac is used to treat depression,
and Halcion sold as a sleeping pill.
Here each drug appears in one and only one level of drug function.
The
Data
The data are
organized similarly to the A x B design.
Note that the data in the following table is identical to that in Table
6.1.
SPSS
commands
The SPSS
commands to do this analysis are
identical to those necessary to do the A x B analysis with one exception; the
DESIGN subcommand must completely specify the design. In this case the design can be specified with the effects A and B
WITHIN A, where A corresponds to the main effect of A, and B WITHIN A, the nested main effect of
B. The following text box contains the
SPSS program to run the analysis.
The
Analysis
RUN NAME PSYCHOLOGY 460 ‑ ASSIGNMENT
4 ‑ DAVID W. STOCKBURGER |
DATA LIST FILE='ANOVA2 DATA A'/A B X X2 X3 X4 1‑12 LIST. MANOVA X BY A(1,2) B(1,3) /PRINT=CELLINFO(MEANS) /DESIGN A B WITHIN A. |
The analysis of
the B(A) design is similar to the A x B design. The Tables of Means will be identical, but the graphs drawn from
the means will be done differently than the A x B design. The ANOVA table will contain only two
effects and interpretation will be discussed.
Cell Means and
Standard Deviations |
FACTOR CODE Mean Std.Dev. N
A 1
B 1 2.000 1.000 3
B 2 6.000 1.000 3
B 3 10.000 1.000 3
A 2
B 1 2.000 1.000 3
B 2 2.000 1.000 3
B 3 2.000 1.000 3 |
The
Table of Means
The Table of
Means produced in this analysis is identical to that produced in the A x B
design.
Graphs
|
|
Because the
qualitative or quantitative meaning of levels of a at each level of b is
different, the graph of the means must be modified. The nested main effects of b are presented side-by-side rather
than on top of one another. This is
necessary because b1 is different depending on whether it appears
under a1 or a2.
The
ANOVA Table
The ANOVA table
produced by design B(A) is presented below.
Source of Variation
SS DF MS
F Sig of F |
WITHIN CELLS 12.00 12 1.00 A 72.00 1
72.00 72.00 .000 B WITHIN A 96.00 4 4.00 24.00
.000 |
Interpretation
of Output
The
interpretation of the ANOVA table is straightforward. The WITHIN CELLS and A main effect SS's, DF's, and MS's are
identical to the analysis done in design A x B and are interpreted similarly.
The B WITHIN A term is called a nested main
effect and is the sum of the two simple B main effects. If this term is significant, then the graph
of the simple main effects should be drawn (Figure 8.1). What significance means is that the points
WITHIN each line are not the same height (value). That is, the means for b1, b2, and b3
within levels a1 and a2 are different. In the example graph the simple main effect
of B under a1 would be significant while the simple main effect of B
under a2 would not. Because
the B WITHIN A effect is the sum of both simple effects, the combined effect
was found to be significant in this case.
Similarities
to the A X B Analysis
As noted
earlier, the data files for the A x B and B(A) designs are identical. Likewise, the MANOVA commands are similar
except that the DESIGN subcommand on the B(A) design must be specified because
it is not a completely crossed design.
The table of means of the B(A) design will be identical to the A x B
design, the difference being that they will be plotted differently. The ANOVA source table will slightly differ
for the two types of designs. Both are
presented below to allow them to be contrasted.
Source Table
for the B(A) Design |
Source of Variation SS DF MS
F Sig of F WITHIN CELLS
12.00 12 1.00 A 72.00
1 72.00 72.00
.000 B WITHIN A 96.00 4 4.00
24.00 .000 Source Table for the A x B Design Source of Variation SS DF MS
F Sig of F WITHIN
CELLS 12.00 12
1.00 A 72.00 1
72.00 72.00 .000 B 48.00 2
24.00 24.00 .000 A BY
B 48.00 2
24.00 24.00 .000 |
Note that the
WITHIN CELLS and A effects are identical in both analyses. Note also that both the SS and DF for B and
A BY B in the A x B design together add up to the SS and DF for B WITHIN A for
the B(A) design. What is happening here
is that the B main effect and the A x B interaction in the A x B design is
collapsed into the nested main effect of B in the B(A) design.
Chapter
12
Contrasts,
Special and Otherwise
Understanding how contrasts may be employed
in the analysis of experiments gives the researcher considerably flexibility in
the specification of
"effects".
In addition, the study of contrasts leads to
even greater flexibility if multiple regression models are employed. Using contrasts, the researcher can test
specific “theory-driven” comparisons between groups.
Definition
A contrast is a set of numbers. When a set of means are being contrasted, a
contrast for the set of means will contain the same number of numbers as the
number of means. For example, if six
means are being contrasted, a contrast for the set of means would contain six
numbers. The order of the numbers in
the contrast corresponds to the order of the means. Examination of the signs of the numbers in the contrast describes
which means are being contrasted.
For example, the following contrast
1 1 1 -1 -1 -1
would compare the first three means with the
last three means.
A contrast of the form
2 2 -1 -1 -1 -1
would compare the first two means with the
last four means. The manner in which
the means are compared is determined by similar number. In the above example, the first two groups
share the same number (2), while the last four groups share the number minus
one (-1). All groups sharing the same
positive number are compared with all groups sharing the same negative
number. If a contrast contains zeros,
then those groups with zeros are not included in the contrast. For example:
0 0 3 -1 -1 -1
Would compare the third mean with the last
three means. The first and second means
would not be included in the calculation or interpretation of the contrast.
Sets of Contrasts
A set of contrasts is simply a number
of contrasts considered simultaneously.
For example, the three contrasts presented above could be combined into
a set of contrasts as follows:
contrast 1 2 2 -1 -1 -1 -1
contrast 2 0 0
3 -1 -1 -1
contrast 3 1 1
1 -1 -1 -1
Orthogonal Contrasts
An orthogonal contrast is two contrasts such
that when corresponding numbers in the contrasts are multiplied together and
then the products are summed, the sum of the products is zero. For example, in the preceding example,
contrasts 1 and 2 are orthogonal as can be seen in the following:
contrast 1 2 2 -1 -1 -1 -1
contrast 2 x
0 x 0 x 3 x -1 x
-1 x -1
products =
0 = 0 = -3 = 1 = 1 = 1
The sum of the products, 0 + 0 + -3 + 1 + 1 +
1 , equals zero, thus contrast 1 and contrast 2 are orthogonal contrasts.
Non-orthogonal Contrasts
Non-orthogonal contrasts are two contrasts such that when
corresponding numbers in the contrasts are multiplied together and then the
products are summed, the sum of the products is not zero. For example, in the preceding example,
contrasts 2 and 3 are non-orthogonal as can be seen in the following:
contrast 3 1 1 1 -1 -1 -1
contrast 2 x
0 x 0 x 3 x -1 x
-1 x -1
products =
0 = 0 = 3 =
1 = 1 = 1
The sum of the products, 0 + 0 + 3 + 1 + 1 +
1 , equals six, thus contrast 2 and contrast 3 are non-orthogonal
contrasts. In a similar manner
contrasts 1 and 3 are non-orthogonal.
Sets of Orthogonal Contrasts
The guiding principle is that the number
of possible orthogonal contrasts will always equal the number of means being
contrasted. If six means are being
contrasted, there will be no more than six contrasts which will be orthogonal
to one another. If six means are being
contrasted and five orthogonal contrasts have already been found, then there
exists a contrast which is orthogonal to the first five contrasts. That contrast is not always easy to find.
Finding Sets of Orthogonal Contrasts
In ANOVA the first contrast, which will be
referred to as contrast 0, is a set of 1's.
This contrast will be seen to be equivalent to the m term in the score
model. For a contrast comparing six
means, the first contrast would be:
contrast 0 1 1 1 1 1 1
In order for any following contrasts to be
orthogonal with contrast 0, it can be seen that the sum of the numbers in the
contrast must equal zero. Both contrasts 1 and 2 described earlier fit this
criterion, and since they are orthogonal to one another, they will be included
in the current set of orthogonal contrasts.
contrast 1 2 2 -1 -1 -1 -1
contrast 2 0 0 3 -1 -1 -1
In finding a fourth contrast which is
orthogonal to the first three, look for patterns of means in the preceding
contrasts which have the same number.
In both contrasts 1 and 2, the first and second means have the same
number, 2 in contrast 1 and 0 in contrast 2.
Working with this subset, finding numbers which sum to zero, and setting
all other numbers to zero , the following contrast is found:
contrast 3 1 -1 0 0 0 0
The student should verify that this contrast
is indeed orthogonal to all three preceding contrasts.
Using the same logic, the fifth contrast will
involve the fourth, fifth, and sixth means.
There are any number of possibilities, but the following will be used:
contrast 4 0 0 0 -2 1 1
The final contrast will compare the fifth
mean and the sixth mean. Any two
numbers which sum to zero (i. e. 2.34
and -2.34) could be used without changing the analysis or interpretation, but 1
and -1 make the contrasts parsimonious.
contrast 5 0 0 0 0 1 -1
Putting the six contrasts together as a set
results in the following:
contrast 0 1
1 1 1 1 1
contrast 1 2
2 -1 -1 -1 -1
contrast 2 0
0 3 -1 -1 -1
contrast 3 1 -1 0
0 0 0
contrast 4 0
0 0 -2 1
1
contrast 5 0
0 0 0 1 -1
G X |
1 1 1 2 1 3 2 5 2 6 2 7 3 9 3 10 3 11 4 1 4 2 4 3 5 1 5 2 5 3 6 1 6 2 6 3 |
Again, the student should verify that all the
contrasts in the set are orthogonal to one another. There is no other contrast available which is orthogonal to all
six of the preceding contrasts. The
non-believing student is welcome to search for such a contrast. I wish him or her well and hope that they
have a lot of spare time on their hands.
The Data
SPSS requires that the data for contrasts be
set up as a single factor design. The
example data is set up for a single factor between groups design with six
levels to the factor. The data appears
in the following text box.
RUN NAME PSYCHOLOGY 460 ‑ ASSIGNMENT 3 |
DATA LIST FILE='ANOVA2
DTA A'/G A B X X2 X3 X4 1‑14. LIST. MANOVA X BY G(1,6) /PRINT=CELLINFO(MEANS) /PRINT=PARAM(ESTIM) /CONTRAST(G)=SPECIAL ( 1
1 1 1 1 1 2
2 ‑1 ‑1 ‑1 ‑1 0
0 3 ‑1 ‑1 ‑1 ‑1 1 0 0
0 0 0
0 0 2 ‑1 ‑1 0
0 0 0 1 ‑1 ) /DESIGN G(1) G(2) G(3) G(4) G(5). |
SPSS commands
The SPSS language requires that the contrasts
be specified as a subcommand on the MANOVA command. The following text box presents the appropriate commands to run
contrasts of means.
The Analysis
The Table of Means
Cell Means and Standard Deviations |
Variable .. X FACTOR CODE Mean
Std. Dev. N G 1 2.000
1.000 3 G 2 6.000
1.000 3 G 3 10.000
1.000 3 G 4 2.000
1.000 3 G 5 2.000
1.000 3 G 6 2.000
1.000 3 4.000
3.254 18 |
The table of means is presented below:
The ANOVA table
The ANOVA table lists each contrast as an
effect.
* * A N A L Y S I S O F V A R I A N C E ‑‑
DESIGN 1 * * |
Tests of Significance for X using UNIQUE
sums of squares Source SS DF MS
F Sig of F WITHIN CELLS 12.00 12 1.00 G(1) .00
1 .00 .00
1.000 G(2) 144.00
1 144.00 144.00 .000 G(3) 24.00
1 24.00 24.00
.000 G(4) .00
1 .00 .00
1.000 G(5) .00
1 .00 .00
1.000 |
Interpretation of Output
Note that only contrasts 2 and 3 are
significant. Contrast 2, comparing the
mean of Group 3 with Groups 4, 5, and 6 has an F-ratio of 144.00,
significant. Contrast 3, comparing the
means of Groups 1 and 2, is also significant.
The other three contrasts are all non-significant, comparing groups with
the same mean.
Constants
Compare the above ANOVA table with one
produced by treating the design as a single factor ANOVA (below).
Source of Variation SS
DF MS F
Sig of F |
WITHIN CELLS
12.00 12 1.00 G
168.00 5 33.60
33.60 .000 |
One can observe that the SS for the single
factor - G (168.00) is the sum of the SS for the five contrasts listed above
(.00 + 144.00 + 24.00 + .00 + .00 = 168.00).
In general it can be stated The sum of the SS for any set of
orthogonal contrasts will equal the SS for that effect. The degrees of freedom (DF) for any effect
will be the number of contrasts summed for that
effect.
To demonstrate this constancy, using the
following set of orthogonal contrasts:
contrast 0 1
1 1 1 1 1
contrast 1 5 -1 -1 -1 -1 -1
contrast 2 0
4 -1 -1 -1 -1
contrast 3 0
0 3 -1 -1 -1
contrast 4 0
0 0 2 -1 -1
contrast 5 0
0 0 0 1 -1
The following ANOVA table is produced:
Source SS DF MS F
Sig of F |
WITHIN CELLS 12.00 12 1.00 G(1) 14.40
1 14.40 14.40 .003 G(2) 9.60
1 9.60 9.60 .009 G(3) 144.00
1 144.00 144.00 .000 G(4) .00
1 .00 .00
1.000 G(5) .00
1 .00 .00
1.000 |
Note that in this case the first three
contrasts are significant and that the sum of the SS for the five contrasts is
equal to 168.00.
Contrasts, Designs, and Effects
It will now be demonstrated that the
different models for the analysis of
experiments are simply special cases of
contrasts.
For example, the following set of contrasts
contrast 0
1 1 1 1
1 1
contrast 1 -1 -1
-1 1 1 1
contrast 2 2 -1
-1 2 -1 -1
contrast 3 0 -1 1 0 -1 1
contrast 4 0 -1 1 0
1 -1
contrast 5 2 -1
-1 -2 1 1
produces the following ANOVA table
Note that in this case all contrasts
are significant. Comparing this with
the ANOVA table produced by an A x B design with A possessing 2 levels and B 3
levels yields the following:
Source
of Variation SS DF
MS F Sig of F |
WITHIN CELLS 12.00 12 1.00 A 72.00
1 72.00 72.00
.000 B 48.00
2 24.00 24.00
.000 A BY B 48.00
2 24.00 24.00
.000 |
Source SS DF MS F
Sig of F |
WITHIN CELLS
12.00 12 1.00 G(1)
72.00 1 72.00
72.00 .000 G(2)
36.00 1 36.00
36.00 .000 G(3)
12.00 1 12.00
12.00 .005 G(4)
12.00 1 12.00
12.00 .005 G(5) 36.00 1 36.00 36.00
.000 |
Note that the A main effect corresponds to
the first contrast and has one degree of freedom. The B main effect corresponds to the sum of contrasts 2 and 3 and
has two degrees of freedom. Finally,
the A x B interaction corresponds to the sum of contrasts 4 and 5 with two
degrees of freedom.
The preceding contrasts were a special form
of the general form of main effects and interactions. For a 2 x 2 design
contrast 0 1 1
1 1
A main effect 1 1 -1 -1
B main effect 1 -1
1 -1
A x B interaction 1 -1 -1
1
In a like manner, contrasts corresponding to
the B(A) design may be specified. From
the general form for a 2 x 2 design
contrast 0 1 1
1 1
A main effect 1 1 -1 -1
simple B1 main effect 1 -1 0
0
simple B2 main effect 0
0 1 -1
the set of contrasts for a 2 x 3 design may
be found as follows
contrast 0
1 1 1
1 1 1
contrast 1 -1 -1 -1
1 1 1
Contrast 2 2 -1
-1 0 0 0
contrast 3 0 -1 1 0
0 0
contrast 4 0
0 0 2 -1 -1
contrast 5 0
0 0 0 -1 1
which produce the following ANOVA table.
Source SS DF
MS F Sig of F |
WITHIN CELLS 12.00
12 1.00 G(1) 72.00 1 72.00 72.00
.000 G(2) 72.00 1 72.00 72.00
.000 G(3) 24.00 1 24.00 24.00
.000 G(4) .00 1 .00 .00
1.000 G(5) .00 1 .00 .00
1.000 |
Comparing this to the ANOVA table produced by
design B(A) yields
Source of Variation
SS DF MS
F Sig of F |
WITHIN CELLS 12.00 12 1.00 A 72.00
1 72.00 72.00
.000 B WITHIN A 96.00 4 4.00
24.00 .000 |
Contrast 1 corresponds to the A main effect,
contrasts 2 and 3 correspond to the simple main effect of B1 and
contrasts 4 and 5 to the simple main effect of B2.
Non-Orthogonal Contrasts
When the specified set of contrasts is
non-orthogonal, the sum of the sum of squares for the contrasts will not equal
the total sum of squared for the effects.
In some cases the sum of the contrast sum of squares will be greater
than the total sum of squares, in others, the sum of the contrast sum of
squares will be less. An example of
each will now be given.
Smaller than Total Sum of Squares
The following set of contrasts produces a sum
of contrast sum of square which is less than the total sum of squares (168.00)
for a set of non-orthogonal contrasts.
Note that while the WITHIN SS remains constant, the sum of the contrast
SS is only 24.00. This was accomplished
by contrasting groups with similar means.
MANOVA X BY G(1,6) |
/CONTRAST(G)=SPECIAL ( 1 1
1 1 1 1 2 2 ‑1 ‑1 ‑1 ‑1 0 0
0 1 0 ‑1 0 0
0 1 ‑1 0 1 ‑1 0
0 0 0 1 0
0 0 0 ‑1 ) /DESIGN G(1)
G(2) G(3) G(4) G(5). Source of Variation SS DF MS F Sig of F WITHIN CELLS
12.00 12 1.00 G(1) .00 1
.00 .00 1.000 G(2) .00 1
.00 .00 1.000 G(3) .00 1
.00 .00 1.000 G(4) 24.00 1
24.00 24.00 .000 G(5) .00 1
.00 .00 1.000 |
Larger than Total Sum of Squares
On the other hand, if non-orthogonal
contrasts are specified to contrast groups which have different means, then the
sum of the SS's will be larger than the
total SS. In the following case, the
sum of the SS's is 424.00 compared with a total SS for orthogonal effects of
168.00. Thus, if non-orthogonal effects
are specified, then the ANOVA procedure will distort the nature of the effects
depending upon how the effects were specified.
MANOVA X BY G(1,6) |
/CONTRAST(G)=SPECIAL ( 1 1
1 1 1 1 2 2 ‑1 ‑1 ‑1 ‑1 1 ‑1 0
0 0 0 0 0
3 ‑1 ‑1 ‑1 0 0
2 ‑1 ‑1 0 0 0
2 ‑1 0 ‑1 ) /DESIGN G(1)
G(2) G(3) G(4) G(5). Source of Variation
SS DF MS F Sig of F WITHIN
CELLS 12.00 12
1.00 G(1) .00 1
.00 .00 1.000 G(2) 24.00 1
24.00 24.00 .000 G(3) 144.00 1
144.00 144.00 .000 G(4) 128.00 1
128.00 128.00 .000 G(5) 128.00 1
128.00 128.00 .000 |
Standard Types of Orthogonal
Contrasts
A number of standard orthogonal contrasts are
built into MANOVA. Instead of
specifying "/CONTRAST(G)=SPECIAL(
)", a standard contrast may be done by using "/CONTRAST(G)=XXX",
where the "XXX" is one of the following:
DIFFERENCE
This is a difference or reverse Helmert
contrast which compares levels of a factor with the mean of the previous levels
of the factor. For the six level
design, the contrast would appear as follows:
contrast 0 1 1 1 1 1 1
contrast 1 1 -1
0 0 0 0
contrast 2 1 1 -2 0 0 0
contrast 3 1 1 1 -3
0 0
contrast 4 1 1 1 1 -4 0
contrast 5 1 1 1 1 1 -5
SIMPLE
Compares each level of a factor to the last
level. For the six level design, the
contrast would appear as follows:
contrast 0 1 1 1 1 1 1
contrast 1 1 0
0 0 0 -1
contrast 2 0 1 0 0 0 -1
contrast 3 0 0 1 0
0 -1
contrast 4 0 0 0 1
0 -1
contrast 5 0 0 0 0 1 -1
POLYNOMIAL
|
|
This contrast requires that the levels
of the independent variable be measured on at least a good approximation to an interval
scale. If this is the case, then
the polynomial contrasts tests for trends.
The first contrast tests for a linear
trend. That is, as the value of the
independent variable increases, the dependent variable increases or decreases
at a steady rate. The increase or
decrease may be described with a straight line.
|
|
|
The second contrast tests for quadratic
trends. The quadratic trend my be
described with a parabola, it changes direction twice. The third trend, the cubic, changes
directions three times. For each trend
the number of changes in direction is the size of the trend. The number of trends for a given set of
contrasts will be one less than the number of groups being contrasted. An illustration of each of the five trends
for six groups is given below.
The set of orthogonal contrasts which tests
for polynomial trends for six means is presented below:
contrast 0 1 1 1 1 1 1
contrast 1
-5 -3 -1 1
3 5
contrast 2 5 -1 -4 -4 -1 5
contrast 3 -5 7 4 -4 -7 5
contrast 4 1 -3 2 2 -3 1
contrast 5 1 -5 10 -10 5 -1
|
Example of Polynomial
Contrasts
Given that the data produced the following
table of means and resulting graph.
The MANOVA of the data using a polynomial
contrast would produce the following source table.
Variable .. X |
FACTOR CODE Mean Std.
Dev. N G 1 2.000 1.000 3 G 2 6.000 1.000 3 G 3 8.000 1.000 3 G 4 8.000 1.000 3 G 5 7.000 1.000 3 G 6 10.000 1.000 3 For entire sample 6.833 2.684 18 |
It
can be seen that the first three contrasts; the linear, the quadratic, and the
cubic, were significant, while the last two were not. The graph of the means changes direction three times, with a
slight overall rise (linear trend). A
model of the following form would be used to describe the relationship between
groups (X) and the dependent measure (Y).
Y' = m
+ c1 * X + c2
* X2 + c3 * X3
Only significant contrasts would be included
in the model.
MANOVA X BY G(1,6) |
/CONTRAST(G)=POLYNOMIAL /PRINT = CELLINFO(MEANS) /DESIGN G(1) G(2) G(3) G(4) G(5). Source
of Variation SS DF
MS F Sig of F WITHIN
CELLS 12.00 12
1.00 G(1) 79.24 1
79.24 79.24 .000 G(2) 10.32 1
10.32 10.32 .007 G(3) 18.15 1
18.15 18.15 .001 G(4) 2.68 1
2.68 2.68 .128 G(5) .11 1
.11 .11 .749 |
Other types of contrast may be
available. Check the SPSS manual.
Conclusion
Contrasts of means provide a very general
mechanism for testing hypotheses about data.
The two-factor and nested main effect designs were seen to be special
cases of orthogonal contrasts.
Contrasts are probably not used as much in ANOVA as they could be.
Chapter
13
ANOVA and
Multiple Regression
Analysis of variance may be conceptualized
and performed as a special case of multiple regression. Doing so provides the user with a flexible
analytical tool which can be applied to a wide variety of data types. In addition,
conceptualization of ANOVA in this manner
permits a greater understanding of the issues involved when unequal number of
subjects appear in between subjects designs with two or more factors.
A review of a one factor ANOVA will be
discussed first, with an additional discussion of orthogonal contrasts. The concept of orthogonal transformations
translates easily into “dummy coding” and the use of multiple regression for
ANOVA. After the discussion of one
factor ANOVA, the conceptualization will be extended to two factor ANOVA in a
similar manner.
ONE FACTOR ANOVA
ANOVA and Multiple Regression
Up until this point, the stated purpose of
multiple regression was the prediction of a continuous dependent variable (Y)
from one or more continuous variables (X’s).
Using the multiple regression approach allowed the testing of hypotheses
about whether a variable or set of variables (X’s) significantly predicted the
variance of the dependent variable (Y).
The purpose of one factor ANOVA was to make a decision as to whether the
differences between means on a continuous dependent variable (Y) for levels of
an independent variable, or factor, were large enough to attribute the
differences between the means to something other than chance.
In order to extend the multiple regression
approach to incorporate ANOVA, the requirement that the independent variable(s)
be continuous must be extended to the case of discrete variables. This is accomplished by recoding the single
independent variable of ANOVA into multiple independent variables for inclusion
into a multiple regression analysis.
This process is called “dummy coding” and if one understands orthogonal
contrasts, the transition is straightforward.
1 23 |
1 26 1 31 2 34 2 32 2 38 3 45 3 41 3 49 4 18 4 24 4 23 5 31 5 21 5 27 6 28 6 34 6 22 |
Example
MANOVA |
x BY group(1 6)
/DESIGN .
GROUP MEAN
1 26.67
2 34.67
3 45.00
4 21.67
5 26.33
6 28.00 |
Start with a one factor between subjects
design, A, where there are six levels of A (A=6) and three subjects in
each group (S=3). An example data file
is presented below, with the SPSS commands, means, and ANOVA table
following. It can be seen that the
means are significantly different from each other with F(5,15)=10.92,
p<.05. The sum of squares for the treatment
effect and error are equal to 1031.61 and 226.67 respectively.
* * * * * * A n a l y s I
s o f V a r I a n c e ‑‑ design 1 * * * * * * |
Tests of
Significance for X using UNIQUE sums of squares Source of
Variation SS DF MS F Sig of F WITHIN+RESIDUAL
226.67 12 18.89 GROUP 1031.61 5
206.32 10.92 .000 (Model)
1031.61 5 206.32 10.92 .000 (Total)
1258.28 17 74.02 R‑Squared
= .820 Adjusted R‑Squared
= .745 |
Example Using Contrasts
MANOVA |
x BY group(1 6) /CONTRAST
(GROUP) SPECIAL ( 1
1 1 1 1 1
1 1 1 1 ‑2 ‑2
1 1 ‑1 ‑1 0
0 1 ‑1 0
0 0 0
0 0 1 ‑1 0 0
0 0 0 0 1 ‑1 ) /DESIGN
GROUP(1) GROUP(2) GROUP(3)
GROUP(4) GROUP(5) . |
As discussed in the previous chapter, the
single factor ANOVA may be performed as a series of orthogonal contrasts. Any orthogonal set of contrasts will account
for exactly the same amount of variance and produce the identical decision as
the single factor ANOVA. As a
demonstration of this equality, the example data presented in the previous
section was re-analyzed using the set of orthogonal contrasts specified below.
The end results of this analysis are
identical to the results of the previous analysis. Note that the sum of the sum of squares for the various contrasts
equals the sum of squares for the Groups factor.
* * A n a l y s I s o
f V a r I a n c e * * |
Tests of
Significance for X using UNIQUE sums of squares Source of
Variation SS DF
MS F Sig of F WITHIN+RESIDUAL
226.67 12 18.89 GROUP(1)
93.44 1 93.44 4.95 .046 GROUP(2)
21.33 1
21.33 1.13 .309 GROUP(3)
96.00 1 96.00 5.08 .044 GROUP(4)
816.67 1 816.67
43.24 .000 GROUP(5)
4.17 1 4.17 .22 .647 (Model)
1031.61 5 206.32
10.92 .000 (Total)
1258.28 17 74.02 R‑Squared
= .820 Adjusted R‑Squared
= .745 |
Dummy Coding
To use a multiple regression approach to
perform an ANOVA, the single independent variable with K levels must be recoded
as K-1 independent variables. The form
of the recoding is identical to the orthogonal contrast. For the examination of the total effect, any
set of orthogonal contrasts will yield the same overall result. Partitioning a single categorical variable
into a set of variables suitable for multiple regression is called “dummy
coding.”
GROUP D1 D2 D3 D4 D5 X |
1 1
1 1 0 0 23 1 1
1 1 0 0 26 1 1
1 1 0 0 31 2 1
1 ‑1 0 0 34 2 1
1 ‑1 0 0 32 2 1
1 ‑1 0 0 38 3 1 ‑1 0 1 0 45 3 1 ‑1 0 1 0 41 3 1 ‑1 0 1 0 49 4 1 ‑1 0 ‑1 0 18 4 1 ‑1 0 ‑1 0 24 4 1 ‑1 0 ‑1 0 23 5 ‑2
0 0 0 1 31 5 ‑2
0 0 0 1 21 5 ‑2
0 0 0 1 27 6 ‑2
0 0 0 ‑1 28 6 ‑2
0 0 0 ‑1 34 6 ‑2
0 0 0 ‑1 22 |
The dummy codes may be created by turning the
set of orthogonal contrasts on their side (called the transpose of a matrix)
and assigning each subject in the group the same codes. For example, dummy coding the example data
set with the example set of contrasts produces the following data set. Note that there are five new variables
created, one for each of the orthogonal contrasts. The first contrast, which is a row of ones, is not included,
because the constant term (b0) in the multiple regression procedure
does the same function.
The dummy coded variables will be orthogonal,
or uncorrelated with one another, if there are an equal number of subjects in
each group. In the preceding example
this was the case and the correlation matrix which results is as follows. Note that all off-diagonal entries are zero.
‑ ‑ Correlation Coefficients
‑ ‑ |
D1 D2 D3 D4 D5 D1
1.0000 .0000 .0000 .0000 .0000 D2
.0000 1.0000 .0000 .0000 .0000 D3
.0000 .0000 1.0000 .0000 .0000 D4
.0000 .0000 .0000 1.0000 .0000 D5
.0000 .0000 .0000 .0000 1.0000 |
ANOVA, Revisited
REGRESSION |
/DEPENDENT x
/METHOD=ENTER d1 d2 d3 d4 d5. |
The ANOVA may now be done by regressing the
dummy coded variables as independent variables, in this case D1, D2, . . . , on the dependent variable, in this
case X. The REGRESSION command in SPSS
does this function and an example of the use of this command is given below.
The output from this command is presented
below:
* * * * M U L
T I P L E R E G R E S S I O N * * * * |
Block Number
1. Method: Enter D1 D2 D3 D4 D5 Multiple R .90546 R Square .81986 Adjusted R Square .74480 Standard Error 4.34613 Analysis of Variance DF
Sum of Squares Mean Square Regression 5
1031.61111 206.32222 Residual 12
226.66667 18.88889 F =
10.92294 Signif F = .0004 ‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑
Variables in the Equation ‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑ Variable B SE
B Beta T Sig T D5
‑.833333 1.774302 ‑.057545 ‑.470 .6470 D4
11.666667 1.774302 .805627 6.575 .0000 D3
‑4.000000
1.774302 ‑.276215 ‑2.254 .0436 D2
‑1.333333
1.254621 ‑.130209 ‑1.063 .3088 D1
1.611111 .724356 .272514 2.224 .0461 (Constant)
30.388889 1.024394 29.665 .0000 |
Note that the values of DF, Sum of Squares,
Mean Square, F, and Sig of F for the “Regression” and “Residual” effects in the
source table are identical to those of the “GROUP” and “WITHIN+RESIDUAL”
effects in the output of the MANOVA command.
A one-factor ANOVA has in effect been performed.
TWO FACTOR ANOVA
In the same manner that orthogonal contrasts
may be used to perform a two factor ANOVA, dummy codes may be selected such
that a multiple regression analysis may perform the same function. The general procedure is identical to dummy
coding procedure for a one factor ANOVA, except that different dummy codes are
used. The recoded variables are
partitioned into subsets and each subset is included as a hypothesis test in
the regression model.
A B X |
1 1 23 1 1 26 1 1 31 1 2 34 1 2 32 1 2 38 1 3 45 1 3 41 1 3 49 2 1 18 2 1 24 2 1 23 2 2 31 2 2 21 2 2 27 2 3 28 2 3 34 2 3 22 |
Example
The raw data, SPSS MANOVA command, and SPSS
output is presented below. The dependent
variable is identical to that in the one factor case and the six levels of the
single factor have been partitioned into two level of a factor called “A” and
three levels of a factor called “B.”
MANOVA |
x BY a(1
2) b(1 3) /DESIGN . |
A n a l y s I s
o f V a r I a n c e |
Tests of
Significance for X using UNIQUE sums of squares Source of
Variation SS DF
MS F Sig of F WITHIN+RESIDUAL
226.67 12 18.89 A 460.06 1
460.06 24.36 .000 B 456.44 2
228.22 12.08 .001 A BY
B 115.11 2
57.56 3.05 .085 (Model)
1031.61 5 206.32
10.92 .000 (Total)
1258.28 17 74.02 R‑Squared
= .820 Adjusted R‑Squared
= .745 |
Example Using Contrasts
The set of orthogonal contrasts may be
selected such that the output of the MANOVA command produces results similar to
those in the above example. The raw
data for this analysis is that presented in the one factor ANOVA section in
this chapter. The MANOVA command and
results of this command are presented below:
MANOVA |
x BY group(1 6)
/CONTRAST(GROUP) SPECIAL ( 1
1 1 1 1 1
1 1 1 ‑1 ‑1 ‑1
1 1 ‑2 1
1 ‑2 1 ‑1 0
1 ‑1 0
1 1 ‑2 ‑1 ‑1 2
1 ‑1 0 ‑1 1 0 ) /DESIGN
GROUP(1) GROUP(2)+GROUP(3) GROUP(4)+GROUP(5). |
A n a l y s I s
o f V a r I a n c e |
Tests of
Significance for X using UNIQUE sums of squares Source of
Variation SS DF
MS F Sig of F WITHIN+RESIDUAL
226.67 12 18.89 GROUP(1)
460.06 1 460.06
24.36 .000 GROUP(2) +
GROUP(3) 456.44 2
228.22 12.08 .001 GROUP(4) +
GROUP(5) 115.11 2
57.56 3.05 .085 (Model)
1031.61 5 206.32
10.92 .000 (Total) 1258.28
17 74.02 R‑Squared
= .820 Adjusted R‑Squared
= .745 |
Note that the results of this analysis are
identical to that of the two factor ANOVA.
Regression Analysis using Dummy
Coding
A
B DA DB1 DB2 DAB1 DAB2 X |
1 1
1 1 1 1 1
23 1 1
1 1 1 1 1
26 1 1
1 1 1 1 1
31 1 2
1 1 ‑1 1 ‑1
34 1 2
1 1 ‑1 1 ‑1
32 1 2
1 1 ‑1 1 ‑1
38 1 3
1 ‑2 0
‑2 0 45 1 3
1 ‑2 0
‑2 0 41 1 3
1 ‑2 0
‑2 0 49 2 1
1 1 1 ‑1 ‑1
18 2 1 ‑1 1 1 ‑1
‑1 24 2 1 ‑1 1 1 ‑1
‑1 23 2 2 ‑1 1 ‑1 ‑1 1 31 2 2 ‑1 1 ‑1 ‑1 1 21 2 2 ‑1 1 ‑1 ‑1 1 27 2 3 ‑1 ‑2 0 2
0 28 2 3 ‑1 ‑2 0 2
0 34 2 3 ‑1 ‑2 0 2
0 22 |
In a manner analogous to the dummy coding of
the one factor ANOVA, the contrast matrix is transposed and expanded such that
each subject in a group is given the same X variables. The variables are labeled “DA, DB1, DB2,
DAB1, and DAB2" where the D indicates that it is a dummy code and the
letter or letters which follow indicate the name of the effect. Because the B and AB factors have two
degrees of freedom each, the effect of these factors will be a combination of
two dummy coded variables, hence the “1" and “2" following the
recoded variable. The data matrix for
the multiple regression analysis is presented below:
The correlation matrix including all the dummy
coded variables is as follows:
‑ ‑ Correlation Coefficients
‑ ‑ |
DA DB1 DB2 DAB1 DAB2 DA
1.0000 .0000 .0000 .0000 .0000 DB1
.0000 1.0000 .0000 .0000 .0000 DB2
.0000 .0000 1.0000 .0000 .0000 DAB1
.0000 .0000 .0000 1.0000 .0000 DAB2
.0000 .0000 .0000 .0000 1.0000 |
Note that the independent variables are
uncorrelated with each other, the definition of orthogonality or
independence. The implication of
independence is that the order of entering the variables into the regression
model will not make a difference.
REGRESSION |
/DEPENDENT x
/METHOD=TEST (DA) (DB1,DB2)
(DAB1,DAB2). |
The REGRESSION command which does the two
factor ANOVA with the dummy coded variables is presented below. The “/METHOD=TEST” subcommand will
sequentially test each of the combinations of variables in the parentheses
following the word “TEST”. In this
case, DA corresponds to the main effect of A, DB1 and DB2 to the main effect of
B, and DAB1 and DAB2 to the interaction effect. The result of this regression analysis is presented below.
* * * * M U L
T I P L E R E G R E S S I O N * * * * |
Hypothesis Tests Sum
of DF Squares Rsq Chg F Sig F Source 1 460.05556 .36562 24.35588 .0003
DA 2 456.44444 .36275 12.08235 .0013
DB1 DB2 2 115.11111 .09148 3.04706 .0851
DAB1 DAB2 5 1031.61111 10.92294
.0004 Regression 12 226.66667 Residual 17 1258.27778 Total ‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑‑ Multiple R .90546 R Square .81986 Adjusted R Square .74480 Standard Error 4.34613 |
Note that the DF, Sum of Squares, F, and Sig
F are identical to those of the two MANOVA analyses.
Conclusion
It at first
appears that dummy coding is a great deal of trouble for very little gain. The results of the analysis are identical to
those of the simpler MANOVA procedures.
There are at least three reasons for viewing ANOVA as a special case of
multiple regression.
The first
reason is to achieve a better understanding of how the MANOVA procedure
works. When one uses the MANOVA
program, the little guy inside the computer dummy codes the independent
variable(s) and then applies a general linear model or regression analysis to
the results. In a like manner, a
greater understanding of other multivariate procedures, such a discriminant
analysis, is gained if one understands dummy coding.
The second
reason for the understanding of dummy coding and the use of multiple regression
to perform an ANOVA is that considerable flexibility is gained in the type of
analysis that may be done. If the
independent variables in an analysis include both interval and categorical
variables, both may be used in the same analysis. The categorical variables are dummy coded and entered either
before, with, or after the interval variables.
The dummy coded variables may be combined into different “effects” and
entered into the regression equation in different orders.
The final and
perhaps overriding reason for including a discussion of the relationship
between multiple regression and ANOVA in this book is that it is difficult to
explain the nature of the difficulty of unequal cell sizes in ANOVA without
recourse to multiple regression. That
is the subject of the next chapter.
Chapter
14
Unequal Cell
Frequencies
Up to this point all computational procedures
and theoretical development have been done for balanced designs. Balanced designs have equal or proportional
cell frequencies in each cell.
Employing unbalanced designs has profound effects on both the interpretation
and theoretical development of ANOVA.
Since an unbalanced design is often employed in real life situations, a
careful examination of unbalanced designs is warranted.
In the following, a review of a balanced
two-way between subjects ANOVA will first be presented. Following that, an unbalanced design of the
same form will be analyzed using various options available from SPSS. Since I entered the problem in a discussion
list (ED-STAT), the solutions resulting from the resulting discussion will be presented
last.
Equal Cell Frequency - Independence
of Effects
In a balanced design the computational
procedures to estimate the terms in the score model produce independent, or
uncorrelated, effects. This result of
independent effects is that the total
of the sum of squares for all effects adds up to the total SS. It makes no difference which effect is
computed first, second, or last, since the size of one effect has no
relationship to the size of any other effect
This can be seen very nicely in a source
table from a balanced A X B design.
Tests of
Significance for X using UNIQUE sums of squares |
Source SS DF MS F Sig of F WITHIN
CELLS 139.36 12
11.61 A .36
1 .36 .03 .863 B 123.08
2 61.54 5.30 .022 A BY B 108.73
2 54.36 4.68 .031 Model 232.17 5
46.43 4.00 .036 Total 271.53 17 |
Note that the SS for the Model is equal to SSA
+ SSB + SSAB.
Here no matter what order the A, B, or A BY B terms are entered into the
model, the result would be the same.
Unequal Cell Frequency - Dependent
Effects
When the AXB design is unbalanced, it makes a
great deal of difference which order the effects are entered into the
model. In order for the following to
make sense, the reader must understand that SPSS has two options for dealing
with unbalanced designs. These options
are selected by including either the subcommand "/METHOD = UNIQUE" or
"/METHOD = SEQUENTIAL" to the MANOVA command. The SEQUENTIAL procedure enters the effects
into the model in the order that they are listed in the "/DESIGN"
subcommand. For example, if the default
"/DESIGN = A B A BY B" was used, the A main effect would be entered
first, the B main effect second, and the interaction A BY B would be last.
The use of
"/DESIGN = A BY B B A" and "/METHOD = SEQUENTIAL"
would result in the interaction term being entered first, the B main effect
second, and the A main effect last. As
shall be seen, the order of effects is critical and greatly effects the results
of the analysis.
The "/METHOD = UNIQUE" subcommand
is the default in SPSS. That is, it is
assumed if no "/METHOD = " subcommand is used. This procedure presents the results as if
every term in the model had been entered last. The following is an article sent to the ED-STAT discussion group
on April 28, 1993.
In trying to understand how SPSS handles
unbalanced A X B designs, I generated a simulated data set. The following is the summary table using the
UNIQUE (default) option.
Tests of Significance for X using UNIQUE sums of squares |
Source of
Variation SS DF MS F Sig of F WITHIN+RESIDUAL
29961.11 76 394.23 FACTORA
927.96 1 927.96 2.35 .129 FACTORB
2500.39 2 1250.20 3.17 .048 FACTORA
BY FACTORB 3504.32 2 1752.16 4.44 .015 (Model)
9391.52 5 1878.30 4.76 .001 (Total)
39352.63 81 485.83 |
Note that the Sum of Squares for the three
effects sum to 6932.67, somewhat less than the Sum of Squares for the model.
Tests of Significance for X using SEQUENTIAL Sums of Squares |
Source of Variation SS DF MS F
Sig of F WITHIN+RESIDUAL 29961.11
76 394.23 FACTORA 797.17
1 797.17 2.02 .159 FACTORB 5090.03
2 2545.01 6.46 .003 FACTORA BY FACTORB 3504.32 2 1752.16 4.44 .015 (Model) 9391.52
5 1878.30 4.76 .001 (Total) 39352.63
81 485.83 |
The same data analysis using the SEQUENTIAL
method produced the following table.
In this case the SS for the factors add up to
the model SS and the interaction SS is the same as in the UNIQUE method. No problem.
The SS for FACTORA, entered into the equation first, is smaller than
that using the UNIQUE method. Listing
the effects in a different order on the DESIGN produces a different SS when the
SEQUENTIAL method is used.
Tests of Significance for X using SEQUENTIAL Sums of Squares |
Source of Variation SS DF MS F
Sig of F WITHIN+RESIDUAL 29961.11
76 394.23 FACTORB 5747.55
2 2873.77 7.29 .001 FACTORA 139.65
1 139.65 .35 .553 FACTORA BY FACTORB 3504.32 2 1752.16 4.44
.015 WITHIN+RESIDUAL 29961.11
76 394.23 FACTORA BY FACTORB 4780.86 2 2390.43 6.06 .004 FACTORA 2110.27
1 2110.27 5.35 .023 FACTORB 2500.39
2 1250.20 3.17 .048 WITHIN+RESIDUAL 29961.11
76 394.23 FACTORA BY FACTORB 4780.86 2 2390.43 6.06 .004 FACTORB 3682.70
2 1841.35 4.67 .012 FACTORA 927.96
1 927.96 2.35 .129 |
In all cases the final SS corresponds to that
using the UNIQUE method. How can the SS for an effect be smaller than the
UNIQUE SS for that effect, especially if that effect is entered into the model
first? To quote a student "I just
don't understand."
To review what occurred. If the "/METHOD = SEQUENTIAL"
subcommand was specified, the sum of the SS's for the effects was equal to the
SS for the model. The problem was that
the SS for each effect was different depending upon where it was listed in the
"/DESIGN" subcommand. The
size of an effect depended upon which effects were already in the model. An effect could be significant or not
depending upon where it was listed on the "/DESIGN" subcommand. In addition, it can be seen that the
"/METHOD = UNIQUE" produced results as if each term had been listed
last on the "/DESIGN" subcommand.
It all seems quite arbitrary.
Solutions for Dealing with Dependent
Effects
The article solicited many comments. Two of them are presented below:
From: Neil W. Henry
<nhenry@cabell.vcu.edu>
You are not going to like this answer: "That's life." Or another: "Don't use the word
'effect' so much." It carries
connotations that interfere with following what the numbers are doing. And 'unique effect' is even more dangerous. If you like correlations and regression, as
I do, more than ANOVA, consider the following example.
Y = X + W (exactly)
Var(X) = Var(Y) = Var(W) = 1.
Cor(X,W) = ‑.5 .
It follows that
Cor(Y,X) = Cor(Y,W) = .5 .
Consequently, W "when entered into the
model first", _explains [only] 25% of the variance in Y_ . But, since together
X and W "explain" 100% of the variance in Y, W's UNIQUE contribution
to the variance is 75% (i.e. its contribution when added to the model after X
is
already in).
Moral 1: Ain't negative numbers fun?
Moral 2: Don't let other people get your
kicks for you; or push their favorite jargon on you either.
*************************************************
`o^o' * Neil W. Henry
(nhenry@cabell.vcu.edu) *
‑<:>‑
* Virginia Commonwealth University *
_/ \_ * Richmond VA 23284‑2014 *
*(804)367‑1301 (math sciences, 2079
Oliver) *
* 7‑6650
(academic computing, B30 Cabell) *
*************************************************
From:
nichols@spss.com (David Nichols)
To:
dws148f@smsvma.bitnet
Subject: Re:
Unbalanced ANOVA
From a
computational perspective, you can think of UNIQUE sums of squares as the
difference between the total model sums of squares accounted for by a model
including all terms and one including all terms except for the one at issue.
Thus each term in the model is "adjusted" for all other terms, or
entered last in a regression. This will produce sums of squares that do not
generally add up to the total (they will in cases where there are the same
number of cases in each cell of the design). The SEQUENTIAL approach fits an
ordered set of terms one at a time, and this will produce a set of sums of
squares that add up to the total. Since the last term entered is after all
other terms in the model, it will always be the same as with UNIQUE sums of
squares.
In the general
case of data in which there are unequal numbers of cases in the cells of the
design (assuming at least one in each cell‑‑the case of empty cells
becomes substantially more complex), the UNIQUE sums of squares test the same
hypotheses that are tested in balanced designs. In the balanced case, both
types of sums of squares test the same hypotheses and thus reduce to the same
results. In unbalanced designs the statistical hypotheses tested by SEQUENTIAL
sums of squares test hypotheses that depend on the number of observations in
the cells, which is generally not considered a positive characteristic. This is
why UNIQUE sums of squares are the default.
As to the
question of how the sums of squares for a factor can be higher when it is
entered after another factor than when it is entered alone, probably the
easiest way to understand it is to think in terms of part correlations and
multiple regression. Recall that ANOVA and regression involve an identical
mathematical/statistical model. You can easily verify the fact that the sums of
squares accounted for by a term in a regression model at a particular stage is
given by the squared part correlation of that regressor with the dependent
variable, multiplied by the mean corrected total sums of squares. Thus if the
squared part correlation between X and Y with the linear effects of Z taken out
of X is higher than the simple squared correlation of X with Y, the sums of
squares of X fitted after Z will be greater than the sums of squares fitted
before Z (at least in the case where these are the only two predictors).
Another way to
look at this (related to the first) and one that may be more helpful, is to try
to picture it in geometric terms. The part correlation of X with Y, removing
the linear effects of Z from X is equivalent to regressing X on Z, taking the
residuals, and then regressing Y on those residuals. If you look at this in
terms of vectors, you can see that the residual vector can in general be
pointing in any direction and thus may be more nearly coincident with the Y
vector than was the original X vector.
It all depends
on the intercorrelations among the variables. There's an excellent description
of this from a geometrical perspective in the February 1993 _American
Statistician_, Vol. 47, No. 1, beginning on p. 26.
UNEQUAL
CELL SIZES FROM A MULTIPLE REGRESSION VIEWPOINT
A B X DA1 DB1 DB2 DAB1 DAB2 |
1 1 23
1 1 1 1 1 1 1 27
1 1 1 1 1 1 1 35
1 1 1 1 1 1 1 25
1 1 1 1 1 1 2 56
1 1 ‑1 1 ‑1 1 2 45
1 1 ‑1 1 ‑1 1 2 62
1 1 ‑1 1 ‑1 1 3 24
1 ‑2 0
‑2 0 1 3 21
1 ‑2 0
‑2 0 1 3 34
1 ‑2 0
‑2 0 1 3 28
1 ‑2 0
‑2 0 1 3 33
1 ‑2 0
‑2 0 1 3 21
1 ‑2 0
‑2 0 1 3 14
1 ‑2 0
‑2 0 1 3 26
1 ‑2 0
‑2 0 1 3 22
1 ‑2 0
‑2 0 1 3 20
1 ‑2 0
‑2 0 1 3 31
1 ‑2 0
‑2 0 1 3 38
1 ‑2 0
‑2 0 1 3 30
1 ‑2 0
‑2 0 1 3 25
1 ‑2 0
‑2 0 2 1 36
‑1 1 1
‑1 ‑1 2 1 41
‑1 1 1
‑1 ‑1 2 2 43
‑1 1 ‑1
‑1 1 2 2 48
‑1 1 ‑1
‑1 1 2 2 35
‑1 1 ‑1
‑1 1 2 2 41
‑1 1 ‑1
‑1 1 2 2 39
‑1 1 ‑1
‑1 1 2 2 44
‑1 1 ‑1
‑1 1 2 2 51
‑1 1 ‑1
‑1 1 2 3 18
‑1 ‑2 0
2 0 2 3 21
‑1 ‑2 0
2 0 |
An example data
set with unequal cell sizes will be analyzed first with MANOVA using both “/METHODS=UNIQUE”
and “/METHODS=SEQUENTIAL” subcommands.
The results of these analyses will correspond directly to those
discussed earlier in this chapter. The
same data set will be dummy coded and analyzed using a multiple regression
procedure. The results of the two
methods of MANOVA will be duplicated.
The example
data set is presented below. Note that
the cells have unequal n’s and the variables have been dummy coded using the
same recoding system as the two factor ANOVA in the last chapter.
The results of
the MANOVA using the “/METHOD=UNIQUE” subcommand is as follows:
A n a l y s I s o
f V a r I a n c e |
Tests of
Significance for X using UNIQUE sums of squares Source of
Variation SS DF
MS F Sig of F WITHIN CELLS 975.02 26 37.50 A 27.63 1
27.63 .74 .399 B 2604.75 2
1302.38 34.73 .000 A BY
B 425.82 2
212.91 5.68 .009 (Model)
3255.94 5 651.19 17.36 .000 (Total)
4230.97 31 136.48 R‑Squared
= .770 Adjusted R‑Squared
= .725 |
A similar
analysis done using the “/METHOD=SEQUENTIAL” subcommand yields the following
table.
A n a l y s I s o f
V a r I a n c e |
Tests of
Significance for X using SEQUENTIAL Sums of Squares Source of
Variation SS DF MS F Sig of F WITHIN
CELLS 975.02 26
37.50 A 398.82 1
398.82 10.63 .003 B 2431.30 2
1215.65 32.42 .000 A BY
B 425.82 2
212.91 5.68 .009 (Model)
3255.94 5 651.19 17.36 .000 (Total)
4230.97 31 136.48 R‑Squared
= .770 Adjusted R‑Squared
= .725 |
The results are
as expected considering the preceding discussion of the effect of unequal n’s
on the ANOVA. Both the cell means for
the groups and the marginal means from the MANOVA command are presented below.
Var Value Mean Std Dev Cases |
A
1 30.47 11.79
21 B 1
27.50 5.25 4 B 2
54.33 8.62 3 B 3
26.21 6.51 14 A
2 37.90 10.25
11 B 1
38.50 3.53 2 B 2
43.00 5.38 7 B 3
19.50 2.12 2 |
Combined Observed Means for A |
A 1 WGT. 30.47619
UNWGT. 36.01587 2 WGT. 37.90909
UNWGT. 33.66667 Combined
Observed Means for B B 1 WGT. 31.16667
UNWGT. 33.00000 2 WGT. 46.40000
UNWGT. 48.66667 3 WGT. 25.37500
UNWGT. 22.85714 |
Note that the
weighted means are the means obtained using the MEANS command and each marginal
mean is computed by dividing the sum of scores by the number of scores. The unweighted marginal means are computed
by finding the mean of the cell means.
For example, the unweighted marginal mean for a1 would be the
mean of 27.50, 54.33, and 26.21.
REGRESSION
ANALYSIS OF UNEQUAL N ANOVA
The regression
analysis uses the dummy codes of the data.
With unequal cell sizes, the dummy coded variables will not be
orthogonal or uncorrelated. The greater
the difference in cell sizes, the greater the correlations between the dummy
codes. The correlation matrix for the
dummy codes in the example data is presented below.
‑ ‑ Correlation Coefficients
‑ ‑ |
DA1 DB1 DB2 DAB1 DAB2 DA1
1.0000 ‑.4606** .3427 ‑.1910 ‑.2835 DB1
‑.4606** 1.0000 ‑.1796 .5069** .2750 DB2
.3427 ‑.1796 1.0000 .0910 ‑.0823 DAB1
‑.1910 .5069** .0910 1.0000 .0296 DAB2
‑.2835 .2750 ‑.0823 .0296 1.0000 |
The implication
of these correlations is that the different dummy coded variables will share
variance in predicting the dependent variable.
The order of entry of a variable into the regression equation will make
a difference.
The REGRESSION
command appearing below places uses the subcommand “/METHOD=TEST(DA1) (DB1
DB2)(DAB1 DAB2).” The result of this
analysis directly corresponds to the “/METHOD=UNIQUE” subcommand in
MANOVA. The REGRESSION command and
output are presented below.
REGRESSION |
/DEPENDENT x
/METHOD=TEST (da1) (db1 db2) (dab1 dab2) . * * *
* M U L T I P L E R E G R E S S I O N * * * * Hypothesis Tests
Sum of DF Squares Rsq Chg F Sig F
Source 1 27.63040 .00653 .73679 .3985
DA1 2 2604.75272 .61564 34.72919 .0000
DB1 DB2 2 425.82185 .10064 5.67749 .0090
DAB1 DAB2 5 3255.94494
17.36462 .0000 Regression 26 975.02381 Residual 31 4230.96875 Total |
To generate the
ANOVA table of the “METHOD=SEQUENTIAL” subcommand in MANOVA, the “/METHOD=TEST”
subcommand is modified as follows in the REGRESSION command. The resultant output follows.
REGRESSION |
/DEPENDENT x
/METHOD=test (da1)
/METHOD=test (db1 db2)
/METHOD=test (dab1 dab2) Variable(s) Entered on Step Number 1..
DA1 Hypothesis Tests
Sum of DF Squares Rsq Chg F Sig F
Source 1 398.82156 .09426 3.12218 .0874
DA1 1 398.82156 3.12218
.0874 Regression 30 3832.14719 Residual 31 4230.96875 Total Block Number
2. Method: Test
DB1 DB2 Hypothesis Tests Sum of DF Squares Rsq Chg F Sig F
Source 2 2431.30153 .57464 24.29834 .0000
DB1 DB2 3 2830.12309 18.85610
.0000 Regression 28 1400.84566 Residual 31 4230.96875 Total Block Number
3. Method: Test
DAB1 DAB2 Hypothesis Tests
Sum of DF Squares Rsq Chg F Sig F
Source 2 425.82185 .10064 5.67749 .0090
DAB1 DAB2 5 3255.94494 17.36462
.0000 Regression 26 975.02381 Residual 31 4230.96875 Total |
Note that the
order of variables entered into the equation is similar to the default order of
entry in the MANOVA. In the case with
the multiple “/METHOD=TEST” commands the REGRESSION analysis tests the
hypothesis that the addition of the variable or variables predicts a
significant amount of the variance that remains after the preceding variables
have been entered into the equation.
RECOMMENDATIONS
The first
recommendation is to avoid multi-factor ANOVA designs with unequal cell
sizes. This recommendation is rather
glib and does not present a realistic alternative in most situations. Unequal cell sizes are a fact of life in
most real-life experimental designs.
The second
recommendation is to use the default “/METHOD=UNIQUE” when faced with unequal
cell sizes. The anomalous results that
this method sometimes generates can be more easily accepted if the user has an
understanding of how the procedure works.
A colleague presented a table of cell means which common sense would
suggest that the two-factor interaction should be significant. It was not, however, because the cell sizes
were grossly unequal. The colleague had
a difficult time accepting a statistical result that went against common sense.
The final
recommendation is to approach ANOVA using a multiple regression approach. Even if the results are similar to that
using MANOVA, at least the user has more control over the analysis and perhaps
a better understanding of the results.
Chapter
15
Subjects
Crossed With Treatments S X A
The Design
In S X A designs each subject
sees each level of factor A.
These designs are sometimes called repeated measures designs
because there must necessarily be more than one score per subject. Another name for the design is within
subjects design, because all effects are within rather than between
subjects.
With respect to purpose and function, the S
X A design is like the A design.
With respect to the score model,
it is much closer to the A X B design.
RUN NAME LOMBARD EFFECT DATA FOR
S X A EXAMPLE |
DATA LIST
FILE='SXA DAT A'/SUBJECT 1‑3 SOFT 4‑6 MEDIUM 7‑9 LOUD 10‑12 NONE 13‑15. LIST. CORRELATION SOFT
TO NONE. MANOVA SOFT
MEDIUM LOUD NONE /WSFACTORS LOMBARD(4) /PRINT = CELLINFO(MEANS) /WSDESIGN /DESIGN. |
An example of an experiment that employs the S
X A design follows. Suppose a
researcher was interested in choral singing behavior. A phenomenon noted by many choral directors, called the Lombard
effect, is that people tend to sing louder when the people around them are
singing louder. The researcher wishes
to test the extent that this occurs.
The researcher has twelve subjects each sing
the Star Spangled Banner four times, measuring the loudness of their singing
each time. The subjects have on
headphones and are singing along with other voices, played soft, medium, and
loud. A forth condition has no
accompaniment. The subjects were
instructed to try to sing at the same level each time. The order of the loudness of the
accompanying voices is randomized across subjects. An apparatus measures the average decibels of the singing for the
entire song.
SUB SFT MD LD NONE |
1 92
93 93 87 2 81
89 89 79 3 88
93 89 81 4 93
99 99 87 5 82
83 83 79 6 88
85 87 90 7 86
92 89 82 8 82
87 84 81 9 82
88 84 82 10 85
88 88 79 11 79
82 82 76 12 88
92 93 84 |
The Data
The data are arranged with each subject's
data as a row. The first column in the
data file is the subject number.
Subject numbers are not necessary
to run the analysis, but may come in very handy in locating incorrectly entered
data or missing subjects. The next four
columns of data are average decibels for the none, soft, medium and loud
accompaniment conditions.
SPSS commands
As in all previous analyses, the DATA LIST
command reads in the data from the data file and the LIST command gives a
picture of how the computer understands the data. The CORRELATION command gives a picture of how the repeated
measures are correlated.
After the keyword MANOVA, the names of the
four variables used for the measures under the different treatment levels are
listed. There is no BY listed because
there are no between subjects factors.
The subcommand WSFACTORS = is followed by a name for the factor and the
number of levels of that factor in parentheses. In this example the factor name selected was LOMBARD and the
number of levels of this factor was four.
The number of levels must correspond to the number of variables listed
after the MANOVA command.
As in the previous MANOVA commands, the PRINT
= CELLINFO(MEANS) will give an optional table of the means for the four
variables. The analysis will also go on
without the WSDESIGN and DESIGN subcommands, but putting them in avoids a
warning message.
The Correlation Matrix
‑ ‑
Correlation Coefficients ‑
‑ |
SOFT MEDIUM LOUD NONE SOFT
1.0000 .7951** .8647**
.8071** MEDIUM
.7951** 1.0000 .9143** .4824 LOUD
.8647** .9143** 1.0000 .5988* NONE
.8071** .4824 .5988*
1.0000 * ‑
Signif.LE.05 ** ‑ Signif.LE.01 (2‑tailed) |
The correlation matrix showing the linear
relationship between the within subjects variables is presented below.
The Table of Means
The following table is an example of the
output of the optional PRINT command.
Note that the results of the experiment are as predicted, the louder the
accompaniment, the louder the subjects sang, except for a slight reversal from
Medium to Loud. The output has been
reduced in size.
|
Graphs
Cell
Means and Standard Deviations |
Mean Std. Dev.
N 95 percent Conf. Interval ‑
‑ ‑ ‑ ‑ ‑ ‑ ‑ ‑ ‑ ‑
‑ ‑ ‑ ‑
Variable .. SOFT For
entire sample 85.500 4.442
12 82.678 88.322 ‑
‑ ‑ ‑ ‑ ‑ ‑ ‑ ‑ ‑ ‑
‑ ‑ ‑ ‑
Variable .. MEDIUM For
entire sample 89.250 4.827
12 86.183 92.317 ‑
‑ ‑ ‑ ‑ ‑ ‑ ‑ ‑ ‑ ‑
‑ ‑ ‑ ‑
Variable .. LOUD For
entire sample 88.333
4.924 12 85.205 91.462 ‑
‑ ‑ ‑ ‑ ‑ ‑ ‑ ‑ ‑ ‑
‑ ‑ ‑ ‑
Variable .. NONE For
entire sample 82.250 4.070
12 79.664 84.836 |
A simple graph is produced from this
data. Because the graph contains so
little information, a table is usually used to present the results.
The ANOVA Table
The results produced by a within subjects
design has several additional statistics besides the ANOVA table. Each section of the results will now be
analyzed in turn.
Interpretation of Output
The first section of the results is presented
below:
‑ ‑ ‑ ‑ ‑
‑ ‑ ‑ ‑ ‑ ‑ ‑ ‑ ‑ ‑
|
Tests of Between‑Subjects Effects. Tests of Significance for T1 using UNIQUE
sums of squares Source of Variation SS DF MS F
Sig of F WITHIN CELLS 746.17 11 67.83 CONSTANT 357765.33
1 357765.33 5274.18
.000 |
* * * * * * * * * * * * * * * * * A N A L Y S I S O F
V A R I A N C E ‑‑ DESIGN 1 * * * * * * * * * * |
Tests involving 'LOMBARD' Within‑Subject
Effect. Mauchly sphericity test, W = .26143 Chi‑square approx. = 13.04327 with 5 D. F. Significance = .023 Greenhouse‑Geisser Epsilon = .53263 Huynh‑Feldt Epsilon = .60890 Lower‑bound Epsilon = .33333 |
Since there are no between subjects effects,
other than the constant (m), in an S X A design, the information
presented above is not very useful. The
analysis says that the grand mean is significantly different from zero.
The second section of the results is
presented below.
The Mauchly sphericity test is an hypothesis
test of the additional assumptions necessary in a within subjects design. When the significance of this statistic is
below a set level (.05 is standard) then the assumptions necessary to do a
within subjects ANOVA have been violated.
The three Epsilons that follow are corrections to the degrees of freedom
to correct for the violations. Both the
numerator and denominator degrees of freedom of the F test that follows must be
multiplied by one of these values and the significance of the F ratio
revaluated. The Greenhouse-Geisser
Epsilon if fairly conservative, especially for small sample. The Huynh-Feldt Epsilon is an attempt at
correction of the conservatism (Norusis, 1990).
EFFECT .. LOMBARD |
Multivariate Tests of Significance (S = 1,
M = 1/2, N = 3 1/2) Test
Name Value Exact F Hypoth. DF Error
DF Sig. of F Pillais .72457 7.89193
3.00 9.00 .007 Hotellings 2.63064 7.89193
3.00 9.00 .007 Wilks .27543 7.89193 3.00
9.00 .007 Roys .72457 Note..
F statistics are exact. |
Multivariate tests of significance are
sometimes recommended for testing within subjects effects because the
assumptions necessary to perform the test are less restrictive. Four different multivariate tests are performed
and appear as follows:
Note that in this case all of the
multivariate tests give the same Exact F and significance level. These hypothesis tests provide evidence for
the reality of effects. The
interpretation of the effect is the same as the interpretation for a
significant main effect of A in design A.
Tests involving 'LOMBARD' Within‑Subject
Effect. |
AVERAGED
Tests of Significance using UNIQUE sums of squares Source of Variation SS DF MS F
Sig of F WITHIN CELLS 176.00
33 5.33 LOMBARD 358.50
3 119.50 22.41
.000 |
The final table is a univariate test of the A
effect. Note that here the Lombard (A)
main effect is statistically significant.
Interpretation of this significant main effect proceeds identically to
the multivariate test for a significant main effect of A.
In any case the results of this study seem
clear, the louder the accompaniment, the louder the person sang. No accompaniment gave the quietest
singing. There may have been a ceiling
effect as moderate and loud accompaniment gave almost identical singing volume.
Additional Assumptions for
Univariate S X A Designs
To avoid distorting the decision-making
process, it is necessary to add an assumption when taking a univariate approach
to testing for the main effect of A in S X A designs. The assumption is that if an infinite number
of subjects were run in the experiment and scores were collected on all levels
of the within subjects treatments, then the correlations between the different
treatments would all be equal. In
addition, the variances would all be equal.
The sample correlation matrix between the treatment effects was
found in the example SPSS program.
Imagine a very large (infinite) data set and a similar correlation
matrix. The assumption underlying the
univariate repeated measures ANOVA is that all the values in the matrix would
be the same.
Norusis (1990) acknowledges difficulties with
the Mauchly Sphericity Test. For small
samples the test of the preceding assumption is not very powerful and for large
samples it may be too powerful. The
major problem here is that it the goal of the analysis is to accept the null
hypothesis that the correlations are not different.
Deciding to use the univariate or
multivariate approach is often a matter of preference. Winer (1962 p. 283) writes: "It is, however, difficult to formulate
a set of general rules by which one type of analysis will be preferred to the
other. The univariate approach pools
information from repeated measures - this kind of pooling usually carries with
it a set of homogeneity assumptions."
SS, MS, and Expected Mean Squares
(EMS)
SS and MS
As in the preceding designs, the SS for any
effect can be found by squaring the score model terms and summing the
squares. The mean squares may be obtained
by dividing the SS by the appropriate degrees of freedom. The F-ratio is then found by dividing the MS
for that term by an error MS. It is not
directly apparent whether the MS for the A effect (MSA)
should be tested by dividing by the MS for Subjects (MSS) or the MS
for error (MSE). For that
reason a new concept, expected mean squares, will be introduced.
Expected Mean Squares
If an experiment could be run with an
infinite number of subjects in all conditions, then the score model would not
need to be estimated, but would be known.
In that case, hypothesis testing for term effects would not be needed
because it would be known whether the terms in the model were zero or not. Since it is impossible and not very
cost-efficient to run an infinite number of subjects in an experiment,
hypothesis testing is standard operating procedure in science.
In an abstract and mathematical sense the
question of what would the world look like if there were and infinite number of
subjects in each group makes sense. In
fact it is the foundation of hypothesis testing theory. If there were an infinite number of subjects
in each group, the score model terms could be found for each score. The procedure for finding the terms is
conceptually similar, but not directly analogous, to the procedure which has
been used in the preceding chapters. It
is not directly possible to solve an infinite number of simultaneous linear
equations for an infinite number of variables.
A mathematical solution, however, is available. At this point the student is asked to believe.
If the real terms for the score model could
be found, the real variances for all the terms in the model could also be
found. There would be a variance for
each term in the score model. For
example,
Factor Model
Term Variance
A aa σA2
S ss σS2
E eas σE2
In real life, only a sample of subjects
(preferably random) is selected to be included in the experiment rather than an
infinite number of subjects. The terms
in the score model are estimated and the mean squares for each term are
computed. Basically, the means squares
for each term are variances of those estimated values.
Moving back again to the conceptual and
abstract arena, visualize an infinite number of similar experiments being
carried out. The score model terms and
associated mean squares are estimated for each experiment. Each time they are different, but seem to
cluster around one value or another. A
mathematician can find the
(theoretical) mean of the model terms and means squares using a technique
called expected value. The expected
mean square (EMS) is the expected value of the infinite number of observed mean
squares.
Using techniques that will not be covered in
this text, it is possible to solve for the EMS in terms of real variances of
model terms. The results show certain
patterns which have been placed in an algorithm so that given any standard
model, the expected mean squares may be found (Lee, 1975). I have written a computer program that uses
the algorithm presented in Lee so that students may find expected mean squares
for many standard ANOVA designs.
Instructions on how to use the program may be found in Homer &
Tessie.
Using the program the following results were found for design S X A.
Factor Model
Term Variance Mean Square EMS
A aa σA2
MSA σE2 + σAS2 + SσA2
S ss σS2 MSS σE2 + AσS2
E eas σE2 MSE σE2 + σAS2
The rule for selecting an error term to test
the effect of a factor is simple: the
error term (denominator) in the F-ratio is the mean square of the term with an
expected mean square containing all the terms in the expected mean square of
the term to be tested plus one - an error term - σE2.
From the above table it can be seen that MSE is the
appropriate error term. Also note that
there is no error term that could be used to test the subjects main effect.
The output from the program described in Homer
& Tessie differs somewhat from the Expected Mean Squares given above,
because of limitations of the screen output within the programming
language. The correspondence is as
follows:
Actual Expected Mean Squares Program Expected Mean Squares
σE2 + σAS2 + SσA2 σE + σAS + SσA
σE2 + AσS2 σE + AσS
σE2 + σAS2 σE + σAS
Note that the terms are not squared and terms
which follow the σ are
not subscripted. With a little effort,
the two presentation modes may be understood as equivalent.
Chapter
16
Subjects
Crossed With Two Treatments -
S X A X B
The Design
In S X A X B designs
each subject sees each level of all possible combinations of
levels of factors A and B. These designs are a more complex example of repeated measures
or within subjects designs. With respect to purpose and function, the S
X A X B design is like the A X B design.
An example of an experiment that employs an S
X A X B design is a variation of the S X A
experiment described in the previous chapter.
That study investigated the Lombard effect by having subjects sing the
national anthem along with a choral group played through earphones. In this variation of the study, the three
levels of factor A were three different locations in the song. Factor B investigated the effects of
instructions to the subjects. On the
first attempt, subjects were not given any instructions about how loudly to
sing, while on the second attempt subjects were instructed to try to sing at the same level throughout the song.
The Data
The data are arranged with each subject's
data as a row. As before, the first
column in the data file is the subject number. Following this are the six measures taken for each subject. Note that the variables are named as L1P1 to
L2P3 for the loudness and place levels for each subject. The MANOVA subcommand does not care what
these variables are named, the variable names could just as well have been V1
to V6. In this case the variable names
were selected to make life more organized for the data analyzer. The GENDER and EXPER variables are not
needed for this analysis, but will be needed in following chapters.
SUB GND EXP I1P1 I1P2 I1P3 I2P1
I2P2 I2P3 |
2
1 3 93 93 93
87 88 89 3
2 2 81 89 89
79 83 79 4
1 2 88 93 89
81 77 76 5
1 1 93 99 99
87 93 95 6
2 1 82 83 83
79 79 79 7
2 2 88 85 87
90 84 83 8
1 3 86 92 89
82 84 86 9
1 2 82 87 84
81 81 79 10 2
2 82 88 84 82
83 83 11 1
1 85 88 88 79
77 83 12 1
3 79 82 82 76
79 82 13 1
1 88 92 93 84
84 88 14 1
3 82 83 83 78
82 80 15 2
1 81 85 86 78
80 80 16 1
1 85 85 87 81
84 87 17 1
3 91 88 85 87
77 78 18 2
3 78 79 79 71
73 78 19 1
3 89 95 102 91
99 98 20 1
2 85 84 86 85
79 78 21 1
3 90 94 93 88
89 90 22 2
3 86 88 89 83
86 87 23 1
2 89 90 90 85
85 87 24 2
2 85 86 87 75
81 85 25 1
1 83 83 78 72
74 73 26 1
1 79 80 80 79
77 78 27 2
1 80 81 81 79
77 78 28 1
2 77 82 84 71
72 76 |
SPSS commands
As in all previous analyses, the DATA LIST
command reads in the data from the data file.
The VALUE LABELS command will make interpretation of output easier. As before, the LIST command gives a picture
of how the computer understands the data and the CORRELATION command gives a
picture of how the repeated measures are correlated.
After the keyword MANOVA, the names of the
six variables used for the measures under the different treatment levels are
listed. Because there are no between
subjects factors, the keyword BY is not used.
The subcommand WSFACTORS = is followed by a name for both the A
and B factors with the number of levels of that factor in parentheses
following the factor name. The slowest
moving variable is listed first and the fastest moving variable is listed last. In this example the factor names selected
were INSTRUCT and PLACE, the former reflecting whether or not instructions were
given to the subject to control for the Lombard effect and the latter
reflecting the position in the song.
The order of the variables following the MANOVA command determines the
order of factor names following the WSFACTORS subcommand. In this case, because the three position
measures under the no instruction condition appeared first and the second three
were under voice control instructions, the order of variables on the WSFACTORS
subcommand had to be INSTRUCT and then PLACE.
Changing the order of these factors would give very different
results. The product of the number of
levels of all factors listed after WSFACTORS must correspond to the number of
variables listed after the MANOVA command.
In this case there are two times three or six variables.
As in the previous MANOVA commands, the PRINT
= CELLINFO(MEANS) will give an optional table of the means for the six
variables. The analysis will also
proceed without the WSDESIGN and DESIGN subcommands, but putting them in avoids
a warning message.
The Correlation Matrix
The correlation matrix showing the linear
relationship between the within subjects variables is presented below. Note that the different measures are highly
correlated, indicating that the experiment gains considerable power by using a
within rather than between subjects design.
‑ ‑ Correlation Coefficients
‑ ‑ |
I1P1 I1P2 I1P3 I2P1
I2P2 I2P3 I1P1
1.0000 .8270** .7578**
.8091** .6620** .6519** I1P2
.8270** 1.0000 .9020** .6882** .7627** .7323** I1P3
.7578** .9020** 1.0000 .7260** .8716** .8696** I2P1
.8091** .6882** .7260** 1.0000 .7667** .6606** I2P2
.6620** .7627** .8716**
.7667** 1.0000 .9224** I2P3
.6519** .7323** .8696**
.6606** .9224** 1.0000 0* ‑
Signif. LE .05 ** ‑ Signif.
LE .01 (2‑tailed) |
The Table of Means
Cell Means and Standard Deviations |
Mean Std. Dev. N 95 percent
Conf. Interval Variable .. I1P1 For entire sample 84.704
4.513 27 82.918 86.489 Variable .. I1P2 For entire sample 87.185 5.054
27 85.186 89.184 Variable .. I1P3 For entire sample 87.037 5.626
27 84.811 89.263 Variable .. I2P1 For entire sample 81.111 5.423
27 78.966 83.256 Variable .. I2P2 For entire sample 81.741 6.010
27 79.363 84.118 Variable .. I2P3 For entire sample 82.778
5.976 27 80.414 85.142 |
The following table is an example of the
output of the optional PRINT subcommand under the MANOVA command. Note that the means must be collapsed over
variables to find effects. For example,
in order to find the size of the INSTRUCT factor, the mean of the first three
means would have to be compared with the mean of the last three means. The POSITION main effect could likewise be
found by comparing the means of I1P1 and I2P1, I1P2 and I2P2, and I1P3 and
I2P3.
Graphs
|
A graph of the interaction of A and B is generated from the
data if the interaction effect is found to be significant. The following presents a graph of the above
means.
The ANOVA Table
* * * * * * * * A N A L Y S I
S O F V A R I A N C E * * * * * * |
Tests
involving 'INSTRUCT' Within‑Subject Effect. Tests of
Significance for T2 using UNIQUE sums of squares Source
of Variation SS DF MS F Sig of F WITHIN
CELLS 280.27 26
10.78
INSTRUCT
795.56 1 795.56 73.80 .000 |
Like the S X A design, the
results produced by a within subjects design has several additional statistics
besides the ANOVA table. Each section
of the results will now be analyzed in turn.
Interpretation of Output
The first section of the results is presented
below:
* * * * A N A L Y S I S O F
V A R I A N C E * * * * |
Tests of Between‑Subjects Effects. Tests of Significance for T1 using UNIQUE
sums of squares Source
of Variation SS DF
MS F Sig of F WITHIN CELLS 3753.11
26 144.35 CONSTANT 1145593.39
1 1145593.4 7936.20
.000 |
Since there are no between subjects effects,
other than the constant (m), in an S X A X B design, the
information presented above is not very useful. The analysis says that the grand mean is significantly different
from zero.
Because the INSTRUCT main effect has only two
levels, univariate and multivariate tests of significance do not differ. In this case the SPSS program does only the
univariate test with the following result.
Note that the INSTRUCT main effect is significant.
The third section of the results is presented
below.
* *
* * * * A N A L Y S I S O F V A R I A N C E * * * * * * * |
Tests involving 'PLACE' Within‑Subject Effect. Mauchly sphericity test, W = .51990 Chi‑square approx. = 16.35277 with 2 D. F. Significance = .000 Greenhouse‑Geisser Epsilon = .67563 Huynh‑Feldt Epsilon = .69951 Lower‑bound Epsilon = .50000 AVERAGED Tests of
Significance that follow multivariate tests are equivalent to univariate or split‑plot
or mixed‑model approach to repeated measures. Epsilons may be used to
adjust d.f. for the AVERAGED results. |
* * * * A N A L Y S I S O F
V A R I A N C E * * * * |
EFFECT
.. PLACE Multivariate
Tests of Significance (S = 1, M = 0, N = 11 1/2) Test
Name Value Exact F Hypoth. DF Error DF
Sig. of F Pillais .23891
3.92384 2.00 25.00
.033 Hotellings .31391 3.92384 2.00
25.00 .033 Wilks .76109
3.92384 2.00 25.00
.033 Roys .23891 Note.. F statistics are exact. |
As in the S X A design, the
Mauchly sphericity test is an hypothesis test of the additional assumptions
necessary in a within subjects design when the factors have more than two
levels. Because the PLACE factor has
more than two level, these assumptions are tested by SPSS. Note that in this case the significance of
the Mauchly sphericity test seriously questions the assumptions necessary to
test the PLACE main effect.
Multivariate tests of significance are
sometimes recommended for testing within subjects effects because the
assumptions necessary to perform the test are less restrictive. Four different multivariate tests are
performed and appear as follows:
Note that in this case all of the
multivariate tests give the same Exact F and significance level. These hypothesis tests provide evidence for
the reality of effects. The
interpretation of the effect is the same as the interpretation for a
significant main effect of A in design A.
The final table is a univariate test of the A
effect. Note that here the Lombard (A)
main effect is statistically significant.
Interpretation of this significant main effect proceeds identically to
the multivariate test for a significant main effect of A.
* * * * A N A L Y S I S O F
V A R I A N C E * * * * |
Tests involving 'PLACE' Within‑Subject
Effect. AVERAGED
Tests of Significance using UNIQUE sums of squares Source of Variation SS
DF MS F
Sig of F WITHIN CELLS 448.89
52 8.63 PLACE 119.11
2 59.56 6.90
.002 |
In any case the results of this study seem
clear, if the accompaniment was louder at some position in the song, the person
sang louder.
Just as in the A X B design,
the S X A X B design has an A by B
interaction. This term is tested by
both multivariate and univariate analyses in SPSS. In addition, the sphericity test is done to test the assumptions
necessary to test this term. The
following presents the combined analysis.
Note the Mauchly sphericity test was not significant and both the
multivariate and univariate tests of the INSTRUCT BY PLACE interaction were
significant. The interaction was
probably mostly due to position 2 in the song where the instructions to control
loudness were relatively more effective.
* * * * A N A L Y S I S
O F V A R I A N C E * * * * |
Tests involving 'INSTRUCT BY PLACE' Within‑Subject
Effect. Mauchly sphericity test, W = .87792 Chi‑square approx. = 3.25506 with 2 D. F. Significance = .196 Greenhouse‑Geisser Epsilon = .89120 Huynh‑Feldt Epsilon = .95230 Lower‑bound Epsilon = .50000 * * * * A N A L Y S I S O F
V A R I A N C E * * * * EFFECT .. INSTRUCT BY PLACE Multivariate Tests of Significance (S = 1,
M = 0, N = 11 1/2) Test
Name Value Exact F Hypoth. DF Error DF Sig. of F Pillais .28144 4.89584 2.00
25.00 .016 Hotellings .39167 4.89584 2.00
25.00 .016 Wilks .71856
4.89584 2.00 25.00 .016 Roys .28144 Note.. F statistics are exact. * * * * A N A L Y S I S O F
V A R I A N C E * * * * Tests involving 'INSTRUCT BY PLACE' Within‑Subject
Effect. Source
of Variation SS DF
MS F Sig of F WITHIN CELLS 166.91 52 3.21 INSTRUCT BY PLACE 23.75 2 11.88 3.70
.031 |
Additional Assumptions for
Univariate S X A X B Designs
To avoid distorting the decision-making
process, it is necessary to add assumptions when taking a univariate approach
to testing for the main effects and interaction effects in an S X A X B design. The assumptions identical to testing the
main effect of A in an S X A design, namely that if an
infinite number of subjects were run in the experiment and scores were
collected on all levels of the within subjects treatments, then the
correlations between the different treatments or combination of treatments
would all be equal. repeated measures
ANOVA is that all the values in the matrix would be the same.
SS, MS, and Expected Mean Squares
(EMS)
SS and MS
As in the preceding designs, the SS for any
effect can be found by squaring the score model terms and summing the
squares. The mean squares may be
obtained by dividing the SS by the appropriate degrees of freedom. The F-ratio is then found by dividing the MS
for that term by an error MS. As in the
S X A design it is not directly apparent whether the MS for the A
effect (MSA) should be tested by dividing by the MS for Subjects (MSS),
the MS for A by Subjects (MSAS), or the MS for error (MSE). Expected mean squares may be used to find
the correct error term.
Expected Mean Squares
The expected mean squares can be found by
using the program described in Homer & Tessie. The results of running the program for design A X B X S has been previously
presented in a text box. Modifying the
order of the terms yields the following:
Factor Model
Term Variance
A aa σA2
S ss σS2
E eas σE2
In real life, only a sample of subjects (preferably
random) is selected to be included in the experiment rather than an infinite
number of subjects. The terms in the
score model are estimated and the mean squares for each term are computed. Basically, the means squares for each term
are variances of those estimated values.
Moving back again to the conceptual and
abstract arena, visualize an infinite number of similar experiments being
carried out. The score model terms and
associated mean squares are estimated for each experiment. Each time they are different, but seem to
cluster around one value or another. A
mathematician can find the
(theoretical) mean of the model terms and means squares using a technique
called expected value. The expected
mean square (EMS) is the expected value of the infinite number of observed mean
squares.
Using techniques that will not be covered in
this text, it is possible to solve for the EMS in terms of real variances of
model terms. The results show certain
patterns which have been placed in an algorithm so that given any standard
model, the expected mean squares may be found (Lee, 1975).
The following results were found for
design S X A X B.
A‑ σE + BσAS + SBσA
AS‑ σE + BσAS
B‑ σE + AσBS + SAσB
BS‑ σE + AσBS
AB‑ σE + σABS + SσAB
ABS‑ σE + σABS
S‑ σE + ABσS
The rule for selecting an error term to test
the effect of a factor can now be recalled.
The error term (denominator) in the F-ratio is the mean square of the
term with an expected mean square containing all the terms in the expected mean
square of the term to be tested plus one - an error term - σE2. From the above table it can be seen that MSAS
is the appropriate error term for the A main effect. In a like manner, the MSBS is
used for B and MSABS is used for AB. As before, there is no error term that
could be used to test the subjects main effect.
The WITHIN CELLS term in the SPSS output
corresponds to the appropriate error term. For example, the MSABS term corresponds to the WITHIN
CELLS term in the test of the AB interaction effect.
Chapter
17
Mixed Designs
- S ( A ) X B
The Design
In S ( A ) X B designs each subject sees each
level of B and appears under one
and only one level of A. These
designs are called mixed designs because they contain both within and between
subjects factors. With respect to
purpose and function, the S ( A ) X B design is like the A
X B or S X A X B designs.
An example of an experiment that employs an S ( A ) X B design is a
variation of the S X A X B experiment described in the
previous chapter. That study
investigated the Lombard effect by having subjects sing the national anthem
along with a choral group played through earphones. In this variation of the study, the three levels of factor B
were three different locations in the song.
Factor A investigated the effects of the subjects previous choral
experience. In this blocking factor,
subjects either had no choral experience, some experience, or a great deal of
experience. The astute student might
realize that this data is identical to the first half the data presented in the
last chapter.
The Data
The data are arranged with each subject's
data as a row. As before, the first
column in the data file is the subject number. Following this are the three measures taken for each
subject. Note that the variables are
named I1P1 to I1P3 for the three places in the song that readings were taken
for each subject. In this analysis only
the condition when the subjects are given no instructions about how loudly to
sing is used.
SUB GND EXP I1P1 I1P2 I1P3 |
2
1 3 93 93 93 3
2 2 81 89 89 4
1 2 88 93 89 5
1 1 93 99 99 6
2 1 82 83 83 7
2 2 88 85 87 8
1 3 86 92 89 9
1 2 82 87 84 10 2
2 82 88 84 11 1
1 85 88 88 12 1
3 79 82 82 13 1
1 88 92 93 14 1
3 82 83 83 15 2
1 81 85 86 16 1
1 85 85 87 17 1
3 91 88 85 18 2
3 78 79 79 19 1
3 89 95 102 20 1
2 85 84 86 21 1
3 90 94 93 22 2
3 86 88 89 23 1
2 89 90 90 24 2
2 85 86 87 25 1
1 83 83 78 26 1
1 79 80 80 27 2
1 80 81 81 28 1
2 77 82 84 |
SPSS commands
After the keyword MANOVA, the names of the
three variables used for the within subjects factor are listed. Following the keyword BY, the between
subjects factors are listed, followed by the beginning and ending levels in
parentheses. Here the between subjects
factor is EXPER with factor levels starting at 1 and ending at 3.
The subcommand WSFACTORS = is followed by a
name for the B factor with the number of levels of that factor in
parentheses following the factor name.
This is done in a manner identical to the design specification if a S
X B design had been employed. In
this case the factor is labeled PLACE and has three levels.
As in the previous MANOVA examples, the PRINT
= CELLINFO(MEANS) will give an optional table of the means for the three place
variables at the three levels of the EXPER variable. As before, the analysis will go on without the WSDESIGN and
DESIGN subcommands, but putting them in avoids a warning message.
RUN NAME ANOVA EXAMPLE MIXED DESIGN ANALYSIS |
DATA LIST FILE='ANOVA4 DAT A'/SUBJECT GENDER EXPER 1‑6
I1P1 I1P2 I1P3 7‑15. VALUE LABELS GENDER 1 'MALE' 2 'FEMALE'/ EXPER 1 'NONE' 2 'SOME' 3 'LOTS'. LIST. MANOVA I1P1 TO I1P3 BY EXPER (1,3) /WSFACTORS PLACE (3) /WSDESIGN /PRINT CELLINFO(MEANS) /DESIGN. |
The Table of Means
Cell Means and Standard Deviations |
Variable .. I1P1 FACTOR CODE
Mean Std. Dev 95 percent Conf. Interval EXPER NONE
84.000 4.387 9
80.627 87.373 EXPER SOME
84.111 3.951 9
81.074 87.148 EXPER LOTS
86.000 5.339 9
81.896 90.104 For entire sample 84.704 4.513
27 82.918 86.489 Variable .. I1P2 EXPER NONE
86.222 6.016 9
81.598 90.847 EXPER SOME
87.111 3.333 9
84.549 89.673 EXPER LOTS
88.222 5.783 9
83.777 92.668 For entire sample 87.185 5.054
27 85.186 89.184 Variable .. I1P3 EXPER NONE
86.111 6.679 9
80.977 91.245 EXPER
SOME 86.667 2.345 9 84.864 88.469 EXPER LOTS
88.333 7.053 9
82.912 93.755 For entire sample 87.037 5.626
27 84.811 89.263 |
The following table is an example of the output
of the optional PRINT subcommand under the MANOVA command for a mixed
design.
|
Graphs
A graph of the interaction of A and B is generated from the
data if the interaction effect is found to be significant. Here the Song Position by Choral Experience
interaction is not significant, but the graph is presented anyway.
The ANOVA Table
The complexity of the results of the S ( A ) X B design falls between the A X B and S X A X B designs. As before, each section of the results will
now be analyzed.
Interpretation of Output
The first section of the results is presented
below:
* * * * A N A L Y S I S O F
V A R I A N C E * * * * |
Tests of Between‑Subjects Effects. Tests of Significance for T1 using UNIQUE
sums of squares Source
of Variation SS DF
MS F Sig of F WITHIN CELLS 1714.37 24 71.43 EXPER 62.91
2 31.46 .44
.649 |
The result of the analysis of the
between-subjects effect, here the main effect of A, is presented first.
In the above example, the main effect of EXPER (Choral Experience) was
not significant.
The Mauchly sphericity test is for the within
subjects factor ( PLACE ) is presented next.
In this case this significance test is again significant and the tests
of the within subjects factors must proceed with some consideration about
whether the assumptions have been violated.
Multivariate tests of significance are
sometimes recommended for testing within subjects
* * * * * * * A N A L Y S I S
O F V A R I A N C E * * * * * * |
Tests involving 'PLACE' Within‑Subject
Effect. Mauchly sphericity test, W = .75133 Chi‑square approx. = 6.57602 with 2 D. F. Significance = .037 Greenhouse‑Geisser Epsilon = .80085 Huynh‑Feldt Epsilon = .92074 Lower‑bound Epsilon = .50000 AVERAGED Tests of Significance that follow
multivariate tests are equivalent to univariate or split‑plot
or mixed‑model approach to repeated measures. Epsilons may be used to adjust d.f. for the
AVERAGED results. |
The multivariate and univariate tests for the
B and A X B effects are presented next. Note that all of the multivariate tests give
the same Exact F and significance level. The univariate tests, while providing a slightly different
significance level, make the same decisions as the multivariate tests.
* * * * A N A L Y S I S O F V A R I A N C E *
* * * |
EFFECT .. EXPER BY PLACE Multivariate
Tests of Significance (S = 2, M = ‑1/2, N = 10 1/2) Test
Name Value Approx. F Hypoth.
DF Error DF sig. of F Pillais
.02538 .15421 4.00 48.00 .960 Hotellings .02598 .14292 4.00 44.00 .965 Wilks
.97465 .14860 4.00 46.00 .963 Roys
.02436 Note.. F statistic for WILK'S Lambda is
exact. * *
* * A N A L Y S I S O F V A R I A N C E * * * * EFFECT .. PLACE Multivariate
Tests of Significance (S = 1, M = 0, N = 10 1/2) Test
Name Value Exact F Hypoth. DF Error DF
Sig. of F Pillais .44336 9.15950 2.00
23.00 .001 Hotellings .79648 9.15950 2.00
23.00 .001 Wilks .55664
9.15950 2.00 23.00 .001 Roys .44336 Note.. F statistics are exact. * *
* * A N A L Y S I S O F V A R I A N C E * * * * Tests involving 'PLACE' Within‑Subject
Effect. AVERAGED
Tests of Significance using UNIQUE sums of squares Source of Variation SS
DF MS F
Sig of F WITHIN CELL 237.41
48 4.95 PLACE 104.62
2 52.31 10.58
.000 EXPER BY PLACE 1.98 4 .49 .10 .982 |
The results of the study seem clear. At positions in the national anthem where
the choir sang louder, the person singing along sang louder. Experience in choral singing had no
significant effect as either a main effect or interaction.
Expected Mean Squares (EMS)
The expected mean squares can be found by
using the program described in Appendix B.
The results of running the
program for design S ( A ) X B has been previously
presented in a text box. Modifying the
order of the terms yields the following:
A‑ σE + BσS + SBσA
S-A σE + BσS
B‑ σE + σBS + SAσB
AB‑ σE + σBS + SσAB
BS‑A σE + σBS
The rule for selecting an error term to test
the effect of a factor can now be recalled.
The error term (denominator) in the F-ratio is the mean square of the
term with an expected mean square containing all the terms in the expected mean
square of the term to be tested plus the error term - σE2. From the above table it is seen that MSBS
is the appropriate error term for the B main effect and AB
interaction effect. The WITHIN CELLS
term in the SPSS output corresponds to the appropriate error term. For example, the MSABS term
corresponds to the WITHIN CELLS term in the test of the B and AB
interaction effect.
Chapter
18
Three Factor
ANOVA
This chapter will focus on four designs which
serve the same function, to test the effects of three factors
simultaneously. The designs which will
be studied include:
S X A X B X C
S ( A ) X B X C
S ( A X B ) X C
A X B X C
Since the naming of the factors is arbitrary,
these designs include all possible three factor designs. In a departure from the last few chapters,
the similarities of these designs will first be studied, followed by the
differences. The advantages and disadvantages
of each will be then be presented.
Effects
The function of the four designs given above
is to test for the reality of three kinds of effects, main, two-way interaction, and three-way interaction. Although the first two have been described
in detail in earlier chapters, the different forms of the effects will be
discussed. The three-way interaction
will be discussed in detail.
A study of effects begins with a table of
means. This table might be constructed
by averaging over subjects in any number of ways, depending upon the
design. An example of a table of means
follows.
c1
c2
b1 b2 b3 b1 b2
b3
┌────┬────┬────┐ ┌────┬────┬────┐
a1 │
5 │
5 │
5 │ 5 │
6 │
7 │
8 │ 7
├────┼────┼────┤
├────┼────┼────┤
a2 │
7 │
7 │
7 │ 7 │
6 │
5 │
4 │ 5
└────┴────┴────┘
└────┴────┴────┘
6 6 6 6 6 6 6 6
Main Effects
Main effects are found in a manner analogous
to finding main effects in a two factor design, except that the data must be
collapsed over two other effects rather than one. In the case of a three factor experiment, there will be three
main effects, one for each factor, A, B, and C
.For example, in order to find the main
effect of factor A, one must find the mean of each level of A,
collapsing over levels of B and C. In the above example _a.. = ( _a11 + _a12 + ... +
_ABC) / BC. Where a=1, B=3,
and C=2, _1.. = ( _111
+ _112 + _121 + _122 + _131 + _132
) / 6 = ( 5 + 5 + 5 + 6 + 7 + 8 ) / 6 = 6. Likewise, _2.. = 6 and there would be no main effect
of A, because these values are similar.
In a like manner, _.1. = _.2.
= _.3. = 6 and there would
be no main effect of B. From the
table above it can be seen that _..1 = _..2 = 6 and
there would be no main effect of C.
Thus, this table is an example of a three factor experiment where no
main effects would be found.
Two-Way Interactions
Each combination of two factors produces a
two-way interaction by collapsing over the third factor. The three two-way interactions are
interpreted just like the single two-way interaction would be in an A X B
design.
By collapsing over the C factor, the AB
interaction yields the following table and graph. Note that an AB interaction is present because the simple
main effect of B does changes over levels of A, in one instance
increasing with B and the other decreasing. This table also clearly illustrates the lack of an A or B
main effect.
|
b1 b2 b3
┌────┬────┬────┐
a1 │ 5.5│ 6 │ 6.5│ 6
├────┼────┼────┤
a2 │ 6.5│ 6 │ 5.5│ 6
└────┴────┴────┘
6 6 6 6
|
By collapsing over the B factor, the AC
interaction produces the following table and graph. The cells in the table
reproduce the numbers which appeared as row means in the full table. In this case there is an AC
interaction present.
c1 c2
┌────┬────┐
a1 │
5 │
7 │ 6
├────┼────┤
a2 │
7 │
5 │ 6
└────┴────┘
6 6 6
|
By collapsing over the A factor, the BC
table and graph are produced. The
numbers in the graph appear as row means on the separate tables in the original
data. In this case the interaction is
absent.
b1 b2 b3
┌────┬────┬────┐
c1 │
6 │
6 │
6 │ 6
├────┼────┼────┤
c2 │
6 │
6 │
6 │ 6
└────┴────┴────┘
6 6 6 6
Three-Way Interaction
The three-way interaction, ABC, is a
change in the simple two-way interaction over levels of the third factor. A simple two-way interaction is a two-way
interaction at a single level of a third factor. For example, going back to the original table of means in this
example, the simple interaction effect of AB at c1 would be
given in the means in the left-hand boxes.
The same simple interaction at c2 would be given in the
right-hand boxes.
A change in the simple two-way interaction
refers a change in the relationship of the lines. If in both simple two-way interactions the lines were parallel,
no matter what the orientation, there would be no three-way interaction. Similarity, if the lines in the simple
two-way interactions intersected at the same angle, again no matter what the
orientation, there would be no three-way interaction.
In the case of the example data, graphed
below, the orientation of the lines comprising the simple interactions changes
from parallel to non-parallel from one graph to the other. In this case a three-way interaction would
exist. It may or may not be significant
depending upon the size of the error term.
|
|
Additional Examples
All Effects Significant
The following table of means was constructed
such that all effects would be significant, given that the error terms were
small relative to the size of the effects.
c1
c2
b1 b2 b3 b1 b2
b3
┌────┬────┬────┐ ┌────┬────┬────┐
a1 │
4 │
5 │
6 │ 5 │
5 │
5 │
5 │ 5
├────┼────┼────┤
├────┼────┼────┤
a2 │
7 │
8 │
9 │ 8 │
6 │
5 │
4 │ 5
└────┴────┴────┘
└────┴────┴────┘
|
5.5 6.5 7.5 6.5
The A X B interaction in table
and graph form follow:
b1 b2 b3
┌────┬────┬────┐
a1 │ 4.5│ 5 │ 5.5│ 5
├────┼────┼────┤
a2 │ 6.5│ 6.5│ 6.5│ 6.5
└────┴────┴────┘
5.5 5.75 6 5.75
The same is now done for the A X C interaction:
|
c1 c2
┌────┬────┐
a1 │
5 │
5 │ 5
├────┼────┤
a2 │
8 │
5 │ 6.5
└────┴────┘
6.5 5 5.75
In a similar fashion the table and graph for
the B X C interaction:
|
b1 b2 b3
┌────┬────┬────┐
c1 │ 5.5│ 6.5│ 7.5│ 6.5
├────┼────┼────┤
c2 │ 5.5│ 5 │ 4.5│ 5
└────┴────┴────┘
5.5 5.75 6 5.75
Finally the graph of the three-way
interaction is given.
|
|
Example 3 - B, AC, and BC
Selecting a somewhat arbitrary combination of
effects, one could ask what table of means could produce a combination of
effects such that B, AC, and BC would possibly be
significant and all other effects would not be significant. The following tables are one solution.
c1
c2
b1 b2 b3 b1 b2
b3
┌────┬────┬────┐ ┌────┬────┬────┐
a1 │
4 │
5 │
6 │ 5 │
7 │
7 │
7 │ 7
├────┼────┼────┤
├────┼────┼────┤
a2 │
6 │
7 │
8 │ 7 │
5 │
5 │
5 │ 5
└────┴────┴────┘
└────┴────┴────┘
5 6 7 6 6
6 6 6
|
The presence of a B main effect and
the lack of an A main effect and AB interaction is seen in the
following table and graph.
b1 b2 b3
┌────┬────┬────┐
a1 │ 5.5│ 6 │ 6.5│ 6
├────┼────┼────┤
a2 │ 5.5│ 6 │ 6.5│ 6
└────┴────┴────┘
5.5 6 6.5 6
|
The AC interaction is seen in the
following.
c1 c2
┌────┬────┐
a1 │
5 │
7 │ 6
├────┼────┤
a2 │
7 │
5 │ 6
└────┴────┘
6 6 6
The BC interaction is seen in the
following.
|
b1 b2 b3
┌────┬────┬────┐
c1 │
5 │
6 │
7 │ 6
├────┼────┼────┤
c2 │
6 │
6 │
6 │ 6
└────┴────┴────┘
5.5 6 6.5 6
The ABC three-way interaction is not
significant because the simple interaction of AB does not change over levels of
C. In this case the lines are parallel
in both cases.
|
|
Two More Examples
Because a three-way interaction does not
always appear as intuitive to students, two additional examples three-way
interactions are now given. In the
first case, the three-way interaction is not significant because the
relationships between the lines in the simple interactions do not change. In the second example, only the three-way
interaction is significant.
c1
c2
b1 b2 b3 b1 b2
b3
┌────┬────┬────┐ ┌────┬────┬────┐
a1 │
5 │
5 │
5 │ 5 │
7 │
7 │
7 │ 7
├────┼────┼────┤
├────┼────┼────┤
a2 │
5 │
7 │
9 │ 7 │
3 │
5 │
7 │ 5
└────┴────┴────┘
└────┴────┴────┘
5 6 7 6 5 6 7 6
The reader should verify that in the above
example there might be a significant main effect of B, an AB interaction,
and an AC interaction, but no other effects could be significant.
c1
c2
b1 b2 b3 b1 b2
b3
┌────┬────┬────┐ ┌────┬────┬────┐
a1 │
4 │
6 │
8 │ 6 │
8 │
6 │
4 │ 6
├────┼────┼────┤
├────┼────┼────┤
a2 │
8 │
6 │
4 │ 6 │
4 │
6 │
8 │ 6
└────┴────┴────┘
└────┴────┴────┘
6 6 6 6 6 6 6 6
In the above example, only the three-way
interaction could be significant. There
could be no other significant effects.
Expected Mean Squares
As describe previously, each term in the
score model has expected mean squares terms which determine which mean square
term is used as an error term to test the significance of an effect. The same four designs are used to illustrate
expected mean squares, but the terms are rearranged to show how the various
terms in each model are tested.
A X B X C
A‑ όE + BCόA
B‑ όE + ACόB
C‑ όE + ABόC
AB‑ όE + CόAB
AC‑ όE + BόAC
BC‑ όE + AόBC
ABC‑
όE + όABC
S-ABC σE
S ( A X B ) X C
A‑ όE + CόS + SBCόA
B‑ όE + CόS + SACόB
AB‑ όE + CόS + SCόAB
S‑AB όE + CόS
C‑ όE + όCS + SABόC
AC‑ όE + όCS + SBόAC
BC‑ όE + όCS + SAόBC
ABC‑ όE + όCS + SόABC
CS‑AB όE + όCS
S ( A ) X B X C
A‑ όE + BCόS + SBCόA
S‑A όE + BCόS
B‑ όE + CόBS + SACόB
AB‑ όE + CόBS + SCόAB
BS‑A όE + CόBS
C‑ όE + BόCS + SABόC
AC‑ όE + BόCS + SBόAC
CS‑A όE + BόCS
BC‑ όE + όBCS + SAόBC
ABC‑
όE + όBCS + SόABC
BCS‑A
όE + όBCS
S
X A X B X C
A‑ όE + BCόAS + SBCόA
AS‑ όE + BCόAS
B‑ όE + ACόBS + SACόB
BS‑ όE + ACόBS
C‑ όE + ABόCS + SABόC
CS‑ όE + ABόCS
S‑ όE + ABCόS
AB‑ όE + CόABS + SCόAB
ABS‑
όE + CόABS
AC‑ όE + BόACS + SBόAC
ACS‑ όE + BόACS
BC‑ όE + AόBCS + SAόBC
BCS‑
όE + AόBCS
ABC‑
όE + όABCS + SόABC
ABCS‑
όE + όABCS
Tests of Significance
Error Terms
The error term used to test an effect will
differ depending upon the design of the experiment. As can be seen from the above EMS for the various designs, if a
term to be tested has any factor which is crossed with subjects, then the
interaction with subjects is used as an error term. For example, in the S(AXB)XC design,
the AB term is tested using the S-AB or subjects mean square,
while the ABC term is tested using the CS-AB mean square.
SPSS Output
The greater the number of factors that are
crossed with subjects (S), the more output the MANOVA package within
SPSS will provide. For every factor
crossed with subjects the program does both a test of the assumptions (Mauchly
Sphericity Test) and three multivariate tests of significance in addition to
the univariate tests of significance.
The output from an SXAXBXC design can be
somewhat forbidding. An example of an
experiment employing each of the functional three-factor designs will now be
presented in addition to the data file, MANOVA commands, and example output.
Examples
S ( A X B X C)
S ( A X B ) X C
S ( A ) X B X C
In its full form, the experiment designed to
test the Lombard effect in choral singers neatly fulfill this design. Factor A was choral singing
experience - None, Some, or Lots. Each
subject sang the national anthem twice, Factor B, the first time with no
instructions and the second with instructions to try to control vocal
intensity. Singing loudness (in
decibels) was measured at three different points (Factor C) in the song.
The data has been presented in its full form
in a previous chapter (Design SXAXB), so it is not necessary
to present it again. The MANOVA
commands necessary to do the analysis is given below. Note that the only difference between this set of commands and
the set of commands do the SXAXB analysis the BY EXPER (1,3) on the
first line following the MANOVA command.
MANOVA I1P1 TO I2P3 BY EXPER (1,3) |
/WSFACTORS INSTRUCT (2)
PLACE (3) /WSDESIGN /PRINT CELLINFO(MEANS) /DESIGN. |
The means and standard deviations are
presented first.
Cell Means and Standard Deviations |
Variable .. I1P1 FACTOR CODE Mean Std.
Dev. N 95 percent Conf. Interval EXPER
NONE 84.000 4.387 9
80.627 87.373 EXPER
SOME 84.111 3.951 9
81.074 87.148 EXPER
LOTS 86.000 5.339 9
81.896 90.104 For entire sample 84.704 4.513
27 82.918 86.489 Variable .. I1P2 EXPER
NONE 86.222 6.016 9
81.598 90.847 EXPER
SOME 87.111 3.333 9
84.549 89.673 EXPER
LOTS 88.222 5.783 9
83.777 92.668 For entire sample 87.185 5.054
27 85.186 89.184 Variable .. I1P3 EXPER
NONE 86.111 6.679 9
80.977 91.245 EXPER
SOME 86.667 2.345 9
84.864 88.469 EXPER
LOTS 88.333
7.053 9 82.912 93.755 For entire sample 87.037 5.626
27 84.811 89.263 Variable .. I2P1 EXPER
NONE 79.778 4.147 9
76.590 82.965 EXPER
SOME 81.000 5.635 9 76.669 85.331 EXPER
LOTS 82.556 6.502 9
77.558 87.554 For entire sample 81.111 5.423
27 78.966 83.256 Variable .. I2P2 EXPER
NONE 80.556 5.725 9 76.155 84.956 EXPER
SOME 80.556 4.065 9
77.431 83.681 EXPER
LOTS 84.111 7.656 9
78.226 89.996 For entire sample 81.741 6.010
27 79.363 84.118 Variable .. I2P3 EXPER
NONE 82.333 6.671 9
77.206 87.461 EXPER
SOME 80.667 3.969 9
77.616 83.717 EXPER
LOTS 85.333 6.576 9
80.278 90.388 For entire sample 82.778 5.976
27 80.414 85.142 |
The results of the hypothesis tests from the
preceding set of commands is presented below in its full form.
* * * * * * * * * A N A L Y S I S O F V A R I A N C E * *
* * * * * * * *
Tests of Between‑Subjects Effects.
Tests of Significance for T1 using UNIQUE sums of squares
Source of Variation
SS DF MS F Sig of F
WITHIN CELLS
3527.19 24 146.97
EXPER
225.93 2 112.96 .77 .475
* * * * * * * * * A N A L Y S I
S O F V A R I A N C E * * * * * * * * * *
Tests involving 'INSTRUCT' Within‑Subject Effect.
Tests of Significance for T2 using UNIQUE sums of squares
Source of Variation
SS DF MS F Sig of F
WITHIN CELLS
260.37 24 10.85
INSTRUCT
795.56 1 795.56
73.33 .000
EXPER BY INSTRUCT
19.90 2 9.95 .92 .413
Tests involving 'PLACE' Within‑Subject Effect.
Mauchly sphericity test, W =
.51268
Chi‑square approx. = 15.36643 with 2 D. F.
Significance = .000
Greenhouse‑Geisser Epsilon = .67235
Huynh‑Feldt Epsilon = .75715
Lower‑bound Epsilon = .50000
AVERAGED Tests of Significance that follow multivariate tests are
equivalent to
univariate or split‑plot or mixed‑model approach to
repeated measures.
Epsilons may be used to adjust d.f. for the AVERAGED results.
* * * * * * * * * A N A L Y S I S O F V A R I A N C
E * * * * * * *
EFFECT .. EXPER BY PLACE
Multivariate Tests of Significance (S = 2, M = ‑1/2, N = 10
1/2)
Test Name
Value Approx. F Hypoth. DF Error DF Sig. of F
Pillais
.06553 .40650 4.00 48.00
.803
Hotellings
.06957 .38266 4.00 44.00
.820
Wilks
.93472 .39481 4.00 46.00
.811
Roys
.06147
Note.. F statistic for WILK'S Lambda is exact.
* * * * * * * * * A N A L Y S I
S O F V A R I A N C E * * * * * * * * * *
EFFECT .. PLACE
Multivariate Tests of Significance (S = 1, M = 0, N = 10 1/2)
Test Name
Value Exact F Hypoth. DF Error DF
Sig. of F
Pillais
.24300 3.69164 2.00 23.00
.041
Hotellings
.32101 3.69164 2.00 23.00
.041
Wilks
.75700 3.69164 2.00 23.00
.041
Roys
.24300
Note.. F statistics are exact.
* * * * * * * * * A N A L Y S I S O F V A R I A N C E * *
* * * *
Tests involving 'PLACE' Within‑Subject Effect.
AVERAGED Tests of Significance for MEAS.1 using UNIQUE sums of
squares
Source of Variation
SS DF MS F
Sig of F
WITHIN CELLS
437.04 48 9.10
PLACE
119.11 2 59.56 6.54 .003
EXPER BY PLACE
11.85 4 2.96 .33 .860
* * * * * * * * * * A N A L Y S I S O F V A R I A N C
E * * * * *
Tests involving 'INSTRUCT BY PLACE' Within‑Subject Effect.
Mauchly sphericity test, W =
.91778
Chi‑square approx. = 1.97333 with 2 D. F.
Significance = .373
Greenhouse‑Geisser Epsilon = .92403
Huynh‑Feldt Epsilon = 1.00000
Lower‑bound Epsilon = .50000
AVERAGED Tests of Significance that follow multivariate tests are
equivalent to
univariate or split‑plot or mixed‑model approach to
repeated measures.
Epsilons may be used to adjust d.f. for the AVERAGED results.
* * * * * * * * A N A L Y S I
S O F V A R I A N C E * * * *
* * * * *
EFFECT .. EXPER BY INSTRUCT BY PLACE
Multivariate Tests of Significance (S = 2, M = ‑1/2, N = 10
1/2)
Test Name
Value Approx. F Hypoth. DF Error DF
Sig. of F
Pillais
.17727 1.16704 4.00 48.00
.337
Hotellings .20882 1.14852
4.00 44.00 .346
Wilks
.82521 1.15950 4.00 46.00
.341
Roys
.16200
Note.. F statistic for WILK'S Lambda is exact.
* * * * * * * * A N A L Y S I
S O F V A R I A N C E * * * * * * *
EFFECT .. INSTRUCT BY PLACE
Multivariate Tests of Significance (S = 1, M = 0, N = 10 1/2)
Test Name
Value Exact F Hypoth. DF Error DF
Sig. of F
Pillais
.29476 4.80658 2.00 23.00
.018
Hotellings
.41796 4.80658 2.00 23.00
.018
Wilks
.70524 4.80658 2.00
23.00 .018
Roys
.29476
Note.. F statistics are exact.
* * * * * * * * A N A L Y S I S
O F V A R I A N C E * * * * * *
*
Tests involving 'INSTRUCT BY PLACE' Within‑Subject Effect.
AVERAGED Tests of Significance for MEAS.1 using UNIQUE sums of
squares
Source of Variation
SS DF MS F Sig of F
WITHIN CELLS
148.30 48 3.09
INSTRUCT BY PLACE 23.75
2 11.88 3.84 .028
EXPER BY INSTRUCT BY
18.62 4 4.65 1.51 .215
PLACE
From the preceding analysis the results are
reasonably clear. Main effects were
found for the INSTRUCT and PLACE factors.
When instructed, the singers sang with less intensity. In different places in the song the singers
also sang with different intensity levels over both levels of instructions.
The significant interactions of EXPER BY
PLACE and INSTRUCT BY PLACE are best seen in the three-way interaction, even
though the three-way interaction was not significant. The graph of the EXPER BY INSTRUCT BY PLACE interaction is
presented below.
|
|
From the above graphs it can be seen that the
more experienced the singer, the more loudly he or she sang. In addition, experienced singers sang
relatively more loudly than inexperienced singers at song positions 2 and
3. Also, the instructions to sing at an
even level seemed to work best at position 2, combined over all groups.
For an explanation of what all this means, it
is perhaps best to refer to the original author (Tonkinson,1990, p. 25)
The original problem was: Is the Lombard effect, to a significant
degree, and unconscious response in choral singing at different levels of
experience and training, and can it be consciously avoided. Most of the choral singers in this study,
regardless of experience, tended to succumb to a Lombard effect when faced with
increasing loss of auditory feedback.
They were, however, able to control the level of vocal intensity with
some brief instructions. It appears
that simple training in awareness would be enough for a member of an amateur
choir to begin regulating the intensity of their voice in a healthy manner.
BIBLIOGRAPHY
CMS User's Manual. (1992). SMSU.
Electronic Mail User's Guide. (1992). SMSU.
Hays, W. L. (1981). Statistics (3rd ed.).
New York: Holt, Rinehart and Winston.
Kerlinger, F. N. (1973). Foundations of Behavioral Research (2nd
Edition). New York: Holt, Rinehart, and Winston.
Letcher, John S. Jr., Marshall, John K.,
Oliver, James C. III, and Salvesen, Nils.
(1987) Stars & Stripes. Scientific American, 257(2), 34-40.
McGuire, L. C. Adults' Recall and Retention of Medical Information: What Does the Patient Really Remember? Dissertation Abstracts, 1993.
Michell, J. (1986). Measurement Scales and Statistics: A Clash of Paradigms. Psychological
Bulletin, 100-3, 398-407.
Norusis, Marija. (1990). The SPSS Guide to Data Analysis for
Release 4. Chicago: SPSS, Inc.
Reed, Valerie. The Student's Knowledge of United States Geography. A project done for PSY 200. Southwest Missouri State University, 1983.
SPSS Reference Guide.
(1990). Chicago: SPSS, Inc.
Stevens, S. S. (1951). Mathematics, measurement and
psychophysics. In S. S. Stevens(Ed.), Handbook
of experimental psychology (pp. 1-49).
New York: Wiley.
Tonkinson, S. E. The Lombard Effect in Choral Singing. Dissertation Abstracts, 1990.
Torgerson, W. S. (1967). Theory and Methods of Scaling. New York:
Wiley.
Tufte, Edward R. (1983). The Visual Display of Quantitative
Information. Cheshire, Conn: Graphics Press.
Winer, B. J. (1971). Statistical Principles in Experimental
Design. New York: McGraw-Hill.
INDEX
dummy
codes......................... 232
MANOVA command............. 166
multiple regression.......... 219,
222
orthogonal transformations...... 214
SPSS
A X B.................................. 173
t-test...................................... 170
Table
of Means...................... 194
theory-driven.......................... 171
/DESIGN................ 166,
226, 228
/METHOD
SEQUENTIA....................... 226
SEQUENTIAL............ 226,
228
UNIQUE............................. 226
/METHOD = SEQUENTIAL. 226
/METHOD = UNIQUE.......... 226
/METHOD=SEQUENTIAL... 233
/METHOD=TEST.................. 234
/METHOD=UNIQUE............ 232
/METHODS=UNIQUE.......... 231
A X B..................................... 172
Analysis of variance................. 214
ANOVA 132,
138, 158, 163, 164, 199, 214, 216, 217, 219, 229
single-factor.......................... 203
summary table....................... 176
ANOVA table 151,
152, 193, 195, 203‑205, 207, 241
Assumption
within-subjects designs.......... 244
Assumptions
A X B X C........................... 258
AVOVA table......................... 195
B WITHIN A.......................... 195
B(A)............................... 193,
207
B(A) ...................................... 141
Balanced designs............. 225, 230
Between groups....................... 173
Between subjects..... 142, 173, 261
Between subjects effects.......... 242
Between subjects factors......... 149
Blocking factor................ 136, 140
Bolded capital letters............... 133
Bulimia.................................... 146
BY.......................................... 166
Carry-over effects................... 140
categorical............................... 224
Causal inferences..................... 136
Cell means............... 175,
190, 233
Central Limit Theorem...... 154‑156
Choir....................................... 145
Collapsing............................... 177
Complex experiments.............. 149
constant term........................... 217
Contrast
difference.............................. 210
non-orthogonal............. 199, 208
orthogonal.................... 198,
200
polynomial............................ 211
simple................................... 210
Contrasts................................. 197
CORRELATION.................... 239
correlation matrix 217, 222, 234,
240, 244, 251
Correlations............................. 258
Counter variable...................... 168
Critical value............................ 152
Crossed.................................. 139
Data file 165,
173, 215, 239, 249, 261
Data files................................. 165
DATA LIST............ 173,
239, 250
Data matrix...................... 148,
222
Data table................................ 190
data-driven.............................. 171
default..................................... 227
/METHOD........................... 227
degrees of freedom 151, 206, 222, 243,
258
orthogonal contrasts.............. 204
Dependent measure................. 147
Dependent variable.......... 133, 165
continuous............................ 214
DESIGN 174,
193, 240, 251, 262
unbalanced........................... 226
Designs
unbalanced........................... 225
Df........................................... 151
Difference contrasts................. 210
discrete variables..................... 215
discriminant analysis................. 223
Dot notation.................... 168,
190
Dots........................................ 189
dummy codes.......................... 234
dummy coding 214,
217, 222, 223
Duncan’s Multiple Range test... 170
E............................................. 137
ED-STAT....................... 225,
227
Effects............. 132,
152, 197, 268
EMS............... 246,
247, 259, 279
Error term 247,
260, 267, 276, 279
Essay...................................... 172
Exact F................................... 265
Example
non-orthogonal contrasts....... 208
Expected mean square 246, 259, 260, 267
Expected mean squares 245, 247, 258,
267, 276
Experiment-wise error rate....... 151
Experimental design 132, 143, 145
experimentwide error rate........ 170
F-distribution................... 159, 162
F-ratio 151,
157‑160, 162, 167, 245, 247, 258,
260, 267
Factor..................... 132,
214, 268
fixed..................................... 137
random................................. 138
Factor name............................ 166
Factors.................................... 133
crossed................................. 139
Fcrit................................ 160,
164
Fixed....................................... 137
FOBS..................................... 164
G............................................ 135
Gender.................................... 136
general linear model................. 223
Geometrical perspective........... 231
Geometrical perspective .......... 231
Grand mean..................... 175,
254
Grand mean ............................ 190
Graph...................................... 264
Graphical presentation............. 153
Greenhouse-Geisser Epsilon.... 243
Group factor............................ 135
Hierarchical design................... 141
Hierarchical designs................. 192
Hierarchical designs ................ 141
Homer & Tessie...................... 258
Huynh-Feldt Epsilon................ 243
hypothesis test......................... 219
Hypothesis testing............ 132, 158
Hypothesis testing theory......... 246
independence.................. 157,
222
Independent effects.................. 225
Independent variable 132, 165, 214, 223
independent variables............... 222
Infinite..................................... 259
Interaction 178,
206, 226, 237, 253, 264, 272, 279, 285
simple................................... 275
simple two-way.................... 271
three-way 268,
271, 273, 275, 285
two-way............................... 269
Interaction effect...... 222, 260, 267
Interaction term....................... 226
Interactions
three-way............................. 271
interval.................................... 224
Interval scale........................... 211
Large standard deviation.......... 188
Levels of a factor..................... 133
Linear effects........................... 231
Linear trend..................... 211,
212
LIST............................... 166,
250
Lombard effect 144, 238, 249, 261,
280
Main effect 206,
222, 226, 243, 265, 267
simple........................... 178,
207
Main effects 177,
255, 269, 285
nested................................... 194
MANOVA 132,
137, 174, 175, 202, 212, 226, 232, 233, 236, 239, 250, 252, 262, 263, 279, 280
/METHOD=TEST................ 236
contrasts............................... 209
METHOD=SEQUENTIAL.. 235
MANOVA command.............. 164
Marginal means....... 175, 177, 233
matrix
transpose.............................. 217
Mauchly sphericity test 243, 245, 255, 256, 265, 279
Mean
cell....................................... 175
grand.................................... 175
Marginal............................... 175
Mean Square................... 151,
247
Mean Square Between............ 161
Mean Square Within................ 161
Mean Squares Between........... 157
MEAN SQUARES WITHIN.. 155
means.............................. 175,
214
unweighted........................... 233
weighted............................... 233
MEANS command.................. 164
Means squares........................ 246
Measures per subject............... 148
Mixed design........................... 263
Model............................. 162,
229
adjusted................................ 230
MS................................. 151,
258
MSB ...................................... 161
MSW...................................... 161
Multiple regression 197, 214, 215,
217, 223, 231
ANOVA.............................. 237
Multiple t-tests......................... 150
Multivariate............................. 254
Multivariate ANOVA.............. 165
Multivariate approach.............. 245
Multivariate test....................... 256
Multivariate tests.............. 243, 265
Multivariate tests of significance 255, 265, 279
Negative numbers.................... 229
Nested.................... 140,
147, 192
Nested main effect................... 195
Nested main effects................. 194
Newman-Keuls....................... 170
Nichols, David......................... 230
Non-orthogonal contrasts 199, 209
Non-parallel............................ 271
Non-significant........................ 203
Nonsignificant effects............... 152
Notational system............ 133, 169
Number of subjects................. 148
orthogonal............................... 234
orthogonal contrast.......... 198, 217
Orthogonal contrasts 199, 200, 213, 215‑217,
219, 220
sum of squares...................... 204
orthogonal transformations....... 214
Parallel.................................... 178
Parameter................................ 156
Parameters...... 153,
154, 159, 178
Parsimonious........................... 200
Part correlations...................... 231
Polynomial Contrasts
example................................ 212
post-hoc.................................. 170
Pre-existing ............................ 135
pre-planned contrasts.............. 170
PRINT.................................... 252
PRINT = CELLINFO(MEANS) 240, 251, 262
PRINT CELLINFO(MEANS) 174
PRINT=CELLINFO(MEANS 166
Probability models................... 153
Probed recall........................... 143
Quadratic trend............... 211, 212
Random factor......................... 138
Ratio....................................... 243
Real effect............................... 177
Real effects.............. 152,
160, 161
Reality Therapy ...................... 150
recoding.................................. 217
Regression....................... 229,
234
/METHOD=TEST 234, 235
METHOD=TEST................. 234
Repeated measures.......... 238, 249
Repeated measures designs...... 140
residual vector......................... 231
Reverse Helmert contrast......... 210
S............................................. 137
S ( A )
X B............................ 261
S X A..................................... 238
S X A X B.............................. 249
S(A)........................................ 142
Sample statistics.............. 153, 155
Sampling distribution 153‑155, 161
Sampling distribution of the mean 154
Score model 225,
238, 245, 246, 258, 276
SEQUENTIAL....... 226,
227, 230
Set of contrasts........................ 198
Set of orthogonal contrasts....... 204
polynomial trends.................. 211
Sheffe’.................................... 170
Sig.......................................... 151
Sig of F........................... 152,
177
Significant................ 177,
203, 272
Significant effects..................... 152
Significantly different................ 164
Simple effects.......................... 195
Simple main effect............ 178, 269
Simple two-way interaction...... 271
Small letters............................. 133
Solitaire................................... 133
Source............................ 151,
177
Sphericity test.................. 243, 256
SPSS 164,
165, 193, 201, 202, 213, 215, 225, 227,
254, 260, 267, 279
MANOVA........... 219,
220, 223
REGRESSION............ 218,
222
SS........... 151,
204, 228, 245, 258
non-orthogonal contrasts....... 209
orthogonal contrasts.............. 204
WITHIN.............................. 208
Standard error of the mean...... 156
Star Spangled Banner.............. 238
Straight line.............................. 211
Subject numbers...................... 239
Subjects.......................... 132,
137
Subjects main effect................. 260
Subscript................................. 168
Subscripted............................. 248
Subscripts....................... 133,
189
Sum of Squares............... 151, 227
non-orthogonal contrasts....... 208
total - balanced designs......... 225
Summary table................. 174,
177
Summation...................... 168,
189
Summation sign........................ 189
Sums of squares.............. 230, 231
Swimsuit issue......................... 146
T............................................. 135
T-distribution........................... 159
T-test.............................. 159,
163
Table of means 175, 202, 268, 272,
274
Tables of Means...................... 193
Theoretical probability distribution 159
Three-way interaction...... 268, 276
Treatment factors..................... 134
Treatments...................... 132,
134
Trend...................................... 211
Trials factor............................. 135
Two-tailed test........................ 163
Two-way interactions.............. 269
Type I error............................. 151
Unbalanced............................. 227
Unbalanced design................... 225
Unbalanced designs 225, 227, 230
uncorrelated............ 217,
222, 234
unequal cell sizes.............. 224, 237
Unequal numbers..................... 230
UNIQUE......................... 226‑230
Unit factor............................... 137
Univariate........................ 254,
265
Univariate approach................. 245
VALUE LABELS........... 174,
250
vectors.................................... 231
Vocal intensity......................... 145
WARNING............................ 166
Warning message..................... 262
WITHIN................................. 193
WITHIN CELLS............ 260,
267
Within method......................... 155
Within subjects 140, 238, 249, 261,
265
Within subjects design
assumptions.......................... 243
Within subjects designs............ 238
Within subjects factor.............. 262
Within subjects factors............. 149
WSDESIGN........... 240,
251, 262
WSFACTORS....................... 251
WSFACTORS = 239,
250, 262
X............................................ 139
X-axis..................................... 176