Why Mixed Effects Models?

Why Mixed Effects Models?
Mixed Effects Models Recap/Intro
●
●
●
Three issues with ANOVA
– Multiple random effects
– Categorical data
– Focus on fixed effects
What mixed effects models do
– Random slopes
– Link functions
Iterative fitting
Problem One: Multiple Random Effects
●
Most studies sample
both subjects and
items
Subject 1
Subject 2
Knight
story
Monkey
story
Problem One: Crossed Random Effects
●
Most studies sample
both subjects and
items
–
Typically, subjects
crossed with items
●
–
Each subject sees a
version of each item
May also be only
partially crossed
●
Each subject sees only
some of the items
...or Hierarchical Random Effects
●
Most studies sample
both subjects and
items
–
Typically, subjects
crossed with items
–
May also have one
nested within the
other (hierarchical)
●
●
e.g. autobiographical
memory
How to incorporate
this into model?
Problem One: Multiple Random Effects
●
Why do we care about items, anyway?
●
#1: Investigate robustness of effects across items
–
Concern is that effect could be driven by just 1 or 2
items – might not really be what we thought it was
–
Psycholinguistics: View is that we studying language
too, not just people
●
–
Other areas of psychology have not tended to care about this
Note: Including items in a model doesn't really “confirm” that the
effect is robust across items. It's still possible to get a reliable
effect driven by a small number of items. But it allows you
investigate how variable the effect is across items and why
different items might be differentially influenced.
Problem One: Multiple Random Effects
●
Why do we care about items?
●
#2: Violations of independence
–
–
–
–
–
A BIG ISSUE
Suppose Amélie and Zhenghan see
items A & B but Tuan sees items C & D
Likely that Amélie's results are more like
Zhenghan's than like Tuan's
But ANOVA assumes observations
independent
Even a small amount of dependency
can lead to spurious results (Quene & van
den Bergh, 2008)
● Dependency you didn't account for makes the variance
look smaller than it actually is
C
A
D
B
What Constitutes an “Item”?
●
●
Items assumed to be independently sampled
sampled from population of relevant
items
ALL POSSIBLE
2 related words / sentences not
DISCOURSES
independently sampled
–
–
–
●
●
“The coach knew you missed practice.”
“The coach knew that you missed practice.”
Not a coincidence both are in your
experiment!
Should be considered the same
item
But 2 unrelated things can be
different items
Problem One: Crossed Random Effects
●
ANOVA solution
–
–
●
Subjects analysis:
Average over
multiple items for
each subject
Items analysis:
Average over
multiple subjects for
each item
Note: not real data or statistical tests
F1 = 18.31, p < .001
Two sets of results
–
–
Sometime combined
with min F'
An approximation of
true min F
F2 = 22.10, p < .0001
Problem One: Crossed Random Effects
Some debate on how
accurate min F' is
●
–
●
Note: not real data or statistical tests
Scott will admit to not be fully
read up on this since I came
in after people started
switching to mixed effects
models
Somewhat less
relevant now that we
can use mixed
effects models
instead
F1 = 18.31, p < .001
F2 = 22.10, p < .0001
Mixed Effects Models Recap/Intro
●
●
●
Three issues with ANOVA
– Multiple random effects
– Categorical data
– Focus on fixed effects
What mixed effects models do
– Random slopes
– Link functions
Iterative fitting
Problem Two: Categorical Data
●
ANOVA assumes our response is continuous
RT: 833 ms
●
But, we often want to look at categorical data
'Lightning hit the
church.”
vs.
“The church was hit
by lightning.”
Choice of
syntactic
structure
Item recalled
or not
Region fixated
in eye-tracking
experiment
Problem Two:
One: Categorical Data
●
●
Traditional solution:
Analyze proportions
Violates assumptions of
ANOVA
–
–
–
Among other issues: ANOVA
assumes normal distribution,
which has infinite tails
But proportions are clearly
bounded
Model could predict
impossible values like 110%
−
0
But
proportions
1
Problem Two:
One: Categorical Data
●
●
Traditional solution:
Analyze proportions
Violates assumptions of
ANOVA
–
–
–
Among other issues: ANOVA
assumes normal distribution,
which has infinite tails
But proportions are clearly
bounded
Model could predict
impossible values like 110%
−
0
But
proportions
1
Problem Two:
One: Categorical Data
●
●
●
Traditional solution:
Analyze proportions
Violates assumptions
of ANOVA
Can lead to:
–
–
Spurious effects (Type
I error)
Missing a true effects
(Type II error)
Problem Two:
One: Categorical Data
●
Transformations improve the situation but
don't solve it
–
–
●
Empirical logit is good (Jaeger, 2008)
Arcsine less so
Situation is worse for very high or very low
proportions (Jaeger, 2008)
–
.30 to .70 are OK
Problem Two:
One: Categorical Data
●
Why can't we just use logistic regression?
–
●
●
Predict if each trial's response is in category A or
category B
This is essentially what we will end up doing
But, if we are looking at things at a trial-bytrial basis...
–
–
Need to control for the different items on each trial
Problem One again!
Mixed Effects Models Recap/Intro
●
●
●
Three issues with ANOVA
– Multiple random effects
– Categorical data
– Focus on fixed effects
What mixed effects models do
– Random slopes
– Link functions
Iterative fitting
Problem Three: Focus on Fixed Effects
●
●
ANOVA doesn't characterize differences
between subjects or items
The bird that they spotted was a ....
ENDING
MEAN READING TIME
●
●
Predictable
283 ms
Unpredictable
309 ms
26 ms
We just have a mean effect
No info. about how much it varies
across participants or items
cardinal
pitohui
Problem Three: Focus on Fixed Effects
●
Can try to account for some of this with an
ANCOVA
–
–
But not typically done
And would have to be done separately for
participants and items (Problem One again)
MEAN
Predictable
283 ms
Unpredictable
309 ms
26 ms
Power of
items
analysis!
Mixed Effects Models Recap/Intro
Power of
subjects
analysis!
●
●
●
Three issues with ANOVA
– Multiple random effects
– Categorical data
– Focused on fixed effects
What mixed effects models do
– Random slopes
– Link functions
Iterative fitting
Captain MLM
to the rescue!
Mixed Effects Models to the Rescue!
●
ANOVA: Unit of analysis is cell mean
●
MLM: Unit of analysis is individual trial!
Mixed Models to the Rescue!
●
●
Look at individual trials
Model outcome using regression
RT
=
Semantic
categorization:
Is it a dinosaur?
+ Subject +
Item
Prime?
Problem One solved!
Mixed Models to the Rescue!
●
This means you will need your data formatted
differently than you would for an ANOVA
–
Each trial gets its own line
Mixed Models to the Rescue!
●
Is this useful for what we care about?
–
–
●
Stereotypical view of regression is that it's about
predicting values
In experimental settings we more typically want to
know if Variable X matters
Yes! We can test individual effects: Do they
contribute to the model?
–
e.g. does priming predict something about RT?
RT
=
+
Prime?
Subject
Jason
+
Item
Mixed Effects Models Recap/Intro
●
●
●
Three issues with ANOVA
– Multiple random effects
– Categorical data
– Focus on fixed effects
What mixed effects models do
– Random slopes
– Link functions
Iterative fitting
Fixed vs. Random Slopes
●
●
Fixed Slope: Same for all participants/items
Random Slope: Can vary by participants/items
RT
=
+
Prime?
+
Laurel
26 ms
+
88 ms
Stego.
Fixed vs. Random Slopes
●
●
Fixed Slope: Same for all participants/items
Random Slope: Can vary by participants/items
RT
=
+
Prime?
Laurel
26 ms
Example: Some items
may show a larger
priming effect than others
+
+
315 ms
Dr. L
Fixed vs. Random Slopes
●
●
●
Fixed Slope: Same for all participants/items
Random Slope: Can vary by participants/items
Can also test what explains variation
RT
=
+
Prime?
Dr. L
Laurel
26 ms
e.g. Adding lexical frequency
to the model may account for
variation in priming effect
+
+
+
15 ms
Lex.Freq.
300 ms
Fixed vs. Random Slopes
●
●
●
Fixed Slope: Same for all participants/items
Random Slope: Can vary by participants/items
Can also test what explains variation
RT
=
+
Prime?
Dr. L
Laurel
26 ms
Problem Three
Solved!
+
+
+
15 ms
Lex.Freq.
300 ms
Mixed Effects Models Recap/Intro
●
●
●
Three issues with ANOVA
– Multiple random effects
– Categorical data
– Focus on fixed effects
What mixed effects models do
– Random slopes
– Link functions
Iterative fitting
Link Functions
Specifies how to connect predictors to
the outcome
●
+
RT
Prime?
+
Subject
Item
1300 ms
Every model has one....
...sometimes, just the identity function
●
●
–
With Gaussian (normal) data
Link Functions
●
Specifies how to connect predictors to
the outcome
Accuracy
Yes/No
●
+
Prime?
+
Subject
Item
For binomial (yes/no) outcomes: Model
log odds to predict outcome
Problem Two solved!
Link Functions
●
Default link function for binomial data is logit
(log odds)
– Odds: p(yes)/p(no) or p(yes)/[1-p(yes)]
●
–
Log Odds: ln(Odds)
●
●
No upper bound, but lower bound at 0
Now unbounded at both ends
Can also use probit
–
–
Based on cumulative distribution function of normal
distribution
Very highly correlated with logit; almost always give
you same results as logit
●
Probit assumes slightly fewer hits at low end of distribution
& slightly more hits at high end
Mixed Effects Models Recap/Intro
●
●
●
Three issues with ANOVA
– Multiple random effects
– Categorical data
– Focus on fixed effects
What mixed effects models do
– Random slopes
– Link functions
Iterative fitting
One Caveat...
Where do model results come from?
(Answer: When a design matrix
and a data matrix really love
each other...)
One Caveat...
●
Fitting ANOVA / linear
regression has easy
solution
b = (X'X)-1X'Y
●
●
A few matrix multiplications a computer can
do easily
– A “closed form solution”
Like a “beta machine” … you put your data in
and automatically get the One True Model
out
One Caveat...
MEMs requires iteration
●
–
–
Check various sets of
betas until you find
the best one
R does this for you
An estimation
●
–
The best model: The one that
smiles with its eyes
Not mathematically guaranteed to be best fit
Complicated models take longer to fit
●
–
–
If too many parameters relative to data, might completely fail
to converge (find the best set of betas)
Scott's only experience with this is with multiple random
slopes of interactions