Skip to content
Prev 1851 / 20628 Next

Teaching Mixed Effects

The accumulation of quoted posts in this thread is quite long, so I 
have trimmed out all but some paragraphs from Doug Bates that prompt my 
comments.
[ ... ]
But a goal of this exploration of the model/data is still to come up 
with an approximation to a sampling distribution (of a test statistic 
or parameter estimate), right?  Ultimately, once all the exploration is 
done, the scientific researcher still wants to be able to tell his/her 
colleagues "Based on these data my conclusion about the effect of 
treatment X is ______ and my confidence in this conclusion is _____."  
That is, the researcher wants to make *inferences* from the data.
This paragraph brings to mind some comments about p-values and
hypothesis 
tests that seem popular on this list.  Among users of lme4 a theme seems
to 
be that "this ignorant editor/referee is insisting that I demonstrate
that 
my discovery of the effect of treatment X is not a false positive.  How
can 
I get around this?"

Every field of science needs to protect its literature from being
overwhelmed 
by false "discoveries".  Since all the professional and economic rewards
go to 
those who make discoveries, there is enormous incentive to claim that
one 
has found an effect or association, and so there needs to be a reality
check.  
The conventional way of judging these claims is via p-values and
hypothesis 
tests.  Granting all of their faults and limitations, what do people
propose 
as a better way?  

Some will probably say we should focus on (interval) estimation of
parameters 
rather than testing.  But this doesn't solve the problem at hand.
Remember 
how this discussion started:  the difficulty of calculating reliable
p-values 
for fixed effects.  Because of the duality between hypothesis tests and
confidence 
intervals, if you can't get reliable p-values, then essentially by
definition 
you can't get reliable confidence intervals either.  So the issue of
testing-
versus-estimation is a red herring with respect to the deeper problem of

inference for fixed effects in mixed models.

Others may try to dodge the problem by claiming to be Bayesians,
calculating 
credible intervals from posterior distributions.  But this won't hold
water 
if they are using lme4 and mcmcsamp.  No honest Bayesian can use an 
"uninformative" prior (if such a thing even exists), or at least not
more 
than once in any area of research:  after analysis of one data set, the 
posterior from that analysis should inform the prior for the next, but
lme4 
has its priors hard-coded.  I think the real rationale for mcmcsamp is
the 
hope that it will produce results with good frequentist properties.  I
am 
not aware that this has been demonstrated for mixed models in the
peer-reviewed 
statistics literature.
Doug, can you elaborate on that last clause?  In what way is the
(absolute) 
ratio |T| informative that the monotonic transformation 2*(1 -
pnorm(abs(T)))
is not?  In other words, if a p-value (based in this case on a standard 
normal) is not reliable for inference, what inferential value does T
have?
Less formally, if T = 2, for example, what exactly do you conclude about

the parameter, and what is your confidence in that conclusion?
 
   [ ... ]
Yes, but the problem is that (as you noted earlier in your post) the
null 
distribution depends on the values of the nuisance parameters, except in

very special cases.  A simple parametric bootstrap conditions on the 
estimated values of those parameters, as if they were known.
Nevertheless, 
a function to do this might be a useful addition to lme4.

Rich Raubertas
Merck & Co.
Notice:  This e-mail message, together with any attachme...{{dropped:12}}