Skip to content

random as fixed effect

3 messages · Ben Bolker, John Maindonald, Andrew Robinson

#
[cc'ing back to r-sig-mixed]
On 12-10-11 09:08 AM, Andrew Koeser wrote:
I would argue this is not really a problem in transitioning from SAS
to R, but from classical method-of-moments ANOVA to modern mixed models;
you will have the same kinds of results with SAS PROC MIXED as you will
with nlme/lme4.  http://glmm.wikidot.com/faq#fixed_vs_random  goes into
more detail.  There is a distinction between _conceptual_ or
_philosophical_ random effects (we don't want to make inferences about
specific values, we want to make inferences about the population) and
_computational_ random effects (we want to estimate effects with
shrinkage, we have enough levels to estimate the variance reasonably
well). I would agree that in the best of all possible worlds you would
somehow be able to generalize from an experiment that was run in two
successive years to the performance of a crop variety across all
possible years (and estimate the variance among years accurately), but
that doesn't work particularly well on statistical grounds (the variance
is extremely poorly determined), and in the case of mixed models it
generally fails for computational reasons as well.
#
1. "we want to make inferences about the population": 
Even making year a random effect is not really enough.  We are dealing with 
a time series, and modelling it as a random effect is a weak concession to that 
issue.  If one does nonetheless fit year as a fixed effect, one should at least 
examine the results for the separate years separately, and check on the extent 
to which they point in the same direction.  Published use of the analysis should 
acknowledge the consequent uncertainty.  

Note however that for certain types of balanced models, the estimates of treatment 
effects will be the same irrespective of whether one fits years as random or fixed.
The model is not allowing for a year by treatment interaction, just as the standard
form of analysis of block designs does not and cannot allow for a block x treatment
interaction.

2. "statistical grounds (the variance is extremely poorly determined)": 
but of course ignoring this component of variance, if it does affect treatment or other 
estimates, does not cause it to go away.

3. "computational reasons": 
The algorithms used in lme4 are general to the extent that they are able to handle
a huge variety of designs.  My experience is using Genstat, which uses quite a
different algorithm. was that it rarely failed for the balanced or approximately
balanced designs that are usual in field and suchlike experimentation.  ASREML
would no doubt perform similarly.

John Maindonald             email: john.maindonald at anu.edu.au
phone : +61 2 (6125)3473    fax  : +61 2(6125)5549
Centre for Mathematics & Its Applications, Room 1194,
John Dedman Mathematical Sciences Building (Building 27)
Australian National University, Canberra ACT 0200.
http://www.maths.anu.edu.au/~johnm
On 12/10/2012, at 12:20 AM, Ben Bolker <bbolker at gmail.com> wrote:

            
#
On Fri, Oct 12, 2012 at 09:46:58AM +1100, John Maindonald wrote:
I echo John's concern.  I would argue that this component of variance
will always affect interval estimates, and it should not be ignored.
I feel uneasy about converting random effects into fixed effects
simply because they have few levels; in so doing we risk
over-confidence in our estimates and tests, because we're assuming
that the contribution is really 0.

My opinion is that the structure of the model should honestly reflect
the structure of the design, at very least.  In an ideal world we
should include the uncertainty around the random effects estimate,
but I do not see that being done.  Maybe two experimental units really
is too few for inference!

Best wishes

Andrew