same old question - lme4 and p-values
On Mon, Apr 7, 2008 at 9:18 PM, David Henderson <dnadavenator at gmail.com> wrote:
Hi Doug:
Perhaps I should clarify. The summary of a fitted lmer model does not provide p-values because I don't know how to calculate them in an acceptable way, not because I am philosophically opposed to them. The estimates and the approximate standard errors can be readily calculated as can their ratio. The problem is determining the appropriate reference distribution for that ratio from which to calculate a p-value. In fixed-effects models (under the "usual" assumptions) that ratio is distributed as a T with a certain number of degrees of freedom. For mixed models it is not clear exactly what distribution it has - except in certain cases of completely balanced data sets (i.e. the sort of data sets that occur in text books). At one time I used a T distribution and an upper bound on the degrees of freedom but I was persuaded that providing p-values that could be strongly "anti-conservative" is worse than not providing any.
Now I understand the situation better and am in agreement that this is clearly the right solution at this point.
The approach that I feel is most likely to be successful in summarizing these models is first to obtain the REML or ML estimates of the parameters then to run a Markov chain Monte Carlo sampler to assess the variability in the parameters (or, if you prefer, the variability in the parameter estimators). (Note: I am not advocating using MCMC to obtain the estimates, I suggest MCMC for assessing the variability.)
I'm a little confused as to what is the Monte Carlo part of this scenario? If you perform REML or ML, theoretically it should always converge to the REML/ML estimates (unless you have a flat or multimodal likelihood which each produce other problems). I understand you are fixing the parameter estimates of something at the REML/ML estimates, but what is the random component?
Although the MCMC chain starts at the REML/ML estimates its iterations are of the form - given the current residuals, sample a new value of $\sigma$ using a random value from a $\chi^2$ distribution - given the current values of $\sigma$ and the variance-covariance of the random effects, sample new values of the fixed-effects and the random effects (in the Bayesian formulation both the fixed-effects parameters and the random effects are regarded as random variables). For a locally uniform prior on the fixed-effects this stage can be reduced to sampling from a multivariate normal distribution. - given the current values of $\sigma$, the fixed effects, the variance-covariance of the random effects and the random-effects themselves sample from the distribution of the variance-covariance parameters. - repeat the above three steps n times It is the third step that gets tricky. The "simple" approach is to condition only on the values of the random effects and use a Wishart distribution. The problem with feedback is in steps 2 and 3. If in step 3 you happen to get a very small value of a variance component then the next set of random effects sampled in step 2 will be small in magnitude, resulting in the next sample of the variance component in step 3 being very small, resulting in ... I enclose a script to illustrate this effect using a model fit to data from an experiment described in the classic book "Statistical Methods in Research and Production" edited by O.L.Davies. The "Yield" variable is the amount of dyestuff in five different analyses of samples from each of six different batches. The REML estimates for a simple random effects model reproduce the estimates of the variance components shown in the book (there were at least 4 editions of the book, the first in 1947, and all contain this example). If you run this script yourself and look at the plot of the samples from the Bayesian posterior distribution of the 3 parameters versus the iteration number for the MCMC sample, you will see the places where the value of ST1, which is the standard deviation for batches divided by the standard deviation for analyses within batch, gets stuck at zero. Those places also coincide with unusually large values of sigma and attenuated variability of the distribution of the mean parameter. (I don't enclose the plots because even the PDF files are very large when you are plotting 10000 samples)
Of course, I could always stop being lazy and just look at the source... ;^)
The current version of the mcmcsamp function suffers from the practical problem that it gets stuck at near-zero values of variance components. There are some approaches to dealing with that. Over the weekend I thought that I had a devastatingly simple way of dealing with such cases until I reflected on it a bit more and realized that it would require a division by zero. Other than that, it was a good idea.
At least the variance estimates were not negative... ;^) Thanks!! Dave H -- David Henderson, Ph.D. Director of Community REvolution Computing 1100 Dexter Avenue North, Suite 250 206-577-4778 x3203 DNADave at Revolution-Computing.Com http://www.revolution-computing.com
-------------- next part -------------- An embedded and charset-unspecified text was scrubbed... Name: Davies_R.txt URL: <https://stat.ethz.ch/pipermail/r-sig-mixed-models/attachments/20080408/7b69f7a7/attachment.txt> -------------- next part -------------- An embedded and charset-unspecified text was scrubbed... Name: Davies_Rout.txt URL: <https://stat.ethz.ch/pipermail/r-sig-mixed-models/attachments/20080408/7b69f7a7/attachment-0001.txt>