Dear Emmanuel,
The questions you raise are sufficiently complicated that it's difficult to address them adequately in an email. My Applied Regression and Generalized Linear Models text, for example, takes about 15 pages to explain the relationships among regressor codings, hypotheses, and tests in 2-way ANOVA, working with the full-rank parametrization of the model, and it's possible (as Russell Lenth indicated) to work things out even more generally.
I'll try to answer briefly, however.
-----Original Message-----
From: Emmanuel Curis [mailto:emmanuel.curis at parisdescartes.fr]
Sent: February 24, 2016 4:54 AM
To: Fox, John <jfox at mcmaster.ca>
Cc: Francesco Romano <francescobryanromano at gmail.com>; r-sig-mixed-
models at r-project.org
Subject: Re: [R-sig-ME] Replicating type III anova tests for glmer/GLMM
Dear Pr Fox,
Thanks for taking time for this discussion. I think I made a few shortcuts that
are wrong, and I still have some not understood issues about the kind of
tests even in the simplest case of linear models...
First, I think I mixed contrast and quadratic forms expectations in my answer,
I apologize for that; what I had in mind when answering Francesco was in fact
No need to apologize. I don't think that these are simple ideas.
the expectation of the quadratic form, and I too quickly deduced that there
was an equivalent linear combination of the parameters as its ? square root
?, but this was obviously wrong since the L matrix in a Lt W L quadratic form
does not have to be a column matrix. Am I wrong thinking that typically in
such tests, the L matrix is precisely a multi-column matrix (hence also several
degrees of freedom associated), and that several contrasts are tested
simultaneously?
Thinking in terms of the full-rank parametrization, as used in R, each type-III hypothesis is that several coefficients are simultaneously 0, which can be simply formulated as a linear hypothesis assuming an appropriate coding of the regressors for a factor. Type-II hypotheses can also be formulated as linear hypotheses, but doing so is more complicated. The Anova() function uses a kind of projection, in effect defining a type-II test as the most powerful test of a conditional hypothesis such as no A main effect given that the A:B interaction is absent in the model y ~ A*B. This works both for linear models, where (unless there is a complication like missing cells), the resulting test corresponds to the test produced by comparing the models y ~ A and y ~ A + B, using Y ~ A*B for the estimate of error variance (i.e., the denominator MS), and more generally for models with linear predictors, where it's in general possible to formulate the (Wald) tests in terms of the coefficient estimates and their covariance matrix.
I precise that I call ? contrast ? a linear combination of the model parameters
with the constraint that the coefficients of this combination sum to 0 ? this is
the definition in French (? contraste ?), but I may use it wrongly in English?
I'd define a "contrast" as the weights associated with the levels of a factor for formulating a hypothesis, where the weights traditionally are constrained to sum to 0, and to differentiate this from a column of the model matrix, which I'd more generally term a "regressor." Often, a traditional set of contrasts for a factor, one less than the number of levels, are defined not only to sum to 0 but also to be orthogonal in the basis of the design. The usage in R is more general, where "contrasts" mean the set of regressors used to represent a factor. Thus, contr.sum() generates regressors that satisfy the traditional definition of contrasts, as do contr.poly() and contr.helmert(), but the default contr.treatment() generates 0/1 dummy-coded regressors that don't satisfy the traditional definition of contrasts.
Second, I may have wrongly understood the definitions of the various tests,
and especially how they generalize from linear model to GLM/GLMM...
I thought type I was by taking the squared distance of the successive
orthogonal projections on the subspaces generated by the various terms, in
the order given in the model; type II, by ensuring that the term tested was
the last amongst terms of same order, after terms of lower order but before
terms of higher order; type III, by projecting on the subspace after removal
of the basis vectors for the term tested ? hence its strong dependency on
the coding scheme, and the ? drop1 ? trick to get them.
Is this definition correct? Does it generalize to other kind models, or is
another definition required? Is it unambiguous? The SAS doc itself suggests
that various procedures call "type II" different kind of things
Yes, if I've followed this correctly, it's correct, and it explains why it's possible to formulate the different types of tests in linear models independently of the contrasts (regressors) used to code the factors -- because fundamentally what's important is the subspace spanned by the regressors in each model, which is independent of coding. This approach, however, doesn't generalize easily beyond linear models fit by least squares. The approach taken in Anova() corresponds to this approach in linear models fit by least squares as long as the models remain full-rank and for type-III tests as long as the contrasts are properly formulated, and generalizes to other models with linear predictors.
However, I cannot see clearly which hypothesis is indeed tested in each case,
especially in terms of cell means or marginal means (and, when I really need
it, I start from them and select the contrasts I need). Is there any
package/software that allows to print the hypotheses testeds in terms of
means starting from the model formula?
Or is there any good reference that makes the link between the two?
For instance, a demonstration that the comparison of marginal means ?
always ? leads to a type XXX sum of square?
This is where a complete explanation gets too lengthy for an email, but a shorthand formulation, e.g., for the model y ~ A*B, is that type-I tests correspond to the hypotheses A|(B = 0, AB = 0), B | AB = 0, AB = 0; type-II tests to A | AB = 0, B | AB = 0, AB = 0; and type-III tests to A = 0, B = 0, AB = 0. Here, e.g., | AB = 0 means assuming no AB interactions, so, e.g., the hypothesis A | AB = 0 means no A main effects assuming no AB interactions. A hypothesis like A = 0 is indeed formulated in terms of marginal means, understood as cell means for A averaging over the levels of B (not level means of A ignoring B).
I realize that this is far from a complete explanation.
Best,
John
Best regards,
On Tue, Feb 23, 2016 at 05:17:29PM +0000, Fox, John wrote:
? Dear Emmanuel,
?
? First, the relevant linear hypothesis is for several coefficients
simultaneously -- for example, all 3 coefficients for the contrasts
representing a 4-level factor -- not for a single contrast. Although it's true
that any linear combination of parameters that are 0 is 0, the converse isn't
true. Second, for a GLMM, we really should be talking about type-III tests not
type-III sums of squares.
?
? Type-III tests are dependent on coding in the full-rank parametrization of
linear (and similar) models used in R, to make the tests correspond to
reasonable hypotheses. The invariance of type-II tests with respect to coding
is attractive, but shouldn't distract from the fundamental issues, which are
the hypotheses that are tested and the power of the tests.
?
? Best,
? John
--
Emmanuel CURIS
emmanuel.curis at parisdescartes.fr
Page WWW: http://emmanuel.curis.online.fr/index.html