Skip to content

Multilevel zero-inflated models: model selection and model assumption checks

1 message · Phillip Alday

#
I won't address everything, but will give a few short pointers
On 28/4/20 1:58 am, Kate R wrote:
For this type of thing, my advise is always "plot the model against the
data!". You can use an effects-type plot overlaid with the data or you
can plot both the fitted and observed values against the predictors (in
addition to the classical fitted vs. observed plot). Which models
capture which aspects of your data? Which models break down in
unacceptable ways? As George Box said, "all models are wrong, but some
models are useful." Which models are most useful (i.e. wrong in ways you
can accept and right in ways you need)?
This is in some sense exactly what a hurdle model does. :)
With difficulty if you want classical tests (they aren't nested so you
can't do likelihood-ratio tests and I find even AIC/BIC less than ideal
here because the different link functions means that you're not quite
comparing the same aspects of the data). You could use stratified
cross-validation to get about predictive power, but this can be
challenging depending on the exact nesting/crossing structure of your
data (sorry, I didn't have time to read and think about your particular
data in detail).? Otherwise, I recommend the graphical comparison I
mentioned above.

Hope that helps,
Phillip