@avehtari Looking at the vignette,
Computed from 4000 by 262 log-likelihood matrix
Estimate SE
elpd_loo -6236.9 725.4
p_loo 284.9 69.1
looic 12473.8 1450.7
------
Monte Carlo SE of elpd_loo is NA.
Pareto k diagnostic values:
Count Pct. Min. n_eff
(-Inf, 0.5] (good) 240 91.6% 206
(0.5, 0.7] (ok) 7 2.7% 48
(0.7, 1] (bad) 8 3.1% 7
(1, Inf) (very bad) 7 2.7% 1
See help('pareto-k-diagnostic') for details.
I didn’t notice the Min. n_eff
column until now. The latest guideline is that n_eff
should be at least 100 times the number of chains. Does this apply to loo also? What about \widehat R? Also, does loo check these statistics itself or is it recommended to use the usual procedure (rstan’s summary function) to evaluate the log_lik
vector?
I understand the logic of looking at observations associated with large k values (outliers or unexpected given the posterior). Is there a useful interpretation of elpd_loo
or the SE of this quantity? Or is elpd_loo
only useful for model comparison?
There was a recent article about Bayesian Comparison of Latent Variable Models: Conditional Versus Marginal Likelihoods. Do I understand correctly that loo should not be used to compare latent variable models without integrating out the latent variables? Apparently, the blavaan
package has some code to integrate out latent variables. Any idea if this code is specific to blavaan
models or if it is generic? If it is generic, maybe it can be moved out of blavaan
some more generic package like latentStan
(I made that up)? What about Pareto k values? Do k values still retain their useful interpretation in the context of latent variable models?