Computed from 4000 by 262 log-likelihood matrix Estimate SE elpd_loo -6236.9 725.4 p_loo 284.9 69.1 looic 12473.8 1450.7 ------ Monte Carlo SE of elpd_loo is NA. Pareto k diagnostic values: Count Pct. Min. n_eff (-Inf, 0.5] (good) 240 91.6% 206 (0.5, 0.7] (ok) 7 2.7% 48 (0.7, 1] (bad) 8 3.1% 7 (1, Inf) (very bad) 7 2.7% 1 See help('pareto-k-diagnostic') for details.
I didn’t notice the
Min. n_eff column until now. The latest guideline is that
n_eff should be at least 100 times the number of chains. Does this apply to loo also? What about \widehat R? Also, does loo check these statistics itself or is it recommended to use the usual procedure (rstan’s summary function) to evaluate the
I understand the logic of looking at observations associated with large k values (outliers or unexpected given the posterior). Is there a useful interpretation of
elpd_loo or the SE of this quantity? Or is
elpd_loo only useful for model comparison?
There was a recent article about Bayesian Comparison of Latent Variable Models: Conditional Versus Marginal Likelihoods. Do I understand correctly that loo should not be used to compare latent variable models without integrating out the latent variables? Apparently, the
blavaan package has some code to integrate out latent variables. Any idea if this code is specific to
blavaan models or if it is generic? If it is generic, maybe it can be moved out of
blavaan some more generic package like
latentStan (I made that up)? What about Pareto k values? Do k values still retain their useful interpretation in the context of latent variable models?