Hi everyone,
I’m fitting per-subject Bayesian latent-state models in Stan for a gambling task (153 trials per participant). Each participant is fit independently.
At each trial, a latent emotional state Et evolves according to past choices and outcomes (RW-style update with linear drift). This latent state:
-
predicts choice at trial t,
-
is linked to sparse self-report emotion ratings, and
-
in an extended model, is also linked to an additional noisy observation channel derived from computer-vision-based affect decoding (Morph).
Schematically:
I’m comparing models that differ only in the measurement model of Et
-
Choice + ratings
-
Choice + ratings + computer-vision-based affect
The choice model (betas) is identical across models; differences are entirely in the additional observation channel for the latent state.
My question is specifically about choice prediction
Does adding the MorphCast-derived affective signal improve prediction of choices, via better inference of the latent emotional state?
I would like to compare the 2 models but focusing only on the choice component.
I am aware that comparing in-sample choice log-likelihoods maybe is problematic here, because adding extra measurement channels (and parameters) can improve latent-state inference and thus indirectly improve in-sample choice fit (although the choice model has the same number of parameters).
To address this, I am considering two approaches and would appreciate guidance on best practice.
Approach 1: PSIS-LOO on choice likelihood (joint models)
Models are fit per subject, so my plan is to:
-
compute PSIS-LOO per participant using only the choice likelihood,
-
sum ELPD across participants to compare models.
My understanding is that this is valid because participants are independent and ELPDs add.
Approach 2: Two-stage model (emotion then choice)
As a secondary / diagnostic approach, I am also considering:
-
Fitting an emotion-only model (ratings ± MorphCast) to infer E[t]
-
Extracting posterior draws of the latent state from generated quantities,
-
Using the inferred pre-choice state E[t] as a predictor in a separate choice-only Stan model,
-
Comparing choice predictive performance across models.
I am aware that:
-
naive plug-in of posterior means ignores uncertainty,
-
propagating uncertainty (e.g., via multiple draws or a measurement-error formulation) would be preferable,
-
and this two-stage approach answers a slightly different question than the fully joint model.
Questions
-
Is PSIS-LOO on loglikelihood of choice only, aggregated across per-subject fits, the recommended way to compare these models with respect to choice prediction?
-
Is summing per-subject ELPDs the correct aggregation strategy in this setting?
-
Does the two-stage emotion then choice approach make sense as a secondary analysis?
Thanks very much for any advice. I mainly want to make sure that the model comparison I report is statistically defensible and aligned with best practice.