Hi.
Thanks for weighing in.
If you can make your parameters uncorrelated and Gaussian shaped (not thick tailed) derivative based sampling with finite precision computers will happily explore.
So, in general, we’re “simply” looking for a way to decouple the parameters from each other? E.g. McElreath (Stat. Rethinking, 2020, p. 424) has an example where
\alpha_j \sim N(\bar{\alpha}, \sigma_{\alpha}) \\
\bar{\alpha} \sim N(0, 1.5) \\
\sigma_{\alpha} \sim \mathrm{Exp}(1)
is reparametrized as
\bar{\alpha} + z \times \sigma_{\alpha} \\
\bar{\alpha} \sim N(0, 1.5) \\
\sigma_{\alpha} \sim \mathrm{Exp}(1) \\
z \sim N(0, 1).
So, in the first parametrization, HMC has a hard time sampling \alpha_j because it is tightly coupled with \bar{\alpha} and \sigma_{\alpha}? In other words, we’ve in effect created a correlation between the distributions of \bar{\alpha} and \sigma_{\alpha}, and the resulting multivariate distribution (that of \alpha_j) has a weird shape that’s difficult to sample?
In the second parametrization, on the other hand, \alpha_j has been broken into the independent distributions of \bar{\alpha}, z and \sigma_{\alpha}, which are all easier to sample?
I have completely accidentally stumbled upon a paper in my Zotero library (Papaspiliopoulos, O., Roberts, G. O., & Sköld, M. (2007). A General Framework for the Parametrization of Hierarchical Models. Statistical Science, 22(1), 59–73) which has the figures below illustrating centered and non-centered parametrizations. Seems to be in line with what you wrote, if I understood both correctly.
They also list some “tricks” for reparametrizations: