Operating System: RHEL7
Interface Version: 2.15.1
Output of writeLines(readLines(file.path(Sys.getenv(“HOME”), “.R/Makevars”))):
Output of devtools::session_info("rstan”):
Is it possible to nest parallelization with rstan? So I have multiple things to run in parallel and I would like to start stan with 2 chains per job. The different things do run in parallel using a foreach loop which uses the doParallel backend. I would like to use 2 CPUs per stan run, but this always fails as it looks to me as if rstan interferes with the doParallel stuff.
Has anyone done this or is this not possible to do?
@wds15 did you manage to work this out? I have the same kind of problem (I think). I have 5 models (4 chains in parallel) that I would like to execute in parallel with foreach %dopar%
I found this article interesting https://www.r-bloggers.com/can-you-nest-parallel-operations-in-r/ where appears that with parLapply (I don’t know whether rstan uses this) is not possible; but with foreach is (meaning that if rstan used foreach this could be accomplished for the more robust way to manage sockets)
I managed to achieve this by using mclapply from the parallel package. In short, if you want to nest two jobs with two chains in parallel, you would do something like this:
I gave up on this and moved on with different parallelization schemes (forgot details). The issue is that the sampling method from rstan creates the cluster used for parallelization itself.
We should probably improve rstan here and make it query if there is a defined default cluster.
However, the best solution I am aware of is to run each chain separately and then merge the results using the tools in rstan… but I have opted against this hassle. The solution from @tiagocc is worth a try - I am just cautious about seed control and all that. I am not sure on that with the proposed solution (I am not saying its not working, but I just don’t know).
I am not sure what parallelisation function rstan uses, but foreach with doParallel should solve the problem and make nested parallelisation much easier (link above).
It is often necessary to test N models at the same time. So this possibility would be incredibly useful.
I haven’t done the decisions for rstan, but the problem with foreach & doParallel is that they add dependencies to the package. The parallel package which provides the mcapply used by rstan is a defacto R standard to my knowledge.
… but I agree that a foreach approach is a good thing (ideally with the doRNG). Still, most of the time I was able to rethink the problem and work around it in a meaningful way. Parallelization is good to avoid thinking about problems - which is a great way to save time.
I know I am late to the party but I want to share my scripts about implementing this.
I have 4 stan models and each of the model needs 8 chains. This should create 32 parallel tasks.
I checked the compute node I am using and although I have 33 cpus merely 15 or so get used. Not exactly sure what is the problem but I know they are computing in parallel.
Comments/corrections welcomed!
# parallel computing part (sampling and rvsims) ----
require(rstan)
require(parallel)
require(foreach)
require(doParallel)
cl <- makeCluster(4) # because I am sampling for four different stan models
registerDoParallel(cl)
seq_models <- c(1, 2, 3, 4)
foreach(i = seq_models,
.export = c(
"input1", "input2", "input3", "input4",
"stan_mod1", "stan_mod2", "stan_mod3", "stan_mod4"
)
.packages = c("rstan")) %dopar% {
# variable names
inputnm <- paste0("input", i)
modelnm <- paste0("stan_mod", i)
fitnm <- paste0("fit", i)
fitcoeffnm <- paste0("fitcoeff", i)
# stan samping
input <- get(inputnm) # e.g. get from input1
stan_mod <- get(modelnm) # e.g. get from stan_mod1
fit <- sampling(stan_mod,
data = input$data,
init = input$inits,
pars = input$para,
iter = niters, thin = nthin,
chains = input$n.chains,
cores = input$n.chains
)
# assign to output variables
assign(fitnm, fit) # e.g. fit1 <- fit
# I actually save each fit variable as Rds / Rdata because it is not accessible outside of the for each loop
save(list = fitnm, file = "save/it/here/data.Rdata") # this does not work for some reason
}
stopCluster(cl)