This vignette provides a short example on how to use the R package bhmbasket
in a high performance computing (HPC) environment using the R packages doFuture
and future.batchtools
.
library(bhmbasket)
library(doFuture)
library(future.batchtools)
<- 5440
rng_seed set.seed(rng_seed)
The code below provides an example for specifying a parallel backend using the future framework with the SLURM job scheduler. Kindly see the documentation of the R packages doFuture
and future.batchtools
for further options. This code is to be run on the master node.
registerDoFuture()
<- 24 # time for job in hours
job_time <- 10 # number of worker nodes
n_workers <- 16 # number of cpus per worker node
n_cpus <- 1 # memory [GB] per worker node
gb_memory
<- tweak(batchtools_slurm,
slurm template = system.file('templates/slurm-simple.tmpl',
package = 'batchtools'),
workers = n_workers,
resources = list(
walltime = 60 * 60 * job_time,
ncpus = n_cpus,
memory = 1000 * gb_memory))
plan(slurm)
The R package bhmbasket
makes use of the foreach framework and runs with every applicable parallel backend. Therefore, no further adjustments to the code are necessary, which can be run on the master node. With a parallel backend registered as shown above, the job scheduler will automatically distribute the parallel jobs to the worker nodes.
Below is some example code, which was taken from the examples section of ?bhmbasket::getEstimates
. Kindly note that running this small example on a HPC environment will most likely not result in a performance improvement.
<- simulateScenarios(
scenarios_list n_subjects_list = list(c(10, 20, 30)),
response_rates_list = list(c(0.1, 0.2, 3)),
n_trials = 10)
<- performAnalyses(
analyses_list scenario_list = scenarios_list,
target_rates = c(0.1, 0.1, 0.1),
calc_differences = matrix(c(3, 2, 2, 1), ncol = 2),
n_mcmc_iterations = 100)
getEstimates(analyses_list)