This vignette gives examples how to restrict CPU and memory usage of parallel workers. This can useful for optimizing the performance of the parallel workers, but also lower the risk that they overuse the CPU and memory on the machines they are running on.
On Unix, we can run any process with a lower CPU priority using the
nice
command. This can be used when we want to lower the risk of
negatively affecting other users and processes that run on the same
machine from our R workers overusing the CPUs by mistake. To achieve
this, we can prepend nice
to the Rscript
call via the rscript
argument using. This works both on local and remote Linux machines,
e.g.
library(parallelly)
cl <- makeClusterPSOCK(2, rscript = c("nice", "*"))
library(parallelly)
workers <- rep("n1.remote.org", 2)
cl <- makeClusterPSOCK(2, rscript = c("nice", "*"))
The special *
value expands to the proper Rscript
on the machine
where the parallel workers are launched.
This example launches two parallel workers each limited to 100% CPU
quota and 50 MiB of memory using Linux CGroups management. The 100%
CPU quota limit constrain each worker to use at most one CPU worth of
processing preventing them from overusing the machine, e.g. through
unintended nested parallelization. For more details, see man systemd.resource-control
.
library(parallelly)
cl <- makeClusterPSOCK(
2L,
rscript = c(
"systemd-run", "--user", "--scope",
"-p", "CPUQuota=100%",
"-p", "MemoryMax=50M", "-p", "MemorySwapMax=50M",
"*"
)
)
Note, depending on your CGroups configuration, a non-privileged user
may or may not be able to set the CPU quota. If not, the -p CPUQuota=100%
will be silently ignored.
The 50 MiB memory limit is strict - if a worker use more than this, the operating system will terminate the worker instantly. To illustrate what happens, we first start by generating 1 million numeric values each consuming 8 bytes, which in total consumes ~8 MB, and then calculate the mean, the memory consumption is within 50-MiB memory limit that each parallel worker has available;
library(parallel)
mu <- clusterEvalQ(cl, { x <- rnorm(n = 1e6); mean(x) })
mu <- unlist(mu)
print(mu)
#> [1] 0.0008072657 -0.0019693992
However, if we generate 10 times more values, the memory consumption will grow to at least 80 MB, which is over then 50-MiB memory limit, and we will get an error:
mu <- clusterEvalQ(cl, { x <- rnorm(n = 10e6); mean(x) })
#> Error in unserialize(node$con) : error reading from connection
This is because the operating system terminated the two background R processes, because they overused the memory. This is why the main R process no longer can communicate with the parallel workers. We can see that both workers are down, by calling:
isNodeAlive(cl)
#> [1] FALSE FALSE
We can use cloneNode()
to relaunch workers that are no longer
alive, e.g.
is_down <- !isNodeAlive(cl)
cl[is_down] <- cloneNode(cl[is_down])
isNodeAlive(cl)
#> [1] TRUE TRUE
This example, works only on MS Windows machines. It launches four local workers, where two are running on CPU Group #0 and two on CPU Group #1.
library(parallelly)
rscript <- I(c(
Sys.getenv("COMSPEC"), "/c",
"start", "/B",
"/NODE", cpu_group=NA_integer_,
"/AFFINITY", "0xFFFFFFFFFFFFFFFE",
"*")
)
rscript["cpu_group"] <- 0
cl_0 <- makeClusterPSOCK(2, rscript = rscript)
rscript["cpu_group"] <- 1
cl_1 <- makeClusterPSOCK(2, rscript = rscript)
cl <- c(cl_0, cl_1)
The special *
value expands to the proper Rscript
on the machine
where the parallel workers are launched.