library(knitr) knitr::opts_chunk$set( eval = FALSE, error = FALSE, tidy = FALSE, message = FALSE, warning = FALSE, fig.width = 6, fig.height = 6, fig.align = "center")
Load the library:
library(bsub)
library(GetoptLong)
We suggest to use bsub directly on the node that has the same file system as the computing nodes. If the file system is different from the computing nodes, you can only monitor jobs status while you cannot submit jobs.
bsub package can submit R code (by bsub_chunk()), R scripts (by
bsub_script()) and bash commands (by bsub_cmd()) to the LSF cluster purely
inside the R session. We suggest to save the output into permanent files in
the jobs while not directly retrieving the results on the fly.
bsub_chunk() submits the R chunk. The code chunk should be embraced by
{...}. For example, NMF::nmf() normally takes very long time to run. We
submit the NMF analysis to the cluster and save the results as an RDS file.
bsub_chunk(name = "example", memory = 10, hours = 10, cores = 4, { fit = NMF::nmf(...) # you better save `fit` into a permanent file in an absolute path saveRDS(fit, file = "/path/to/fit.rds") })
In the following examples, we use Sys.sleep(5) to simulate a chunk of code which
runs for a short time.
bsub_chunk( { Sys.sleep(5) })
The bsub_chunk() prints the bsub command and the value returned by
bsub_chunk() is the job ID from LSF cluster.
Set job name, memory, running time and number of cores:
bsub_chunk(name = "example", memory = 10, hours = 10, cores = 4, { Sys.sleep(5) })
If name is not specified, an internal name calculated by digest::digest()
on the chunk is automatically assigned. The unit of memory is GB.
The R chunk is saved into a temporary R script and called by Rscript command
when it is executed on the cluster. A lot of LSF clusters have customized installation
of R, which means, calling Rscript is specific for every LSF cluster, thus, you
need to configure how to call the Rscript command. By default, it simply calls Rscript
with the default R version installed on the cluster.
To set Rscript calling with a specific version or in a specific path, you need to
configure the bsub_opt$call_Rscript option. The value for bsub_opt$call_Rscript
should be a user-defined function where the R version in the only argument. The default
value for bsub_opt$call_Rscript is
function(version) "Rscript"
which ignores the R version. If you want to specify Rscritp with a specific path, you
can set bsub_opt$call_Rscript as:
bsub_opt$call_Rscript = function(version) "/the/absolute/path/of/Rscript"
To make it more flexible, the R version can be used when setting how to call Rscript.
By default, when installing R, R will installed into folder with the version name of e.g.
/.../3.6/..., thus, if there are several R versions are installed on your cluster,
you can set bsub_opt$call_Rscript as:
library(GetoptLong) bsub_opt$call_Rscript = function(version) { qq("/the/absolute/path/of/@{gsub('\\.\\d+$', '', version)}/Rscript") }
Here qq() is from GetoptLong package which does variable
interpolation. You can use
similar packages such as glue here.
Later, the R version can be easily switched by setting bsub_opt$R_version
or the R_version argument in bsub_chunk() (The value of R_version is sent
to call_Rscript function). E.g:
bsub_chunk(name = "example", R_version = "3.6.0", { Sys.sleep(5) })
Or set R_version as a global parameter:
bsub_opt$R_version = "3.6.0" bsub_chunk(name = "example", { Sys.sleep(5) })
On DKFZ ODCF cluster, software with different versions are managed by
Environment Modules.
bsub_opt$call_Rscript was set as follows:
function(version) { qq("module load gcc/7.2.0; module load java/1.8.0_131; module load R/@{version}; Rscript") }
The module loading for gcc/7.2.0 and java/1.8.0_131 ensures that R
packages depending on specific C/Java libraries can be successfully loaded.
So, if R_version is set to 4.0.0, the Rscript call would be
module load gcc/7.2.0; module load java/1.8.0_131; module load R/4.0.0; Rscript
which makes sure the Rscript from R-4.0.0 is used.
Similarlly, if you use conda for managing
different versions of software, you can also choose R with different versions
by setting a proper bsub_opt$call_Rscript. Let assume you have conda
environments for different R versions with the name schema R_$version (e.g.
R_3.6.0), then you can set bsub_opt$call_Rscript as:
bsub_opt$call_Rscript = function(version) { qq("conda activate R_@{version}; Rscript") }
In previous examples, we load the gcc/7.2.0 and java/1.8.0_131 modules, or
activate the conda environment as parts of the command callling Rscript.
These bash-level initialization can also be set by sh_head which adds shell
commands as header in the bash script that is used for job submission. E.g.,
we can do the other way:
bsub_opt$call_Rscript = function(version) qq("module load R/@{version}; Rscript") bsub_chunk(name = "example", sh_head = c("module load gcc/7.2.0", "module load java/1.8.0_131"), { Sys.sleep(5) })
Or set sh_head as a global option:
bsub_opt$call_Rscript = function(version) qq("module load R/@{version}; Rscript") bsub_opt$sh_head = c("module load gcc/7.2.0", "module load java/1.8.0_131") bsub_chunk(name = "example", { Sys.sleep(5) })
One usage of this functionality is to load pandoc module if the rmarkdown
is used in the code chunk (on DKFZ ODCF cluster):
bsub_chunk(name = "example", sh_head = "module load pandoc/2.2.1", { library(rmarkdown) render(...) })
The packages that are needed can be directly added in the code chunk:
bsub_chunk(name = "example", { library(package1) library(package2) Sys.sleep(5) })
Or assign by packages argument:
bsub_chunk(name = "example", packages = c("package1", "package2"), { Sys.sleep(5) })
Or set it as a global parameter:
bsub_opt$packages = c("package1", "package2") bsub_chunk(name = "example", { Sys.sleep(5) })
There is a special value _in_session_ for packages argument that loads all packages in the current R session.
library(foo) library(bar) bsub_chunk(name = "example", packages = "_in_session_", { Sys.sleep(5) })
The R variables that are defined outside the code chunk and need to be used
inside the code chunk can by specified by variables argument:
foo = 1 bsub_chunk(name = "example", variables = "foo", { bar = foo Sys.sleep(5) })
variables argument has a special value _all_functions_ that loads all functions defined in the global environment.
f1 = function() 1 f2 = function() 2 bsub_chunk(name = "example", variables = "_all_functions_", { f1() f2() Sys.sleep(5) })
If multiple jobs use the same variables, they can be specified via share argument. In this case, the shared variables are only saved into
temporary files once. Note these temporary are not deleted automatically since they do not know whether all jobs which reply on them are finished.
Users need to manually delete them when all jobs are done.
foo = 1 for(i in 1:10) { bsub_chunk(name = paste0("example", i), share = "foo", { bar = foo Sys.sleep(5) }) }
If you have too many external variables that are used in the code chunk or
they are used in multiple jobs, you can directly save the workspace or the objects as an
image and specify the image argument:
save.image(file = "/path/foo.RData") # or # save(var1, var2, ..., file = "...") bsub_chunk(name = "example", image = "/path/foo.RData", { ... Sys.sleep(5) })
Or set the image file as a global parameter:
save.image(file = "/path/foo.RData") bsub_opt$image = "/path/foo.RData" bsub_chunk(name = "example", { ... Sys.sleep(5) })
Absolute paths should be used instead of relative paths.
Please note, image files can be shared between different jobs and they are not
deleted after all the jobs are finished, as a comparison, variables are
saved into separated temporary files for different jobs even when the variable
names are the same, and they are deleted after the jobs are finished.
If the code chunk replies on the working directory, it can be specified by working_dir argument:
bsub_chunk(name = "example", working_dir = "/path" { Sys.sleep(5) })
Or set it as a global parameter:
bsub_opt$working_dir = "/path" bsub_chunk(name = "example", { Sys.sleep(5) })
Note it is not recommended to let all file pathes in the jobs be relative or be affected by the working directory. It is recommended to use absolute path everywhere in the job.
The last variable in the code chunk can be saved by setting save_var = TRUE
and retrieved back by retrieve_var() by specifying the job ID.
retrieve_var() waits until the job is finished.
job_id = bsub_chunk(name = "example2", save_var = TRUE, { Sys.sleep(10) 1+1 }) retrieve_var(job_id)
However, it is not recommended to directly retrieve the returned value from the code chunk. Better choice is to save the variable into permanent file in the code chunk so you don't need to rerun the code in the future which normally has very long runing time, E.g.:
bsub_chunk(name = "example", { ... save(...) # or saveRDS(...) })
There is a flag file to mark whether the job was successfully finished or not.
If the job has been successfully done, the job with the same name will be
skipped. enforce argument controls how to rerun the jobs with the same
names. If it is set to TRUE, jobs will be rerun no matter they are done or not.
Sys.sleep(10)
bsub_chunk(name = "example", enforce = FALSE, { Sys.sleep(5) })
enforce can be set as a global parameter:
bsub_opt$enforce = FALSE bsub_chunk(name = "example", { Sys.sleep(5) })
Since bsub_chunk() returns the job ID, it is can be used to specify the dependency in other jobs.
The value for dependency can be a vector of job IDs.
job1 = bsub_chunk(name = "example1", { Sys.sleep(5) }) bsub_chunk(name = "example2", dependency = job1, { Sys.sleep(5) })
bsub_chunk() has two arguments temp_dir and output_dir. temp_dir is used for the temporary R script
and sh files. output_dir is used for the flag files and the output files from LSF cluster.
bsub_chunk(name = "example", temp_dir = ..., output_dir = ..., { Sys.sleep(5) })
They can be set as global parameters. The value of output_dir is by default set as the same as temp_dir.
bsub_opt$temp_dir = ... bsub_opt$output_dir = ... bsub_chunk(name = "example", { Sys.sleep(5) })
To remove temporary files in temp_dir, run clear_temp_dir() function.
You can run code chunk from a script by specifying the starting line number
and the ending line number. The R script is specified by script argument,
the starting line number and the ending line number are specified by start
and end arguments. (Note this functionality has not been tested yet.)
bsub_chunk(name = "example", script = "/path/foo.R", start = 10, end = 20, ...)
Assuming you are editing foo.R very offen and the line numbers that you want
to run change from time to time, you can add tags in the R script and
specifying start and end by those tags. In following example which is the
source code of foo.R, we add tags for the code chunk we want to run:
... # BSUB_START you code chunk here # BSUB_END ...
Then you can specify start and end by regular expressions to match them:
bsub_chunk(name = "example", script = "/path/foo.R", start = "^# BSUB_START", end = "^# BSUB_END", ...)
Setting local = TRUE directly runs the code chunk in the same R session (do not submit
to the cluster).
bsub_chunk(name = "example", local = TRUE, { cat("blablabla...\n") })
The nice thing for bsub package is you can programmatically submit many of
jobs. Assuming we have a list of samples where the sample IDs are saved in
sample_id variable, and a list of parameters (in parameters variable) to
test, we want to apply the analysis by analyze() function to each sample
with each parameter per single job. We can submit all the jobs as follows:
library(GetoptLong) for(sid in sample_id) { for(param in parameters) { bsub_chunk(name = qq("analysis_@{sid}_@{param}"), variables = c("sid", "param"), packages = ..., other_arguments..., { res = analyze(sid, param) saveRDS(res, file = qq("/path/to/result_@{sid}_@{param}.rds")) }) } }
bsub_script() submits the job from R scripts. The major arguments are the same as in bsub_chunk().
bsub_script("/path/of/foo.R", name = ..., memory = ..., cores = ..., ...)
If the R script needs command-line arguments, they can be specified by argv.
bsub_script("/path/of/foo.R", argv = "--a 1 --b 3", ...)
When you have a list of jobs with the same argument names but with different
argument values, you can construct argv by glue::glue() or
GetoptLong::qq() to construct the argv string:
library(GetoptLong) for(a in 1:10) { for(b in 11:20) { bsub_script("/path/foo.R", argv = qq("-a @{a} --b @{b}"), ...) } }
The command-line arguments of your R script can also specified as arguments of bsub_script(),
but with . prefix.
bsub_script("/path/foo.R", .a = 1, .b = 3, ...)
Then for the same example previously for submitting a list of jobs, it can be written as:
for(a in 1:10) { for(b in 11:20) { bsub_script("/path/foo.R", .a = a, .b = b, ...) } }
The R scripts should be used in the absolute paths.
Note the bash environment can be initialized by setting the sh_head option.
bsub_cmd()submits shell commands. Basically it is similar as bsub_script():
bsub_cmd("samtools sort ...", name = ..., memory = ..., cores = ..., ...) bsub_cmd(c("cmd1", "cmd2", ...), name = ..., memory = ..., cores = ..., ...)
The binary and the arguments should all be set in the first argument of
bsub_cmd(). Remember to use glue::glue() or GetoptLong::qq() to
construct the commands if they contain variable arguments, e.g:
for(bam in bam_file_list) { bsub_cmd(qq("samtools sort @{bam} ... "), name = qq("sort_@{basename(bam)}"), memory = ..., cores = ..., ...) }
bjobs() or just entering bjobs gives a summary of running and pending
jobs. Job status (by default is RUN and PEND) is controlled by status
argument. Number of most recent jobs is controlled by max argument.
Filtering on the job name is controlled by filter argument. In the following
example, we submit four tiny jobs.
for(i in 1:4) { bsub_chunk(name = paste0("example_", i), { Sys.sleep(5) }) } bjobs
There is one additional column RECENT in the summary table which shows the order
of the jobs with the same job name. The most recent job has the value 1.
for(i in 1:2) { bsub_chunk(name = "example", { Sys.sleep(5) }) }
bjobs(status = "all", filter = "example")
brecent() by default returns 20 most recent jobs of "all" status. You can
simply type brecent without the brackets.
brecent
There are some helper functions which only list running/pending/done/failed jobs:
bjobs_runningbjobs_pendingbjobs_donebjobs_exitbjobs_barplot() makes a barplot of numbers of jobs per day.
bjobs_barplot()

bjobs_timeline() draws the duration of each job. In the plot, each segment represents
a job and the width corresponds to its duration.
bjobs_timeline()

bkill(job_id) kills a job or a list jobs.job_log(job_id) prints the log of a specified running/finished/failed job. A vector
of jobs can also be sent at the same time that last 10 lines of each job are printed.check_dump_files() searches the dump files (core.xxx by LSF cluster or .RDataTmpxxx by R).ssh_connect() establishes the SSH connection to the submission node if it is lost.Type bsub_opt gives you a list of global options. Values can be set by in a
form of bsub_opt$opt = value. All the values can be reset by bsub_opt(RESET = TRUE).
bsub_opt
Or a more readable text:
bconf
Simply running monitor() opens a shiny app where you can query and manage jobs.
monitor()
Following are examples of the job monitor.
The job summary table:

Job log:

Job dependency tree:

Kill jobs:

sessionInfo()
Add the following code to your website.
For more information on customizing the embed code, read Embedding Snippets.