Array Jobs

From Research Computing Center Wiki
Revision as of 10:26, 17 June 2021 by Ben (talk | contribs)
Jump to navigation Jump to search

Introduction

An array job is a collection of jobs (called array job "elements") initiated from a single submission script. Array jobs work well for problems that are embarassingly parallel, meaning a problem can be easily split up into concurrently running tasks that are not dependent on one another. Imagine you have 10 input files that you want to perform the same action(s) against. Rather than looping through the input one at a time, or rather than writing 10 almost identical submission scripts, you could write and submit one array job submission script.

Example Submission Scripts

Numbered Input Files

Writing an array job submission script is hardly different from any other type Slurm submission script. The two key things to remember are the Slurm array header (#SBATCH --array), and the SLURM_ARRAY_TASK_ID environment variable. Below is an array job submission script in which there are 5 input files to be ran as arguments for myScript.R, assuming the input files were named myinput-1, myinput-2, myinput-3, etc...

#!/bin/bash

#SBATCH --job-name=array-test
#SBATCH --partition=batch
#SBATCH --ntasks=1
#SBATCH --mem=20gb
#SBATCH --time=1:00:00
#SBATCH --array=1-5

ml R/4.0.0-foss-2019b

Rscript myScript.R myinput-${SLURM_ARRAY_TASK_ID}

Submitting the above script would create five array job elements as shown below:

bc06026@b1-24 arraytest$ sbatch sub.sh
Submitted batch job 3341751
bc06026@b1-24 arraytest$ squeue --me
             JOBID PARTITION     NAME     USER ST       TIME  NODES NODELIST(REASON)
         3341751_1     batch array-te  bc06026  R       0:08      1 c4-9
         3341751_2     batch array-te  bc06026  R       0:08      1 c4-21
         3341751_3     batch array-te  bc06026  R       0:08      1 c4-21
         3341751_4     batch array-te  bc06026  R       0:08      1 c4-21
         3341751_5     batch array-te  bc06026  R       0:08      1 c4-11

As you can see in the squeue --me output, by submitting this one submission script, we have 5 jobs running concurrently. Each one of these jobs is allocated the resources requested in the submission script and is running the commands:

ml R/4.0.0-foss-2019b

Rscript myScript.R myinput-${SLURM_ARRAY_TASK_ID}

with ${SLURM_ARRAY_TASK_ID} being replaced by one of the numbers in the range defined in the --array Slurm header. This ensures each array job element is working one unique input file, but as you can see, this does require naming your files a certain way to leverage the SLURM_ARRAY_TASK_ID environment variable, which may not always be desirable.

Non-Numbered Input Files

Sometimes it will make more sense for the names of your input files to not have a numbered naming scheme. In this scenario input files can be mapped to a SLURM_ARRAY_TASK_ID by creating a separate file in your working directory, containing just a list of your input files, one file name per line. Then awk can be used to map each file name's line number to SLURM_ARRAY_TASK_ID. For example, say have three files we want to distribute among three array job elements. We can create a file called input.lst, listing each input file, one line at a time:

testdata.txt
aninputfile.txt
moredata.txt

Then in our submission script, we can reference each file name like this:

#!/bin/bash

#SBATCH --job-name=array-test
#SBATCH --ntasks=1
#SBATCH --partition=batch
#SBATCH --mem=20gb
#SBATCH --time=1:00:00
#SBATCH --array=1-3

ml R/4.0.0-foss-2019b

file=$(awk "NR==${SLURM_ARRAY_TASK_ID}" input.lst)

Rscript myScript.R $file

In the above submission script, ${SLURM_ARRAY_TASK_ID} is being replaced by one of the integers in the range defined by #SBATCH --array, for each array job element. The awk built-in variable NR is getting the content of the line number given of input.lst. This ensures each array job element gets a single unique input file. One way that you can quickly create a file containing a list of your input files would be to redirect the output of ls. For example, say you had 100+ .csv's in your current directory that you wanted to spread across an array job. You could create your input list file with the command ls *.csv > input.lst.

Further Reading

For more information on Slurm array jobs, please see Slurm's documentation.