OpenMP: Difference between revisions
No edit summary |
No edit summary |
||
(13 intermediate revisions by 2 users not shown) | |||
Line 1: | Line 1: | ||
[[Category:Sapelo2]] | |||
===Compiling OpenMP code on | ===Compiling OpenMP code on Sapelo2=== | ||
The compilers installed on | The compilers installed on Sapelo2 support shared memory applications using OpenMP. Here are the compiler options to include when using OpenMP: | ||
{| width="100%" border="1" cellspacing="0" cellpadding="2" align="center" class="wikitable unsortable" | {| width="100%" border="1" cellspacing="0" cellpadding="2" align="center" class="wikitable unsortable" | ||
Line 8: | Line 9: | ||
! scope="col" | Compiler | ! scope="col" | Compiler | ||
! scope="col" | Commands | ! scope="col" | Commands | ||
! scope="col" | Option | ! scope="col" | Compile Option | ||
|- | |- | ||
|- | |- | ||
| PGI || pgcc, pgCC, | | PGI || pgcc, pgCC, pgfortran, etc. || -mp | ||
|- | |- | ||
| Intel || icc, ifort || -openmp | | Intel || icc, icpc, ifort || -openmp | ||
|- | |- | ||
| GNU || gcc | | GNU || gcc, g++, gfortran, etc. || -fopenmp | ||
|- | |- | ||
|} | |} | ||
For more information about the compilers on Sapelo2, please see [[Code Compilation on Sapelo2]]. | |||
===Defining number of OpenMP threads=== | ===Defining number of OpenMP threads=== | ||
The number of OpenMP threads can be specified using the environment variable OMP_NUM_THREADS. For example, to define 4 threads: | The number of OpenMP threads can be specified using the environment variable '''OMP_NUM_THREADS'''. For example, to define 4 threads: | ||
For bash/sh: | For bash/sh: | ||
Line 35: | Line 38: | ||
setenv OMP_NUM_THREADS 4 | setenv OMP_NUM_THREADS 4 | ||
</pre> | </pre> | ||
=== Setting thread-core binding === | |||
In general, an OpenMP code will run a lot more efficiently when threads are pinned to a given core on a compute node. Please set the environment variable '''OMP_PROC_BIND=true''' to specify threads' affinity policy (thread-core bind). | |||
For bash/sh: | |||
<pre class="gcommand"> | |||
export OMP_PROC_BIND=true | |||
</pre> | |||
For csh/tcsh: | |||
<pre class="gcommand"> | |||
setenv OMP_PROC_BIND true | |||
</pre> | |||
===Sample script to run a shared memory job using OpenMP=== | ===Sample script to run a shared memory job using OpenMP=== | ||
Sample script to run a program that uses 4 OpenMP threads: | |||
Sample job submission script (sub.sh) to run a program that uses 4 OpenMP threads: | |||
<pre class="gscript"> | <pre class="gscript"> | ||
#!/bin/bash | #!/bin/bash | ||
cd | #SBATCH --job-name=mctest # Job name | ||
#SBATCH --partition=batch # Partition (queue) name | |||
#SBATCH --ntasks=1 # Run a single task | |||
#SBATCH --cpus-per-task=4 # Number of CPU cores per task | |||
#SBATCH --mem=4gb # Job memory request | |||
#SBATCH --time=02:00:00 # Time limit hrs:min:sec | |||
#SBATCH --output=mctest.%j.out # Standard output log | |||
#SBATCH --error=mctest.%j.err # Standard error log | |||
#SBATCH --mail-type=END,FAIL # Mail events (NONE, BEGIN, END, FAIL, ALL) | |||
#SBATCH --mail-user=username@uga.edu # Where to send mail | |||
cd $SLURM_SUBMIT_DIR | |||
export OMP_NUM_THREADS=4 | export OMP_NUM_THREADS=4 | ||
time ./ | export OMP_PROC_BIND=true | ||
module load foss/2022a # load the appropriate module, e.g. foss/2022a | |||
time ./a.out | |||
</pre> | |||
'''Please note''' that the number of cores requested by this job ('''--cpus-per-task=4''') is the same as the number of OpenMP threads that the application will use ('''export OMP_NUM_THREADS=4'''). | |||
===Sample job submission command on Sapelo2=== | |||
To submit a shared memory job that uses 4 OpenMP threads: | |||
<pre class="gcommand"> | |||
sbatch sub.sh | |||
</pre> | </pre> | ||
For more information about running jobs on Sapelo2, please see [[Running Jobs on Sapelo2]]. |
Latest revision as of 11:48, 6 September 2023
Compiling OpenMP code on Sapelo2
The compilers installed on Sapelo2 support shared memory applications using OpenMP. Here are the compiler options to include when using OpenMP:
Compiler | Commands | Compile Option |
---|---|---|
PGI | pgcc, pgCC, pgfortran, etc. | -mp |
Intel | icc, icpc, ifort | -openmp |
GNU | gcc, g++, gfortran, etc. | -fopenmp |
For more information about the compilers on Sapelo2, please see Code Compilation on Sapelo2.
Defining number of OpenMP threads
The number of OpenMP threads can be specified using the environment variable OMP_NUM_THREADS. For example, to define 4 threads:
For bash/sh:
export OMP_NUM_THREADS=4
For csh/tcsh:
setenv OMP_NUM_THREADS 4
Setting thread-core binding
In general, an OpenMP code will run a lot more efficiently when threads are pinned to a given core on a compute node. Please set the environment variable OMP_PROC_BIND=true to specify threads' affinity policy (thread-core bind).
For bash/sh:
export OMP_PROC_BIND=true
For csh/tcsh:
setenv OMP_PROC_BIND true
Sample job submission script (sub.sh) to run a program that uses 4 OpenMP threads:
#!/bin/bash #SBATCH --job-name=mctest # Job name #SBATCH --partition=batch # Partition (queue) name #SBATCH --ntasks=1 # Run a single task #SBATCH --cpus-per-task=4 # Number of CPU cores per task #SBATCH --mem=4gb # Job memory request #SBATCH --time=02:00:00 # Time limit hrs:min:sec #SBATCH --output=mctest.%j.out # Standard output log #SBATCH --error=mctest.%j.err # Standard error log #SBATCH --mail-type=END,FAIL # Mail events (NONE, BEGIN, END, FAIL, ALL) #SBATCH --mail-user=username@uga.edu # Where to send mail cd $SLURM_SUBMIT_DIR export OMP_NUM_THREADS=4 export OMP_PROC_BIND=true module load foss/2022a # load the appropriate module, e.g. foss/2022a time ./a.out
Please note that the number of cores requested by this job (--cpus-per-task=4) is the same as the number of OpenMP threads that the application will use (export OMP_NUM_THREADS=4).
Sample job submission command on Sapelo2
To submit a shared memory job that uses 4 OpenMP threads:
sbatch sub.sh
For more information about running jobs on Sapelo2, please see Running Jobs on Sapelo2.