Parallel Jobs (Slurm)

Audience

The instructions in this section are for users who have been asked to test applications in the Slurm batch system on the upgraded CSF3, thereby helping us to ensure the upgraded system is running as expected.

PLEASE DO NOT SUBMIT A REQUEST ASKING FOR ACCESS TO THE NEW SYSTEM – WE WILL CONTACT YOU AT THE APPROPRIATE TIME!!!
(answering these requests will slow down our upgrade work!)

Parallel batch job submission (Slurm)

For jobs that require 2-168 CPU cores – running on one AMD Genoa compute node. A jobscript template is shown below. Please also consult the Partitions page for details on available compute resources.

Please also consult the software page specific for the code / application you are running for advice running your application.

A parallel job script will run in the directory (folder) from which you submit the job. The jobscript takes the form:

#!/bin/bash --login
#SBATCH -p multicore  # Partition is required. Runs on an Intel hardware.
#SBATCH -n numcores   # (or --ntasks=) where numcores is between 2 and 168.
#SBATCH -t 4-0        # Wallclock limit (days-hours). Required!
                      # Max permitted is 7 days (7-0).

# Load any required modulefiles. A purge is used to start with a clean environment.
module purge
module load apps/some/example/1.2.3

# $SLURM_NTASKS will be set to the numcores given above on the -n line
export OMP_NUM_THREADS=$SLURM_NTASKS

# Now the commands to be run by the job.
openmp-app.exe

# Or if running MPI apps (mpirun knows how many cores to use):
mpirun mpi-app.exe

In the above jobscript, we must explicitly specify the number of cores.

#SBATCH -n 84        # Causes the $SLURM_NTASKS env var to be set (to 84)

If you require the $SLURM_CPUS_PER_TASKS env var in your jobscript, then you should specify instead of -n:

#SBATCH -c 84        # Causes the $SLURM_CPUS_PER_TASKS env var to be set (to 84)

Available Hardware and Resources

Please see the Partitions page for details on available compute resources.

Last modified on April 2, 2025 at 3:38 pm by George Leaver