Parallel Jobs (Slurm)
Audience
The instructions in this section are for users who have been asked to test applications in the Slurm batch system on the upgraded CSF3, thereby helping us to ensure the upgraded system is running as expected.
(answering these requests will slow down our upgrade work!)
Parallel batch job submission (Slurm)
For jobs that require 2-168 CPU cores – running on one AMD Genoa compute node. A jobscript template is shown below. Please also consult the Partitions page for details on available compute resources.
Please also consult the software page specific for the code / application you are running for advice running your application.
A parallel job script will run in the directory (folder) from which you submit the job. The jobscript takes the form:
#!/bin/bash --login #SBATCH -p multicore # Partition is required. Runs on an Intel hardware. #SBATCH -n numcores # (or --ntasks=) where numcores is between 2 and 168. #SBATCH -t 4-0 # Wallclock limit (days-hours). Required! # Max permitted is 7 days (7-0). # Load any required modulefiles. A purge is used to start with a clean environment. module purge module load apps/some/example/1.2.3 # $SLURM_NTASKS will be set to the numcores given above on the -n line export OMP_NUM_THREADS=$SLURM_NTASKS # Now the commands to be run by the job. openmp-app.exe # Or if running MPI apps (mpirun knows how many cores to use): mpirun mpi-app.exe
In the above jobscript, we must explicitly specify the number of cores.
#SBATCH -n 84 # Causes the $SLURM_NTASKS env var to be set (to 84)
If you require the $SLURM_CPUS_PER_TASKS
env var in your jobscript, then you should specify instead of -n
:
#SBATCH -c 84 # Causes the $SLURM_CPUS_PER_TASKS env var to be set (to 84)
Available Hardware and Resources
Please see the Partitions page for details on available compute resources.