Gaussian16 Linda (multinode)

Gaussian is a general purpose suite of electronic structure programs. The “Linda” version adds multi-node (larger parallel jobs) capability to Gaussian.

Versions g16c01 Linda is installed on the CSF. It is available as binaries only. The source code is not available on the CSF.

Gaussian 16 (non-Linda) is also available on the CSF.
Gaussian 09 (non-Linda) is also available on the CSF.

Restrictions on use

Gaussian Linda is only available to members for Prof. Kaltsoyannis’s group. All requests to access this version must be approved by Prof. Kaltsoyannis.

Please contact us via our help form to request access to Gaussian 16 Linda.

Set up procedure

There is only one version available on the CSF – optimized for Intel “Haswell” architectures or equivalently any CPU that provides AVX2 vector extensions. This includes all Intel and AMD compute nodes in the CSF.

The detectcpu modulefile must be loaded inside your jobscript, not on the login node.

After being added to the relevant unix group, you will be able to access the executables by loading the modulefile:

module load apps/binapps/gaussian/g16c01_em64t_detectcpu_linda

Gaussian MUST ONLY be run in batch. Please DO NOT run g16 on the login nodes. Computational work found to be running on the login nodes will be killed WITHOUT WARNING.

Gaussian Scratch

Gaussian uses an environment variable $GAUSS_SCRDIR to specify a directory path for where to write scratch (temporary) files (two-electron integral files, integral derivative files and a read-write file for temporary workings).

It is set to your scratch directory (~/scratch) when you load the modulefile. This is a Lustre filesystem which provides good I/O performance. Do not be tempted to use your home directory for Gaussian scratch files – the files can be huge making the home area at risk of going over quota. So this gives you a sensible default location for temp files if you wish.

Using a Scratch Directory per Job

We recommend using a different scratch directory for each job. This improves file access times if you run many jobs – writing 1000s of scratch files to a single directory can slow down your jobs. It is much better to create a directory for each job within your scratch area. It is also then easy to delete the entire directory if Gaussian has left unwanted scratch files behind.

# In the jobscript:
export GAUSS_SCRDIR=/scratch/$USER/gau_temp_$SLURM_JOB_ID
mkdir -p $GAUSS_SCRDIR

See below for example jobscripts that do this.

Using a node-local directory

A faster, but smaller, local $TMPDIR on each compute node is available should users prefer to use that. It can be more efficient if you have a need to create lots of small files, but space is limited. The AMD 168-core Genoa nodes in the multicore partition all have a 1.6TB local disk. Use this with caution – if you fill the local disk, your job will crash!.

# Slurm creates the temp $TMPDIR. FOR SMALL FILES ONLY.
# It also deletes the folder (and all files) at the end of the job.
# You should take a copy of any files from this area within your jobscript.
# You will NOT have access to this temp area after the job has finished.
export GAUSS_SCRDIR=$TMPDIR

If your job writes a huge .rwf, say, then it will likely run out of space in this area and your job will crash.

Cleaning up scratch files

Gaussian should delete scratch files automatically when a job completes successfully or dies cleanly. However, it often fails to do this. Scratch files are also not deleted when a job is killed externally or terminates abnormally so that you can use the scratch files to restart the job (if possible). Consequently, leftover files may accumulate in the scratch directory, and it is your responsibility to delete these files. Please check periodically whether you have a lot of temporary Gaussian files that can be deleted.

Very large Gaussian scratch files

Occasionally some jobs create .rwf files which are very large (several TB). The batch system will not permit a job to create files bigger than 4TB. If your gaussian job fails and the .rwf file is 4TB then it may be that this limit has prevented your job from completing. You should re-run the job and in your input file request that the .rwf file be split into multiple files. For example to split the file into two 3TB files:

%rwf=/scratch/$USER/myjob/one.rwf,3000GB,/scratch/$USER/myjob/two.rwf,3000GB

Number of Cores

If your Gaussian input file contains the number of cores – e.g.:

%NProcsShared
%nprocs

then you must remove this line from your input file.

The Linda modulefile will work out the correct number of cores to use, based on your jobscript settings.

Parallel batch job

You must submit your job to the hpcpool partition on the CSF – this is the only multi-node capable partition. Our installation will use the fast InfiniBand network for Linda communication. Submitting to the multicorepartition (to run on a single compute node) will not work (you can simply use non-Linda Gaussian for those jobs.)

#!/bin/bash --login
#SBATCH -p hpcpool          # (or --partition=) Use the HPC Pool nodes
#SBATCH -N 4                # (or --nodes=) Number of compute nodes (each has 32 cores.)
#SBATCH --ntasks-per-node=1 # Number of Linda workers to run on each compute node (1 is recommended!)
#SBATCH -c 32               # (or --cpus-per-task=) Number of cores each Linda worker will use on its assigned compute node.
#SBATCH -t 1-0              # Maximum wallclock time (1-0 is 1 day, max permitted is 4-0.)
##### You must have been given permission to run in an HPC Pool project
#SBATCH -A hpc-projectcode  # The HPC Account code is required

## Set up scratch dir (please do this!)
export GAUSS_SCRDIR=/scratch/$USER/gau_temp_$SLURM_JOB_ID
mkdir -p $GAUSS_SCRDIR

# Load g16 linda
module purge
module load apps/binapps/gaussian/g16c01_em64t_detectcpu_linda

$g16root/g16/g16 < file.inp > file.out

Submit the job using sbatch jobscript where jobscript is the name of your jobscript.

Automatic Settings

Note that the Gaussian Linda module will make automatically a lot of the environment settings required by Gaussian. These include:

# The Linda modulefile will set these variables automatically. In general, DO NOT SET THESE YOURSLEF!
GAUSS_PDEF or GAUSS_CDEF   # to specify the number of cores each Linda worker will use (DO NOT SET YOURSELF)
GAUSS_MDEF                 # to specify the memory to use (default is 5GB per core on HPC Pool nodes)
GAUSS_WDEF                 # the names of the nodes to use for Linda (DO NOT SET YOURSELF)
GAUSS_SDEF                 # the remote process launcher - Slurm's "srun" (DO NOT SET YOURSELF)
GAUSS_LFLAGS2              # the linda stacksize (default is "--LindaOptions -s 20000000")
GAUSS_LFLAGS               # optional flags, e.g., "-vv" for debugging (default is unset)

If you want to see what the above variables have been set to, add the following to your jobscript after the module load line:

env | grep GAUSS_

Optional Settings

NOTE: We DO NOT recommend that you change the defaults. But we document them here for completeness.

It is possible to override the default amount of memory per core used by Gaussian, by setting the following before you load the modulefile:

export GB_PER_CORE=4      # Some number of GB per core. Using too much will crash your job. Default is 5.
module load ...

It is also possible to increase the number of Linda workers per node, which defaults to one. Each Linda worker will use all of the cores available to it on the compute nodes. The default of 1 is strongly recommended, and increasing this will likely reduce performance. But you can set this before loading the modulefile if you want to experiment:

#SBATCH --ntasks-per-node=2   # Some number of Linda workers per node. The default of 1 is usually best.
module load ...

Further info

Last modified on August 7, 2025 at 11:53 am by George Leaver