Gromacs 2016.4 (CPU only, with and without Plumed)
Overview
GROMACS is a package for computing molecular dynamics, simulating Newtonian equations of motion for systems with hundreds to millions of particles. GROMACS is designed for biochemical molecules with complicated bonded interactions (e.g. proteins, lipids, nucleic acids) but can also be used for non-biological systems (e.g. polymers).
-v
flag to your mdrun
command.
It will write to a log file every second for the duration of your job and can lead to severe overloading of the file servers.
Significant Change in this Version
Within Gromacs 2016, the different gromacs commands (e.g., mdrun
, grompp
, g_hbond
) should now be run using the command:
gmx command
where command
is the name of the command you wish to run (without any g_
prefix), for example:
gmx mdrun
The gmx
command changes its name to reflect the gromacs flavour being used but the command
does not change. For example, if using the mdrun
command:
# New 2016.4 method # Previous 5.0.4 method # ================= # ===================== gmx mdrun mdrun gmx_d mdrun mdrun_d mpirun -n $NSLOTS gmx_mpi mdrun mpirun -n $NSLOTS mdrun_mpi mpirun -n $NSLOTS gmx_mpi_d mdrun mpirun -n $NSLOTS mdrun_mpi_d
The complete list of command
names can be found by running the following on the login node:
gmx help commands
# The following commands are available: anadock dyecoupl mdmat sans anaeig dyndom mdrun sasa analyze editconf mindist saxs angle eneconv mk_angndx select bar enemat morph sham bundle energy msd sigeps check filter nmeig solvate chi freevolume nmens sorient cluster gangle nmtraj spatial clustsize genconf order spol confrms genion pairdist tcaf convert-tpr genrestr pdb2gmx traj covar grompp pme_error trjcat current gyrate polystat trjconv density h2order potential trjorder densmap hbond principal tune_pme densorder helix rama vanhove dielectric helixorient rdf velacc dipoles help rms view disre hydorder rmsdist wham distance insert-molecules rmsf wheel do_dssp lie rotacf x2top dos make_edi rotmat xpm2ps dump make_ndx saltbr
Notice that the command names do NOT start with g_
and do NOT reference the flavour being run (e.g., _mpi_d
). Only the main gmx
command changes its name to reflect the flavour (see below for list of modulefiles for the full list of flavours available).
To obtain more help about a particular command run:
gmx help command
For example
gmx help mdrun
Helper scripts
To assist with moving to the new command calling method, we have recreated some of the individual commands that you may have used in your jobscript. For example, you can continue to use mdrun
(or mdrun_d
) instead of the new gmx mdrun
(or gmx_d mdrun
) in this release. These extra commands are automatically included in your environment when you load the gromacs modulefiles. This old method uses the flavour of gromacs in the command name (see above for comparison of new and old commands).
However, please note that the following commands are new to 2016.4 and so can only be run using the new method (gmx command
):
# New commands that can only be run using: gmx command check help convert-tpr insert-molecules distance pairdist dump sasa freevolume solvate gangle view
Available Flavours
For version 2016.4 we have compiled multiple versions of Gromacs, each of which is optimised for a particular CPU architecture. In addition we have compiled versions with and without the Plumed plugin. The module file has been written to auto-detect which CPU the compute node is using and to automatically select the correct Gromacs executable. If you want to ensure you get a particular level of opimisation specify an architecture in the jobscript e.g. -l skylake
.
2016.4 for and Ivybridge (and Haswell, Broadwell and Skylake nodes) only
With AVX optimisation.
2016.4 for Haswell and Broadwell (and Skylake) nodes only
With AVX2 optimisation.
2016.4 for Skylake nodes only
With AVX-512 optimisation.
Restrictions on use
GROMACS is free software, available under the GNU General Public License.
Set up procedure
You must load the appropriate modulefile:
module load modulefile
replacing modulefile with one of the modules listed in the table below. The module file will auto-detect and pick a version of Gromacs with AVX optimisations to match the CPU of the compute node(s) you are assigned.
Version | Modulefile | Notes | Typical Executable name |
---|---|---|---|
Single precision multi-threaded (single-node) | apps/intel-17.0/gromacs/2016.4/single | non-MPI | mdrun or gmx mdrun |
Double precision multi-threaded (single-node) | apps/intel-17.0/gromacs/2016.4/double | non-MPI | mdrun_d or gmx_d mdrun |
Single precision MPI | apps/intel-17.0/gromacs/2016.4/single_mpi | For MPI | mdrun_mpi or gmx_mpi mdrun |
Double precision MPI (single-node) | apps/intel-17.0/gromacs/2016.4/double_mpi | For MPI | mdrun_mpi_d or gmx_mpi_d mdrun |
Single precision multi-threaded (single-node) with Plumed | apps/intel-17.0/gromacs/2016.4-plumed/single | non-MPI | mdrun or gmx mdrun |
Double precision multi-threaded (single-node) with Plumed | apps/intel-17.0/gromacs/2016.4-plumed/double | non-MPI | mdrun_d or gmx_d mdrun |
Single precision MPI with Plumed | apps/intel-17.0/gromacs/2016.4-plumed/single_mpi | For MPI | mdrun_mpi or gmx_mpi mdrun |
Double precision MPI (single-node) with Plumed | apps/intel-17.0/gromacs/2016.4-plumed/double_mpi | For MPI | mdrun_mpi_d or gmx_mpi_d mdrun |
Running the application
Please do not run GROMACS on the login node.
Important notes regarding running jobs in batch
Ensure you have created a directory containing the required input data files.
Please NOTE the following which is important for running jobs correctly and efficiently:
Ensure you inform gromacs how many cores it can use. This is done using the $NSLOTS
variable which is automatically set for you in the jobscript to be the number of cores you request in the jobscript header (see later for complete examples). You can use either of the following methods depending whether you want a multi-core job (running on a single compute node) or a larger job running across multiple compute nodes:
# Multi-core (single-node) or Multi-node MPI jobs mpirun -n $NSLOTS mdrun_mpi # Old method (v5.0.4 and earlier) mpirun -n $NSLOTS mdrun_mpi_d # Old method (v5.0.4 and earlier) mpirun -n $NSLOTS gmx_mpi mdrun # New method (v5.1.4 and later) mpirun -n $NSLOTS gmx_mpi_d mdrun # New method (v5.1.4 and later)
or
# Single-node multi-threaded job export OMP_NUM_THREADS=$NSLOTS # Do this for all versions mdrun # Old method (v5.0.4 and earlier) mdrun_d # Old method (v5.0.4 and earlier) export OMP_NUM_THREADS=$NSLOTS # Do this for all versions gmx mdrun # New method (v5.1.4 and later) gmx_d mdrun # New method (v5.1.4 and later)
The examples below can be used for single precision or double precision gromacs. Simply run mdrun
(single precision) or mdrun_d
(double precision).
-v
flag to your mdrun
command.
It will write to a log file every second for the duration of your job and can lead to severe overloading of the file servers.
Multi-threaded single-precision on Intel nodes, 2 to 32 cores
Note that GROMACS 2016.4 (unlike v4.5.4) does not support the -nt
flag to set the number of threads when using the multithreaded OpenMP (non-MPI) version. Instead set the OMP_NUM_THREADS
environment variable as shown below.
An example batch submission script to run the single-precision mdrun executable with 12 threads:
#!/bin/bash --login #$ -cwd #$ -pe smp.pe 12 # Can specify 2 to 32 cores in smp.pe # 2-16 includes Ivybridge, Haswell, Broadwell, Skylake # 17-24 includes Haswell, Broadwell, Skylake # 25-28 includes Broadwell and Skylake # 28-32 uses Skylake only # Can force use of a particular architecture with #$ -l Ivybridge/Haswell/Broadwell/Skylake module load apps/intel-17.0/gromacs/2016.4/single export OMP_NUM_THREADS=$NSLOTS mdrun # # This is the old naming convention (it will still work in this release) # The new gromacs convention is to run: gmx mdrun
Submit with the command: qsub scriptname
The system will run your job on a Ivybridge, Haswell, Broadwell or Skylake node depending on what is available. This option goes to the biggest pool of nodes. To get a more optimised run on Haswell/Broadwell you should specify the architecture you require using -l broadwell or -l haswell, see the instructions below.
Multi-threaded double-precision on Intel nodes, 2 to 32 cores
An example batch submission script to run the double-precision mdrun_d
executable with 8 threads:
#!/bin/bash --login #$ -cwd #$ -pe smp.pe 24 module load apps/intel-17.0/gromacs/2016.3/double export OMP_NUM_THREADS=$NSLOTS mdrun_d # # This is the old naming convention (it will still work in this release) # The new gromacs convention is to run: gmx_d mdrun
Submit with the command: qsub scriptname
Single precision MPI (single-node), 2 to 32 cores
An example batch submission script to run the double-precision mdrun_mpi
executable on 8 cores using mpi:
#!/bin/bash --login #$ -cwd #$ -pe smp.pe 8 module load apps/intel-17.0/gromacs/2016.4/single_mpi mpirun -n $NSLOTS mdrun_mpi
Submit with the command: qsub scriptname
Double precision MPI (single-node), 2 to 32 cores
An example batch submission script to run the double-precision mdrun_mpi_d
executable on 8 cores using mpi:
#!/bin/bash --login #$ -cwd #$ -V #$ -pe smp.pe 8 module load apps/intel-17.0/gromacs/2016.4/double_mpi mpirun -n $NSLOTS mdrun_mpi_d # # This is the old naming convention (it will still work in this release) # The new gromacs convention is to run: mpirun -n $NSLOTS gmx_mpi_d mdrun
Submit with the command: qsub scriptname
Single-precision, MPI, 48 cores or more in multiples of 24
An example batch submission script to run the single precision mdrun_mpi
executable with 48 MPI processes (48 cores on two 24-core nodes) with the mpi-24-ib.pe
parallel environment (Intel Haswell nodes using infiniband):
#!/bin/bash --login
#$ -cwd
#$ -pe mpi-24-ib.pe 48 # EG: Two 24-core Intel Haswell nodes
module load apps/intel-17.0/gromacs/2016.4/single_mpi
mpirun -n $NSLOTS gmx_mpi mdrun
Submit with the command: qsub scriptname
Double-precision, MPI, 48 cores or more in multiples of 24
An example batch submission script to run the single precision mdrun_mpi
executable with 48 MPI processes (48 cores on two 24-core nodes) with the mpi-24-ib.pe
parallel environment (Intel Haswell nodes using infiniband):
#!/bin/bash --login #$ -cwd #$ -pe mpi-24-ib.pe 48 # EG: Two 24-core Intel Haswell nodes module load apps/intel-17.0/gromacs/2016.4/double_mpi mpirun -n $NSLOTS gmx_mpi_d mdrun Submit with the command:qsub scriptname
Error about OpenMP and cut-off scheme
If you encounter the following error:
OpenMP threads have been requested with cut-off scheme Group, but these are only supported with cut-off scheme Verlet
then please try using the mpi version of the software. Note that is is possible to run mpi versions on a single node (example above).
Further info
- You can see a list of all the installed GROMACS utilities with the command:
ls $GMXDIR/bin
- GROMACS web page
- GROMACS manuals
- GROMACS user mailing list
Updates
Oct 2018 – 2016.4 installed with AVX, AVX2 and AVX-512 support enabled and patched with Plumed 2.4.0
Oct 2018 – 2016.4 installed with AVX, AVX2 and AVX-512 support enabled
Oct 2018 – 2016.3 installed with AVX, AVX2 and AVX-512 support enabled