LAMMPS at TACC
Last update: June 1, 2022
LAMMPS is a classical molecular dynamics code developed at Sandia National Laboratories and is available under the GPL license. LAMMPS (Large-scale Atomic/Molecular Massively Parallel Simulator) makes use of spatial-decomposition techniques to partition the simulation domain and runs in serial or in parallel using MPI. The code is capable of modeling systems with millions or even billions of particles on a large HPC machine. A variety of force fields and boundary conditions are provided in LAMMPS which can be used to model atomic, polymeric, biological, metallic, granular, and coarse-grained systems.
LAMMPS is installed on the Stampede2, Lonestar6 and Frontera systems.
As of this date, the default versions are 9Jan20 (Stampede2), 20Sep21 (Lonestar6) and 15Apr20 (Frontera). Users are welcome to install different versions of LAMMPS in their own directories (see Building Third Party Software in the Stampede2 User Guide).
$ module spider lammps # list installed LAMMPS versions $ module load lammps # load default version
The LAMMPS module defines a set of environment variables for the locations of the LAMMPS home, binaries and more with the prefix
TACC_LAMMPS. Use the
env command to display the variables:
$ env | grep "TACC_LAMMPS"
Note that each installation's executable name differs. The name of the executable is in the format of
machine can be either
frontera depending on the system. The Stampede2 versions 9Jan20 and 16Mar18, use
lmp_knl and must be submitted to the Stampede2's KNL (not SKX) queues. The LAMMPS GPU executables,
lmp_gpu, can only be submitted to Frontera's GPU queues.
LAMMPS uses spatial-decomposition techniques to partition the simulation domain into small 3d sub-domains, one of which is assigned to each processor. You will need to set suitable values of
-N (number of nodes),
-n (total number of MPI tasks), and
OMP_NUM_THREADS (number of threads to use in parallel regions) to optimize the performance of your simulation.
Sample Job Script: LAMMPS on Stampede2
Refer to Stampede2's Running Jobs section for more Slurm options. To configure this script for Lonestar6 and Frontera, vary the
#!/bin/bash ##SBATCH -J test # Job Name ##SBATCH -A myProject # Your project name ##SBATCH -o test.o%j # Output file name (%j expands to jobID) ##SBATCH -e test.e%j # Error file name (%j expands to jobID) ##SBATCH -N 1 # Requesting 1 node ##SBATCH -n 16 # and 16 tasks ##SBATCH -p normal # Queue name (normal, skx-normal, etc.) ##SBATCH -t 24:00:00 # Specify 24 hour run time module load intel/18.0.2 module load impi/18.0.2 module load lammps/9Jan20 export OMP_NUM_THREADS=1 ibrun lmp_knl -in lammps_input
Example command-line invocations:
LAMMPS with USER-OMP package (e.g. using 2 threads)
ibrun lmp_stampede -sf omp -pk omp 2 -in lammps_input
LAMMPS with USER-INTEL package (e.g. using 2 threads)
ibrun lmp_stampede -sf intel -pk intel 0 omp 2 -in lammps_input
LAMMPS with GPU package (Frontera 15Apr20 only)
lmp_gpu. On Frontera GPU nodes, you could set
-pk gpu 4to utilize all four RTX GPUs available on each node. Set the
-ndirective to a value > 1 to let more than one MPI task share one GPU.
#SBATCH -N 1 # Requesting 1 node #SBATCH -n 16 # and 16 tasks that share 4 GPU #SBATCH -p rtx # Frontera rtx queue # Use all 4 GPUs ibrun lmp_gpu -sf gpu -pk gpu 4 -in lammps_input
You can also run LAMMPS within an
idev session as demonstrated below:
login1$ idev ... c123-456$ module load lammps c123-456$ lmp_stampede < lammps_input
-h option to print out a list of all supported functions and packages:
c123-456$ lmp_stampede -h