VASP at TACC
Last update: May 06, 2024
Vienna Ab initio Simulation Package (VASP) is a computer program for atomic scale materials modeling, e.g. electronic structure calculations and quantum-mechanical molecular dynamics, from first principles.
VASP Licenses
VASP requires an individual license. TACC's HPC support license allows us to install only the compiled VASP executables and grant the access to licensed users who have accounts on our systems. We are not allowed to share source code.
If you have your own license and want to use your own compilation, you may install it in your own account. If you wish to use TACC's installed version, then TACC will have to verify your license. Submit a support ticket and include the following information:
- full name
- affiliated institution
- TACC login ID
- email address that was used by your VASP PI to register you under that license on VASP portal
TACC will confirm the license by using this email address on the VASP portal. If you get the license from Material Design, TACC will use that information to contact Material Design and confirm the license.
Installations
The latest stable release of VASP is installed on Stampede3, Lonestar6 and Frontera systems.
Important
You cannot load any VASP module until your license is confirmed. See above.
Once your license is active, use Lmod's module
commands to explore other VASP installations, for example:
login1$ module spider vasp
login1$ module spider vasp/6.3.0
Then, either interactively or via a batch script, load the appropriate module:
login1$ module load vasp/6.3.0
Running VASP Jobs
You may use and customize the following sample job scripts for VASP jobs on TACC's Stampede3, Lonestar6 and Frontera resources.
Sample Job Script: VASP on Stampede3
Important
DO NOT run VASP using Stampede3's SPR nodes!
TACC staff has noticed many VASP jobs causing issues on the SPR nodes and impacting overall system stability and performance.
Please run your VASP jobs using either the SKX or ICX nodes.
The issue appears to be memory overuse: each SPR node has 112 cores on 2 sockets and only 128 GB memory in total, so each core has about 1GB memory. Compare to the SKX (48 nodes, 192GB) and ICX (80 cores and 256GB) nodes, the SPR nodes have much less total and per task memory. Therefore jobs running on the SPR nodes will require more nodes, yet need less tasks per node. This is not an efficient use of resources.
Tip
TACC staff recommends that former Stampede2 users first conduct numerical consistency tests using the skx
compute nodes. From there, determine the best node and queue configuration for your application.
The script below requests 4 nodes and 192 tasks, for a maximum of four hours in Stampede3's skx
queue (SKX compute nodes).
#!/bin/bash
#SBATCH -J vasp
#SBATCH -o vasp.%j.out
#SBATCH -e vasp.%j.err
#SBATCH -n 192
#SBATCH -N 4
#SBATCH -p skx
#SBATCH -t 4:00:00
#SBATCH -A projectnumber
module load vasp/5.4.4.pl2
ibrun vasp_std > vasp_test.out
Sample Job Script: VASP on Frontera
The script below submits a VASP job to Frontera's normal
queue (CLX compute nodes), requesting 4 nodes and 224 tasks for a maximum of 4 hours.
#!/bin/bash
#SBATCH -J vasp
#SBATCH -o vasp.%j.out
#SBATCH -e vasp.%j.err
#SBATCH -n 224
#SBATCH -N 4
#SBATCH -p normal
#SBATCH -t 4:00:00
#SBATCH -A projectnumber
module load vasp/5.4.4.pl2
ibrun vasp_std > vasp_test.out
Sample Job Script: VASP on Lonestar6
The script below submits a VASP job to Lonestar6's normal
queue ( Milan compute nodes), requesting 2 nodes and 256 tasks for a maximum of 4 hours.
#!/bin/bash
#SBATCH -J vasp
#SBATCH -o vasp.%j.out
#SBATCH -e vasp.%j.err
#SBATCH -n 256
#SBATCH -N 2
#SBATCH -p normal
#SBATCH -t 4:00:00
#SBATCH -A projectnumber
module load vasp/5.4.4.pl2
ibrun vasp_std > vasp_test.out