Table of Contents | ||||||||||||||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
|
...
There are mainly three approaches in preparing an environment on LANTA:
Module system Manual installation (this guide)
Conda environment (please visit Mamba / Miniconda3)
Container (please visit Container (Apptainer / Singularity))
Users should select one over another. They should not be mixed, since library conflicts may occur.
...
Expand | ||
---|---|---|
| ||
|
GPU acceleration
For building an application with GPU acceleration, users can use either PrgEnv-nvhpc
, cudatoolkit/<version>
or nvhpc-mixed
. We recommend using PrgEnv-nvhpc
for completeness.
Expand | ||
---|---|---|
| ||
|
Build target
To enable optimizations that depend on the hardware architecture of LANTA, the following modules should be loaded together with PrgEnv
.
...
Module name
...
Hardware target
...
Note
...
craype-x86-milan
...
AMD EPYC Milan (x86)
...
-
...
craype-accel-nvidia80
...
NVIDIA A100
...
Load after PrgEnv-nvhpc
, cudatoolkit
or nvhpc-mixed
Expand | ||
---|---|---|
| ||
|
Cray optimized libraries
Most Cray optimized libraries become accessible only after loading a PrgEnv
, ensuring compatibility with the selected compiler suite. Additionally, some libraries, such as NetCDF, require loading other specific libraries first. Below is the hierarchy of commonly used cray-*
modules.
...
Expand | |||||
---|---|---|---|---|---|
| |||||
|
CPE version
To ensure backward compatibility after a system upgrade, it is recommended to fix the Cray Programming Environment version using either cpe/<version>
or cpe-cuda/<version>
. Otherwise, the most recent version will be loaded by default.
Expand | ||
---|---|---|
| ||
When unloading
This simply means that the remaining modules are still of the On the other hand, if you use |
1.2 ThaiSC pre-built modules
For user convenience, we provide several shared modules of some widely used software and libraries. These modules were built on top of the HPE Cray Programming Environment, using the CPE toolchain.
CPE toolchain
A CPE toolchain module is a bundle of craype-x86-milan
, PrgEnv-<compiler>
and cpe-cuda/<version>
. The module is defined as a toolchain for convenience and for use with EasyBuild, the framework used for installing most ThaiSC modules.
...
title | [Feb 2024] Current CPE toolchains |
---|
...
CPE toolchain
...
Note
...
cpeGNU/23.03
...
GCC 11.2.0
...
cpeCray/23.03
...
CCE 15.0.1
...
cpeIntel/23.03
...
Deprecated and hidden. It will be removed in the future.
...
cpeIntel/23.09
...
Intel Compiler 2023.1.0
ThaiSC modules
All ThaiSC modules are located at the same module path, so there is no module hierarchy. Executing module avail
on LANTA will display all available ThaiSC modules. For a more concise list, you can use module overview
, then, use module whatis <name>
or module help <name>
to learn more about a specific module.
Users can readily use ThaiSC modules and CPE toolchains to build their applications. Some popular application software are pre-installed as well, for more information, refer to Applications usage.
Code Block | |||||
---|---|---|---|---|---|
| |||||
username@lanta-xname:~> module overview
------------------------------ /lustrefs/disk/modules/easybuild/modules/all ------------------------------
ADIOS2 (2) EasyBuild (1) MAFFT (1) Trimmomatic (1) libaec (1)
Amber (1) Eigen (1) METIS (2) VASP (3) libdeflate (1)
Apptainer (1) FDS (1) MPC (1) VCFtools (1) libgeotiff (1)
Armadillo (1) FastQC (1) MPFR (1) WPS (2) libiconv (1)
AutoDock-vina (1) FortranGIS (1) Mamba (1) WRF (1) libjpeg-turbo (2)
Autoconf (1) FreeXL (1) MrBayes (1) WRFchem (1) libpciaccess (1)
Automake (1) GATK
|
GPU acceleration
For building an application with GPU acceleration, users can use either PrgEnv-nvhpc
, cudatoolkit/<version>
or nvhpc-mixed
. We recommend using PrgEnv-nvhpc
for completeness.
Expand | ||
---|---|---|
| ||
|
Build target
To enable optimizations that depend on the hardware architecture of LANTA, the following modules should be loaded together with PrgEnv
.
Module name | Hardware target | Note |
---|---|---|
craype-x86-milan | AMD EPYC Milan (x86) | - |
craype-accel-nvidia80 | NVIDIA A100 | Load after |
Expand | ||
---|---|---|
| ||
|
Cray optimized libraries
Most Cray optimized libraries become accessible only after loading a PrgEnv
, ensuring compatibility with the selected compiler suite. Additionally, some libraries, such as NetCDF, require loading other specific libraries first. Below is the hierarchy of commonly used cray-*
modules.
...
Expand | |||||
---|---|---|---|---|---|
| |||||
|
CPE version
To ensure backward compatibility after a system upgrade, it is recommended to fix the Cray Programming Environment version using either cpe/<version>
or cpe-cuda/<version>
. Otherwise, the most recent version will be loaded by default.
Expand | ||
---|---|---|
| ||
When unloading
This simply means that the remaining modules are still of the On the other hand, if you use |
1.2 ThaiSC pre-built modules
For user convenience, we provide several shared modules of some widely used software and libraries. These modules were built on top of the HPE Cray Programming Environment, using the CPE toolchain.
CPE toolchain
A CPE toolchain module is a bundle of craype-x86-milan
, PrgEnv-<compiler>
and cpe-cuda/<version>
. The module is defined as a toolchain for convenience and for use with EasyBuild, the framework used for installing most ThaiSC modules.
Expand | ||||||||||
---|---|---|---|---|---|---|---|---|---|---|
| ||||||||||
|
ThaiSC modules
All ThaiSC modules are located at the same module path, so there is no module hierarchy. Executing module avail
on LANTA will display all available ThaiSC modules. For a more concise list, you can use module overview
, then, use module spider <name>
to learn more about each specific module.
Users can readily use ThaiSC modules and CPE toolchains to build their applications. Some popular application software are pre-installed as well, for more information, refer to Applications usage.
Code Block | ||
---|---|---|
| ||
username@lanta-xname:~> module overview --------------------------------- /lustrefs/disk/modules/easybuild/modules/all --------------------------------- ADIOS2 (2) GATK (1) NASM (1) NASM Tcl (12) XZgroff (32) ATK libpng (3) Autotools (12) GDAL (1) Nextflow (2) NLopt Xerces-C++ (1) libreadline (32) BCFtools Tk (1) GEOS (2) (1)hwloc OpenEXR (1) Amber aws-ofi-nccl (1) libtirpc (1) (1)GEOS BEDTools (1) GMP (2) OpenFOAMNextflow (2) beagle-lib Trimmomatic (1) libtool intltool (1) Apptainer BLAST+ (1) GLM (1) GROMACS (2) OpenJPEG (12) binutilsNinja (1) libunwind (1) BLASTDB UDUNITS2 (1) GSL jbigkit (12) Armadillo OpenSSL (2) GLib (1) bzip2 (32) OSPRay libxml2 (2) BWA VASP (13) GaussianlibGLU (1) PCRE (2) AutoDock-vina (1) GMP (1) cURL (13) lz4 OpenCASCADE (32) BamTools VCFtools (1) Go libaec (12) Autoconf PCRE2 (1) GObject-Introspection (12) canuOpenEXR (12) WPS minimap2 (1) Beast (2) libdeflate (12) Automake HDF-EOS (1) GROMACS PROJ (12) OpenFOAM cpeCray (12) ncclWRF (1) Bison libdrm (12) Autotools HDF (1) GSL (1) ParallelIO (1) cpeGNU (13) OpenJPEG ncurses (12) Boost WRFchem (42) libepoxy HTSlib (12) BCFtools Perl (1) Gaussian (1) cpeIntel (1) OpenMPI numactl (1) Bowtie Wayland (12) libffi HYPRE (1) BEDTools PostgreSQL (1) GenericIO ecCodes (1) pkgconf (2) OpenSSL (1) Bowtie2 (1) ICUX11 (32) libgeotiff QuantumESPRESSO (2) expatBLAST+ (21) Gmsh tbb (1) C-Blosc2 (1) ImathOpenTURNS (12) RAxML-NGXZ (1) flex(3) libglvnd (12) BLASTDB xorg-macros (1) CFITSIO Go (1) JasPer (2) SAMtools (1) PCRE (1) gettext (1) Xerces-C++ zfp (1) libiconv (2) BWA CGAL (1) JavaHDF-EOS (2) SCOTCH (2) PCRE2 (1) git-lfs (1) zlib Yasm (21) CMake libjpeg-turbo (3) BamTools (21) KaHIPHDF (1) SPAdes (2) PDAL (1) arpack-ng gperftools (12) libpciaccess zstd(2) Beast (31) HTSlib CrayNVHPC (1) LMDB (1) PETSc SQLite (12) groffassimp (12) DB libpng (3) Bison (1) LibTIFF (1) HYPRE SWIG (2) hwlocPROJ (12) ESMF at-spi2-atk (2) libreadline (12) Blosc M4 (1) Tcl (1) jbigkit (1) |
Expand | |||||
---|---|---|---|---|---|
| |||||
|
2. Building an application software
After an appropriate environment is loaded, this section provides guidelines on how to use it to build an application software on LANTA.
2.1 Compiler wrapper
...
<wrapper> command
...
Description
...
Manual
...
In substitution for
...
cc
...
C compiler wrapper
...
man cc
or cc --help
...
mpicc / mpiicc
...
CC
...
C++ compiler wrapper
...
man CC
or CC --help
...
mpic++ / mpiicpc
...
ftn
...
Fortran compiler wrapper
...
man ftn
or ftn --help
...
mpif90 / mpiifort
The Cray compiler wrappers, namely, cc
, CC
and ftn
, become available after loading any PrgEnv-<compiler>
or CPE toolchain. Upon being invoked, the wrapper will pass relevant information about the cray-*
libraries, loaded in the current environment, to the underlying <compiler>
to compile source code. It is recommended to use these wrappers for building MPI applications with the native Cray MPICH library cray-mpich
.
Adding -craype-verbose
to the wrapper when compiling a source file will display the final command executed. To see what will be added before compiling, try <wrapper> --cray-print-opts=all
.
Expand | |||||
---|---|---|---|---|---|
| |||||
The output of the final command indicates that include search paths ( ${CPATH} , ${LIBRARY_PATH} and ${LD_LIBRARY_PATH} , and users have to manually add appropriate linking flags (-l<lib> ). However, these steps should be covered automatically if using a build tool. (2) HarfBuzz (2) Pango (2) at-spi2-core (2) libtirpc (2)
Boost (4) ICU (3) ParFlow (1) aws-ofi-nccl (1) libtool (1)
Bowtie (1) Imath (2) ParMETIS (2) beagle-lib (1) libunwind (2)
Bowtie2 (1) JasPer (3) ParaView (1) binutils (1) libxml2 (3)
Brotli (2) Java (2) ParallelIO (1) bzip2 (3) lz4 (3)
C-Blosc2 (2) KaHIP (2) Perl (2) cURL (1) minimap2 (1)
CFITSIO (2) LAME (1) PostgreSQL (2) cairo (2) nccl (1)
CGAL (2) LLVM (1) QuantumESPRESSO (2) canu (1) ncurses (2)
CMake (2) LMDB (1) RAxML-NG (1) cpeCray (1) nlohmann_json (1)
CrayNVHPC (1) LibTIFF (2) SAMtools (1) cpeGNU (1) numactl (1)
DB (2) M4 (1) SCOTCH (2) cpeIntel (1) pixman (1)
DBus (1) MAFFT (1) SDL2 (2) ecCodes (2) pkgconf (1)
ESMF (2) METIS (2) SLEPc (2) expat (2) tbb (1)
EasyBuild (1) MPC (2) SPAdes (1) flex (1) termcap (1)
Eigen (1) MPFR (2) SQLite (2) fontconfig (2) x264 (1)
FDS (1) MUMPS (3) SWIG (3) freetype (2) x265 (1)
FFmpeg (2) Mako (2) SYCL (1) gettext (1) xorg-macros (2)
FastQC (1) Mamba (1) SZ (2) git-lfs (1) xprop (2)
FortranGIS (2) Mesa (2) SpectrA (1) googletest (1) zfp (2)
FreeXL (2) Meson (2) SuiteSparse (2) gperf (1) zlib (2)
FriBidi (2) MrBayes (1) SuperLU_DIST (2) gperftools (2) zstd (3) |
Expand | |||||
---|---|---|---|---|---|
| |||||
|
...
2. Building an application software
After an appropriate environment is loaded, this section provides guidelines on how to use it to build an application software on LANTA.
2.1 Compiler wrapper
<wrapper> command | Description | Manual | In substitution for |
| C compiler wrapper |
| mpicc / mpiicc |
| C++ compiler wrapper |
| mpic++ / mpiicpc |
| Fortran compiler wrapper |
| mpif90 / mpiifort |
The Cray compiler wrappers, namely, cc
, CC
and ftn
, become available after loading any PrgEnv-<compiler>
or CPE toolchain. Upon being invoked, the wrapper will pass relevant information about the cray-*
libraries, loaded in the current environment, to the underlying <compiler>
to compile source code. It is recommended to use these wrappers for building MPI applications with the native Cray MPICH library cray-mpich
.
Adding -craype-verbose
to the wrapper when compiling a source file will display the final command executed. To see what will be added before compiling, try <wrapper> --cray-print-opts=all
.
Expand | |||||
---|---|---|---|---|---|
| |||||
The output of the final command indicates that include search paths ( For ThaiSC libraries, their include and library search paths are stored in |
Expand | |||||
---|---|---|---|---|---|
| |||||
|
Expand | |||||
---|---|---|---|---|---|
| |||||
|
Expand | |||||
---|---|---|---|---|---|
| |||||
|
Expand | |||||
---|---|---|---|---|---|
| |||||
|
2.2 Build tools
Several tools exist to help us build large and complex programs. Among them, GNU make and CMake are commonly used. The developer team for each software chooses what build tool they will support. Therefore, it is important to thoroughly read the software documentation. For some software, users might need to additionally load the latest CMake or Autotools modules on the system (e.g., module load CMake/3.26.4
).
There are three general stages in building a program using a build tool: configure
, make
and make install
. For more information, see Basic Installation.
Build tools typically detect compilers through environment variables such as CC
, CXX
, and FC
at the configure
stage. Therefore, setting these variables before running configure
should be sufficient to make the tool use the Cray compiler wrappers.
Code Block | ||
---|---|---|
| ||
export CC=cc CXX=CC FC=ftn F77=ftn F90=ftn
# ./configure --prefix=<your-install-location> ...
# or
# cmake -DCMAKE_INSTALL_PREFIX=<your-install-location> ... |
Nevertheless, if the CMake cache is not clean, you might need to explicitly use:
Code Block | ||
---|---|---|
| ||
cmake -DCMAKE_C_COMPILER=cc -DCMAKE_CXX_COMPILER=CC -DCMAKE_Fortran_COMPILER=ftn -DCMAKE_INSTALL_PREFIX=<your-install-location> ... |
We encourage users to manually specify the installation path using --prefix=
or -DCMAKE_INSTALL_PREFIX=
as shown above. This path can be within your project home, such as /project/ltXXXXXX-YYYY/<software-name>
, allowing you to manage permissions and share the installed software with your project members. By default on LANTA, your team will be able to read and execute your software but cannot make any changes inside the directory you own.
After these steps, you should be able to execute make
and make install
, then build your software as you would on any other system.
Expand | |||||
---|---|---|---|---|---|
| |||||
|
2.3 Related topics
Local module & EasyBuild
A separate page is dedicated for explaining how to manage and install local modules in the user’s home/project paths using EasyBuild → Local module & EasyBuild (In progress)
Useful compiler flags
Intel oneAPI (In progress)
3. Running the software
Every main application software must run on compute/gpu/memory nodes. The recommended approach is to write a job script and send it to Slurm scheduler through sbatch
command.
Note |
---|
Only use |
3.1 Writing a job script
Code Block | ||
---|---|---|
| ||
#!/bin/bash
#SBATCH -p gpu # Partition
#SBATCH -N 1 # Number of nodes
#SBATCH --gpus=4 # Number of GPU cards
#SBATCH --ntasks=4 # Number of MPI processes
#SBATCH --cpus-per-task=16 # Number of OpenMP threads per MPI process
#SBATCH -t 5-00:00:00 # Job runtime limit
#SBATCH -A ltXXXXXX # Billing account
# #SBATCH -J <JobName> # Job name
module purge
# --- Load necessary modules ---
module load <...>
module load <...>
# --- Add software to Linux search paths ---
export PATH=<software-bin-path>
|
2.2 Build tools
Several tools exist to help us build large and complex programs. Among them, GNU make and CMake are commonly used. The developer team for each software chooses what build tool they will support. Therefore, it is important to thoroughly read the software documentation. For some software, users might need to additionally load the latest CMake or Autotools modules on the system (e.g., module load CMake/3.26.4
).
There are three general stages in building a program using a build tool: configure
, make
and make install
. For more information, see Basic Installation.
Build tools typically detect compilers through environment variables such as CC
, CXX
, and FC
at the configure
stage. Therefore, setting these variables before running configure
should be sufficient to make the tool use the Cray compiler wrappers.
Code Block | ||
---|---|---|
| ||
export CC=cc CXX=CC FC=ftn F77=ftn F90=ftn
# ./configure --prefix=<your-install-location> ...
# or
# cmake -DCMAKE_INSTALL_PREFIX=<your-install-location> ... |
Nevertheless, if the CMake cache is not clean, you might need to explicitly use:
Code Block | ||
---|---|---|
| ||
cmake -DCMAKE_C_COMPILER=cc -DCMAKE_CXX_COMPILER=CC -DCMAKE_Fortran_COMPILER=ftn -DCMAKE_INSTALL_PREFIX=<your-install-location> ... |
We encourage users to manually specify the installation path using --prefix=
or -DCMAKE_INSTALL_PREFIX=
as shown above. This path can be within your project home, such as /project/ltXXXXXX-YYYY/<software-name>
, allowing you to manage permissions and share the installed software with your project members. By default on LANTA, your team will be able to read and execute your software but cannot make any changes inside the directory you own.
After these steps, you should be able to execute make
and make install
, then build your software as you would on any other system.
Expand | |||||
---|---|---|---|---|---|
| |||||
|
2.3 Related topics
...
3. Running the software
Every main application software must run on compute/gpu/memory nodes. The recommended approach is to write a job script and send it to Slurm scheduler through sbatch
command.
Note |
---|
Only use |
3.1 Writing a job script
Code Block | ||
---|---|---|
| ||
#!/bin/bash
#SBATCH -p gpu # Partition
#SBATCH -N 1 # Number of nodes
#SBATCH --gpus=4 # Number of GPU cards
#SBATCH --ntasks=4 # Number of MPI processes
#SBATCH --cpus-per-task=16 # Number of OpenMP threads per MPI process
#SBATCH -t 5-00:00:00 # Job runtime limit
#SBATCH -A ltXXXXXX # Billing account
# #SBATCH -J <JobName> # Job name
module purge
# --- Load necessary modules ---
module load <...>
module load <...>
# --- Add software to Linux search paths ---
export PATH=<software-bin-path>:${PATH}
export LD_LIBRARY_PATH=<software-lib/lib64-path>:${LD_LIBRARY_PATH}
# export PYTHONPATH=<software-python-site-packages>:${PYTHONPATH}
# source <your-software-specific-script>
# --- (Optional) Set related environment variables ---
# export OMP_NUM_THREADS=${SLURM_CPUS_PER_TASK} # MUST specify --cpus-per-task above
# --- Run the software ---
# srun <srun-options> ./<software>
# or
# ./<software> |
The above job script template consists of five sections:
1. Slurm sbatch header
Anchor | ||||||
---|---|---|---|---|---|---|
|
The
#SBATCH
directives can be used to specify sbatch
options that mostly unchanged, such as partition, time limit, billing account, and so on. For optional options like job name, users can specify them when submitting the script (see Submitting a job). For more details regarding sbatch
options, please visit Slurm sbatch. Mostly, Slurm sbatch
options only define and request computing resources that can be used inside a job script. The actual resources used by a software/executable can be different depending on how it will be invoked/issued (see Stage 5), although these sbatch
options are passed and become the default options for it. For GPU jobs, using either --gpus
or --gpus-per-node
to request GPUs at this stage will provide the most flexibility for the next stage, GPU binding.
If your application software only supports parallelization by multi-threading, then your software cannot utilize resources across nodes; in this case, therefore, -N
, -n, --ntasks
and --ntasks-per-node
should be set to 1.
2. Loading modules
It is advised to load every module used when installing the software in the job script, although build dependencies such as CMake, Autotools, and binutils can be omitted. Additionally, those modules should be of the same version as when they were used to compile the program.
3. Adding software paths
The Linux OS will not be able to find your program if it is not in its search paths. The commonly used ones are namely PATH
(for executable/binary), LD_LIBRARY_PATH
(for shared library), and PYTHONPATH
(for python packages). Users MUST append or prepend them using syntax such as export PATH=<software-bin-path>:${PATH}
, otherwise, prior search paths added by module load
and others will disappear.
If <your-install-location>
is where your software is installed, then putting the below commands in your job script should be sufficient in most cases.
Code Block | ||
---|---|---|
| ||
export PATH=<your-install-location>/bin:${PATH} export LD_LIBRARY_PATH=<software<your-install-location>/lib/lib64-path>:${LD_LIBRARY_PATH} # export PYTHONPATH=<software-python-site-packages>LD_LIBRARY_PATH=<your-install-location>/lib64:${PYTHONPATH} # source <your-software-specific-script> # --- (Optional) Set related environment variables --- # export OMP_NUM_THREADS=${SLURM_CPUS_PER_TASK} # MUST specify --cpus-per-task above # --- Run the software --- # srun <srun-options> ./<software> # or # ./<software> |
The above job script template consists of five sections:
...
It should be noted that Slurm sbatch options only define and request computing resources that can be used inside a job script. The actual resources used by a software/executable can be different depending on how it will be invoked/issued (see Stage 5).
2. Loading modules
It is advised to load every module used when installing the software in the job script, although build dependencies such as CMake, Autotools, and binutils can be omitted. Additionally, those modules should be of the same version as when they were used to compile the program.
...
LD_LIBRARY_PATH} |
Some of them can be omitted if there no such sub-directory when using ls <your-install-location>
.
Expand | |||||
---|---|---|---|---|---|
| |||||
If the software also generates
|
Expand | |||||
---|---|---|---|---|---|
| |||||
When executing your program, if you encounter
Preliminary check could be performed on frontend node by doing something like
|
4. Setting environment variables
Some software requires additional environment variables to be set at runtime; for example, the path to the temporary directory. Output environment variables set by Slurm sbatch (see Slurm sbatch - output environment variables) could be used to set software-specific parameters.
For application with OpenMP threading, OMP_NUM_THREADS
, OMP_STACKSIZE
, ulimit -s unlimited
are commonly set in a job script. An example is shown below.
Code Block | ||
---|---|---|
| ||
export PATH=<your-install-location>/bin:${PATHXXX_TMPDIR=/scratch/ltXXXXXX-YYYY/${SLURM_JOBID} export LDOMP_LIBRARY_PATH=<your-install-location>/lib:${LD_LIBRARY_PATH}STACKSIZE="32M" export LD_LIBRARY_PATH=<your-install-location>/lib64:${LD_LIBRARY_PATH} |
Some of them can be omitted if there no such sub-directory when using ls <your-install-location>
.
Expand | |||||
---|---|---|---|---|---|
| |||||
If the software also generates
|
...
title | More information |
---|
If some software dependencies were installed locally, their search paths should also be added.
We do NOT recommend specifying these search paths in
~/.bashrc
directly, as it could lead to library conflicts when having more than one main software.Some software provides a script to be sourced before using. In this case, sourcing it in your job script should be equivalent to adding its search paths manually by yourself.
When executing your program, if you encounter
If 'xxx' is not a typo you can use command-not-found to lookup ...
, then, your currentPATH
variable may be incorrect.xxx: error while loading shared libraries: libXXX.so: cannot open shared object file
, then,If
libXXX.so
seem to be related to your software, then you may setLD_LIBRARY_PATH
variable in Step 3 incorrectly.If
libXXX.so
seem to be from a module you used to build your software, then loading that module should fix the problem.
ModuleNotFoundError: No module named 'xxx'
, then, your currentPYTHONPATH
may be incorrect.
Preliminary check could be performed on frontend node by doing something like
Code Block | ||
---|---|---|
| ||
bash # You should check them in another bash shell
module purge
module load <...>
module load <...>
export PATH=<software-bin-path>:${PATH}
export LD_LIBRARY_PATH=<software-lib/lib64-path>:${LD_LIBRARY_PATH}
export PYTHONPATH=<software-python-site-packages>:${PYTHONPATH}
<executable> --help
<executable> --version
exit |
4. Setting environment variables
Some software requires additional environment variables to be set at runtime; for example, the path to the temporary directory. Parameters set by Slurm sbatch (see Slurm sbatch - output environment variables) could be utilized in setting up software-specific environment variables.
For application with OpenMP threading, OMP_NUM_THREADS
, OMP_STACKSIZE
, ulimit -s unlimited
are commonly set in a job script. An example is shown below.
Code Block | ||
---|---|---|
| ||
export XXX_TMPDIR=/scratch/ltXXXXXX-YYYY/${SLURM_JOBID}
export OMP_STACKSIZE="32M"
export OMP_NUM_THREADS=${SLURM_CPUS_PER_TASK}
ulimit -s unlimited |
...
Usually, either srun
, mpirun
, mpiexec
or aprun
is required to run MPI programs. On LANTA, srun
command MUST be used instead. The table below compares a few options of those commands.
...
Command
...
Total MPI processes
...
CPU per MPI process
...
MPI processes per node
...
srun
...
-n, --ntasks
...
-c, --cpus-per-task
...
--ntasks-per-node
...
mpirun/mpiexec
...
-n, -np
...
--map-by socket:PE=N
...
--map-by ppr:N:node
...
aprun
...
-n, --pes
...
-d, --cpus-per-pe
...
-N, --pes-per-node
There is usually no need to add options to srun
since, by default, Slurm will automatically derive them from sbatch
. However, we recommend explicitly adding GPU binding options such as --gpus-per-task
or --ntasks-per-gpu
according to your software specification to srun
. Please visit Slurm srun for more details.
Note |
---|
For multi-threaded OMP_NUM_THREADS=${SLURM_CPUS_PER_TASK}
ulimit -s unlimited |
5. Running your software
Anchor | ||||||
---|---|---|---|---|---|---|
|
Each software has its own command to be issued. Please read the software documentation and forum. Special attention should be paid to how the software recognizes and maps computing resources (CPU-MPI-GPU); occasionally, users may need to insert additional input arguments at runtime. The total resources concurrently utilized in this stage should be less than or equal to the resources previously requested in Stage 1. Oversubscribing resources can reduce overall performance and could cause permanent damage to the hardware.
Usually, either srun
, mpirun
, mpiexec
or aprun
is required to run MPI programs. On LANTA, srun
command MUST be used to launch MPI processes. The table below compares a few options of those commands.
Command | Total MPI processes | CPU per MPI process | MPI processes per node |
---|---|---|---|
srun | -n, --ntasks | -c, --cpus-per-task | --ntasks-per-node |
mpirun/mpiexec | -n, -np | --map-by socket:PE=N | --map-by ppr:N:node |
aprun | -n, --pes | -d, --cpus-per-pe | -N, --pes-per-node |
There is usually no need to explicitly add option to srun
since, by default, Slurm will automatically derive them from sbatch
, with the exception of --cpus-per-task.
Anchor | ||||
---|---|---|---|---|
|
Expand | ||||||||||
---|---|---|---|---|---|---|---|---|---|---|
| ||||||||||
|
Note |
---|
For hybrid (MPI+Multi-threading) applications, it is essential to specify |
...
Info |
---|
You can test your initial script on compute-devel or gpu-devel partitions, using |
Your entire job script will only run on the first requested node (${SLURMD_NODENAME}). Only the lines starting with srun
could initiate process and run on the other nodes.
...
Example
MiniWeather (cpeCray + cudatoolkit)
Installation guide
...