Skip to main content Link Search Menu Expand Document (external link)

Installation on LUMI (lumi.csc.fi) Supercomputer

Here are a couple of recipes for commonly used computing facilities, which can be easily adapted.

To compile the code, first load the required modules:

module swap PrgEnv-cray PrgEnv-gnu
module load LUMI/23.03
module load cray-hdf5-parallel
module load buildtools/23.03

then set-up the build:

export TREXIO_DIR=/users/shindera
export QMCKL_DIR=/users/shindera

cmake -S. -Bbuild -DCMAKE_Fortran_COMPILER=ftn -DCMAKE_C_COMPILER=cc -DENABLE_TREXIO=ON -DENABLE_QMCKL=ON

and finally build:

cmake --build build -j --clean-first

To run the code, you need to submit a job to the queue system:

sbatch job.sh

where job.sh is a SLURM job script. Here are some sample scripts:

Sample VMC job script

#!/bin/bash -l
#SBATCH --job-name=ex01         # Job name
#SBATCH --output=ex01.o%j       # Name of stdout output file
#SBATCH --error=ex01.e%j        # Name of stderr error file
#SBATCH --partition=standard    # Partition (queue) name
#SBATCH --nodes=1               # Total number of nodes
#SBATCH --ntasks=128            # Total number of mpi tasks
#SBATCH --mem=0                 # Allocate all the memory on the node
#SBATCH --time=0:10:00          # Run time (d-hh:mm:ss)
#SBATCH --mail-type=all         # Send email at begin and end of job
#SBATCH --account=              # Project for billing
#SBATCH --reservation=          # Reservation name

# Any other commands must follow the #SBATCH directives
export PMI_NO_PREINITIALIZE=y

# Launch MPI code
srun champ/bin/vmc.mov1 -i vmc_h2o_hf_trexio.inp -o vmc_h2o_hf_trexio.out  -e error

Sample DMC job script

#!/bin/bash -l
#SBATCH --job-name=ex05         # Job name
#SBATCH --output=ex05.o%j       # Name of stdout output file
#SBATCH --error=ex05.e%j        # Name of stderr error file
#SBATCH --partition=standard    # Partition (queue) name
#SBATCH --nodes=1               # Total number of nodes
#SBATCH --ntasks=128            # Total number of mpi tasks
#SBATCH --mem=0                 # Allocate all the memory on the node
#SBATCH --time=0:05:00          # Run time (d-hh:mm:ss)
#SBATCH --mail-type=all         # Send email at begin and end of job
#SBATCH --account=              # Project for billing
#SBATCH --reservation=          # Reservation for training

# Any other commands must follow the #SBATCH directives
export PMI_NO_PREINITIALIZE=y

# Launch MPI code
srun champ/bin/vmc.mov1 -i vmc_h2o_hf_jas2body_trexio.inp -o vmc_h2o_hf_jas2body_trexio.out -e error

cat mc_configs_new* >> mc_configs
rm mc_configs_new*

srun champ/bin/dmc.mov1 -i dmc_h2o_hf_jas2body_trexio.inp -o dmc_h2o_hf_jas2body_trexio.out -e error