Newer
Older
##############################################
# ARIS slurm script template #
# #
# Submit script: sbatch GPU.slurm n1 n2 ... #
# #
##############################################
#SBATCH --output=GPU.out
#SBATCH --error=GPU.err
#SBATCH --ntasks=1 # Number of processor cores (i.e. tasks)
#SBATCH --nodes=1 # Number of nodes requested
#SBATCH --ntasks-per-node=1 # Tasks per node
#SBATCH --gres=gpu:1 # GPUs per node
#SBATCH --time=00:40:00 # walltime
#SBATCH --mem=32G # memory per NODE
#SBATCH --partition=gpu # Partition
#SBATCH --account=testproj # Accounting project
## LOAD MODULES ##
module purge # clean up loaded modules
# load necessary modules
module load gnu
module load intel
module load intelmpi
module load binutils
module load cuda
export I_MPI_FABRICS=shm:dapl
gpu_prog="./cuda_SingleGPU.exe"
gpu_prog1="./cuBLAS.exe"
gpu_prog2="./cuBLAS_MultiGPU.exe"
srun $gpu_prog $n $n
srun $gpu_prog1 $n $n
# Important note: In MultiGPU version you must use gres=ntasks-per-node values in order to utilize all GPUs !!!
# srun $gpu_prog2 $n $n