diff --git a/wk/sub_heat3d_A100.sh b/wk/sub_heat3d_A100.sh new file mode 100644 index 0000000..98da6bd --- /dev/null +++ b/wk/sub_heat3d_A100.sh @@ -0,0 +1,45 @@ +#!/bin/bash +#PJM -L "node=1" +#PJM -L "rscgrp=regular-a" +#PJM -L "elapse=10:00" +#PJM -s +#PJM -g jh220031a +#PJM --mpi proc=1 + +. /etc/profile.d/modules.sh # Initialize module command + +module purge + +# Load spack +export HOME=/work/jh220031a/i18048 +. $HOME/spack/share/spack/setup-env.sh + +spack load gcc@11.3.0 +spack load cmake@3.24.3%gcc@8.3.1 +module load /work/04/jh220031a/i18048/lib/nvidia/hpc_sdk23.3/modulefiles/nvhpc/23.3 +module list + +# Need GPUs to build the code appropriately +# So compile inside a batch job, wherein GPUs are visible +if [ ! -d "../build" ] +then + cd ../ + rm -rf build + mkdir build && cd build + cmake -DCMAKE_CXX_COMPILER=nvc++ -DBACKEND=CUDA .. + cmake --build . -j 8 + cd ../wk/ +fi + +export UCX_MEMTYPE_CACHE=n +export UCX_IB_GPU_DIRECT_RDMA=no +export UCX_RNDV_FRAG_MEM_TYPE=cuda + +echo "stdpar" +../build/mini-apps/heat3d/stdpar/heat3d-stdpar --nx 512 --ny 512 --nz 512 --nbiter 1000 --freq_diag 0 + +echo "thrust" +../build/mini-apps/heat3d/thrust/heat3d-thrust --nx 512 --ny 512 --nz 512 --nbiter 1000 --freq_diag 0 + +echo "executors" +../build/mini-apps/heat3d/executors/heat3d-executors --nx 512 --ny 512 --nz 512 --nbiter 1000 --freq_diag 0 diff --git a/wk/sub_heat3d_Icelake.sh b/wk/sub_heat3d_Icelake.sh new file mode 100644 index 0000000..386860e --- /dev/null +++ b/wk/sub_heat3d_Icelake.sh @@ -0,0 +1,51 @@ +#!/bin/bash +#PJM -L "node=1" +#PJM -L "rscgrp=regular-a" +#PJM -L "elapse=20:00" +#PJM -s +#PJM -g jh220031a +#PJM --mpi proc=1 + +. /etc/profile.d/modules.sh # Initialize module command + +module purge + +# Load spack +export HOME=/work/jh220031a/i18048 +. $HOME/spack/share/spack/setup-env.sh + +spack load gcc@11.3.0 +spack load cmake@3.24.3%gcc@8.3.1 +module load /work/04/jh220031a/i18048/lib/nvidia/hpc_sdk23.3/modulefiles/nvhpc/23.3 +module list + +# Need GPUs to build the code appropriately +# So compile inside a batch job, wherein GPUs are visible +if [ ! -d "../build" ] +then + cd ../ + rm -rf build + mkdir build && cd build + cmake -DCMAKE_CXX_COMPILER=nvc++ -DBACKEND=OPENMP .. + cmake --build . -j 8 + cd ../wk/ +fi + +export UCX_MEMTYPE_CACHE=n +export UCX_IB_GPU_DIRECT_RDMA=no +export OMP_NUM_THREADS=36 +export OMP_PROC_BIND=true + +numactl -H + +echo "stdpar" +numactl --physcpubind=+0-35 ../build/mini-apps/heat3d/stdpar/heat3d-stdpar --nx 512 --ny 512 --nz 512 --nbiter 1000 --freq_diag 0 +#../build/mini-apps/heat3d/stdpar/heat3d-stdpar --nx 512 --ny 512 --nz 512 --nbiter 1000 --freq_diag 0 + +echo "thrust" +numactl --physcpubind=+0-35 ../build/mini-apps/heat3d/thrust/heat3d-thrust --nx 512 --ny 512 --nz 512 --nbiter 1000 --freq_diag 0 +#../build/mini-apps/heat3d/thrust/heat3d-thrust --nx 512 --ny 512 --nz 512 --nbiter 1000 --freq_diag 0 + +echo "executors" +numactl --physcpubind=+0-35 ../build/mini-apps/heat3d/executors/heat3d-executors --nx 512 --ny 512 --nz 512 --nbiter 1000 --freq_diag 0 +#numactl --cpunodebind=0 --membind=0 ../build/mini-apps/heat3d/executors/heat3d-executors --nx 512 --ny 512 --nz 512 --nbiter 1000 --freq_diag 0