Skip to content

Commit

Permalink
LUMI wip
Browse files Browse the repository at this point in the history
  • Loading branch information
luraess committed Nov 24, 2023
1 parent 866c055 commit 580e439
Show file tree
Hide file tree
Showing 5 changed files with 27 additions and 30 deletions.
16 changes: 0 additions & 16 deletions scripts_future_API/profileme.sh

This file was deleted.

23 changes: 16 additions & 7 deletions scripts_future_API/runme.sh
Original file line number Diff line number Diff line change
@@ -1,14 +1,23 @@
#!/bin/bash

module load LUMI/22.08
module load partition/G
module load rocm/5.3.3
source /users/lurass/scratch/setenv_lumi.sh
# module load LUMI/22.08
# module load partition/G
# module load rocm/5.3.3

# ROCm-aware MPI set to 1, else 0
export MPICH_GPU_SUPPORT_ENABLED=1
export IGG_ROCMAWARE_MPI=1

# Needs to know about location of GTL lib
export LD_PRELOAD=${CRAY_MPICH_ROOTDIR}/gtl/lib/libmpi_gtl_hsa.so
## basic
# srun --cpu-bind=map_cpu:49,57,17,25,1,9,33,41 -N1 -n8 --gpus-per-node=8 profileme.sh

julia --project rocmaware.jl
## optimal using only single GCD per MI250x Module
# srun --cpu-bind=map_cpu:49,17,1,33 -N1 -n1 --gpus-per-node=8 profileme.sh
# srun --cpu-bind=map_cpu:49,17,1,33 -N4 -n16 --gpus-per-node=8 profileme.sh
export ROCR_VISIBLE_DEVICES=0,2,4,6

# julia --project benchmark_diffusion_3D.jl
julia --project --color=yes tm_stokes_mpi_wip.jl

# Profiling
# ENABLE_JITPROFILING=1 rocprof --hip-trace --hsa-trace -d ./prof_out${SLURM_PROCID} -o ./prof_out${SLURM_PROCID}/results${SLURM_PROCID}.csv julia --project bench3d.jl
15 changes: 9 additions & 6 deletions scripts_future_API/submit.sh
Original file line number Diff line number Diff line change
Expand Up @@ -2,14 +2,17 @@
#SBATCH --job-name="FastIce3D"
#SBATCH --output=FastIce3D.%j.o
#SBATCH --error=FastIce3D.%j.e
#SBATCH --time=00:05:00
#SBATCH --nodes=16
#SBATCH --ntasks=64
# #SBATCH --ntasks-per-node=8
#SBATCH --time=00:10:00
#SBATCH --nodes=4
#SBATCH --ntasks=16
# #SBATCH --ntasks-per-node=8 # this somehow fails...
#SBATCH --gpus-per-node=8
#SBATCH --partition=standard-g
#SBATCH --account project_465000557

# export ROCR_VISIBLE_DEVICES=0,2,4,6
# CPU_BIND="map_cpu:49,57,17,25,1,9,33,41"

srun --cpu-bind=map_cpu:49,17,1,33 ./profileme.sh
# export ROCR_VISIBLE_DEVICES=0,2,4,6 # -> done in runme.sh
CPU_BIND="map_cpu:49,17,1,33"

srun --cpu-bind=${CPU_BIND} ./runme.sh
2 changes: 1 addition & 1 deletion scripts_future_API/tm_stokes_mpi_wip.jl
Original file line number Diff line number Diff line change
Expand Up @@ -36,7 +36,7 @@ function main(; do_visu=false, do_save=false)
MPI.Init()

backend = ROCBackend()
dims = (2, 1, 1)
dims = (4, 2, 2)
# dims = (4, 2, 2)
# dims = (2, 1, 1)
topo = CartesianTopology(dims)
Expand Down
1 change: 1 addition & 0 deletions scripts_future_API/tm_stokes_wip.jl
Original file line number Diff line number Diff line change
Expand Up @@ -15,6 +15,7 @@ const SBC = BoundaryCondition{Slip}
using LinearAlgebra, Printf
using KernelAbstractions
# using CUDA
# using AMDGPU

using CairoMakie
# using GLMakie
Expand Down

0 comments on commit 580e439

Please sign in to comment.