Usage on HOKUSAI at RIKEN ACCC
GreatWave MPC (FX100)
GreatWave MPC (GW-MPC) has 32 CPU cores in each node. The followings are recommended installation scheme, example of the batch script file for hybrid MPI/OpenMP computation with 64 MPI processors and 4 OpenMP threads (256 CPU cores in total), and a command to submit a job. Please see also the user guide of the computer system.
Installation
$ cd /home/user/genesis
$ module load sparc
$ ./configure --host=k
$ make
$ make install
Do not forget “module load sparc”.
In K computer and other FX10/FX100 machines, prst_setup
(Parallel I/O tool) is not installed, because Fujitsu compilers are not suitable for compiling the program.
Batch script
#!/bin/bash
#PJM -L rscunit=gwmpc
#PJM -L rscgrp=batch
#PJM -L node=8
#PJM -L elapse=24:00:00
#PJM -o log -j
#PJM --mpi "proc=64"
bindir=/home/user/genesis/bin
export OMP_NUM_THREADS=4
mpirun ${bindir}/spdyn INP
In this case, node=8
is specified, because each node has 32 CPU cores (64*4/32=8).
Usage
# Execute run1.sh to run3.sh sequentially $ pjsub --step run1.sh $ pjsub --step --sparam "jid=XXXXXXX" run2.sh $ pjsub --step --sparam "jid=XXXXXXX" run3.sh # Check running jobs $ pjsub -E # Delete a job $ pjdel JOB_ID # Check CPU time $ listcpu
The job ID (jid=XXXXXXX
) given at run1 is specified at run2 and run3.
BigWaterfall
BigWaterfall MPC (BW-MPC) has 40 CPU cores in each node. The followings are recommended installation scheme, example of the batch script file for hybrid MPI/OpenMP computation with 64 MPI processors and 5 OpenMP threads (320 CPU cores in total), and a command to submit a job. Please see also the user guide of the computer system.
Installation
$ cd /home/user/genesis
$ module load intel
$ ./configure FC=mpiifort CC=mpiicc F77=mpiifort
$ make
$ make install
Do not forget “module load intel”.
Batch script
#!/bin/sh
#PJM -L rscunit=bwmpc
#PJM -L rscgrp=batch
#PJM -L vnode=8
#PJM -L vnode-core=40
#PJM -L elapse=24:00:00
export OMP_NUM_THREADS=5
bindir=/home/user/genesis/bin
mpirun -np 64 -ppn 8 ${bindir}/spdyn INP > log
Usage
# Execute run1.sh to run3.sh sequentially $ pjsub --step run1.sh run2.sh run3.sh # Check running jobs $ pjstat # Delete a job $ pjdel JOB_ID # Check CPU time $ listcpu
GPU server (GWACSG)
GPU server (GWACSG) has 24 CPU cores and 4 GPU cards in each node. The followings are recommended installation scheme, example of the batch script file for hybrid MPI/OpenMP computation with 32 MPI processors and 3 OpenMP threads + 16 GPU cards, and a command to submit a job. Please see also the user guide of the computer system.
Installation
$ pjsub --interact -L rscunit=gwacsg -L vnode=1 -x gpu_per_vnode=1
$ cd /home/user/genesis
$ module load cuda/8.0
$ autoreconf
$ ./configure --enable-gpu --enable-single --with-cuda=/bwfefs/opt/x86_64/cuda/8.0/ FC=mpiifort CC=mpiicc
$ make
$ make install
$ exit
Compilation is carried out on a computational node.
Batch script
#!/bin/sh #PJM -L rscunit=gwacsg #PJM -L rscgrp=gpu #PJM -L vnode=4 #PJM -L vnode-core=24 #PJM -x gpu_per_vnode=4 #PJM -L elapse=24:00:00 #PJM -j export OMP_NUM_THREADS=3 module load cuda/8.0 mpirun -np 32 -ppn 8 gpurun /home/user/genesis/bin/spdyn INP > log