====== How to use compute nodes ======
==== Compilers ====
^Compilers^Version^Licenses^
|ifort, icc, icpc|Intel oneApi 2021|2 floating|
|mpiifort, mpiicc, mpiicpc|:::|:::|
|ifort2020, icc2020, icpc2020|Intel Parellel Studio 2020|:::|
|mpiifort2020, mpiicc2020, mpiicpc2020|:::|:::|
|ifort2017, icc2017, icpc2017|Intel Parellel Studio 2017|5 floating|
|mpiifort2017, mpiicc2017, mpiicpc2017|:::|2 floating|
==== Compile Options ====
(a) optimal for Fortran
|$ ifort –ipo –ip –O3 –xCASCADELAKE prog.f90
|Serial|
|$ mpiifort –ipo –ip –O3 –xCASCADELAKE prog.f90
|MPI parallel|
(b) optimal for C
|$ icc –ipo –ip –O3 –xCASCADELAKE prog.c
|Serial|
|$ mpiicc –ipo –ip –O3 –xCASCADELAKE prog.c
|MPI parallel|
(c) optimal for C++
|$ icpc –ipo –ip –O3 –xCASCADELAKE prog.cpp
|Serial|
|$ mpiicpc –ipo –ip –O3 –xCASCADELAKE prog.cpp
|MPI parallel|
(d) other options
|-qopenmp|Use OpenMP|
|-parallel|Use automatic parallelization|
|-check|For debug|
|-traceback|For debug|
==== Job Queues ====
^Item \ Queue^IDLbatch^SHARED^NODEX1^NODEX4^
|Node usage|2 IDL nodes|2 Shared nodes|12 Occupied nodes|12 Occupied nodes|
|Number of nodes|1|1|1|2 - 4|
|Number of cores per node|1 - 2|1 - 26|52|52|
|Default elapsed time|24h|24h|24h|24h|
|Maximum elapsed time|120h |120h |168h |120h |
|Default memory size per node|4GB|4GB|350GB|350GB|
|Maximum memory size per node|87GB|175GB|350GB|350GB|
|Number of submissions per user|1|8|4|2|
|Number of executions per user |1|4|2|1|
|Number of executions per system|52|unlimited|unlimited|2|
NODEX1 and NODEX4 queues are for ISEE computational joint research program only.
==== Job queuing system ====
(a) Submit a job
$ qsub
(b) Check status of jobs
|$ qstat
|List all your jobs|
|$ qstat -J
|Show all nodes used by your jobs|
|$ qstat -S
|Show status of all nodes|
|$ qstat -Q
|Show the number of jobs in queues|
|$ qstat -f
|Show detailed information of a job|
(c) Delete a job
$ qdel
==== Sample Job Scripts ====
(a) IDLbatch queue
|
#!/bin/bash
#PBS -q IDLbatch
#PBS –b 1
#PBS -l elapstim_req=96:00:00
#PBS -l cpunum_lhost=1
#PBS -l memsz_lhost=8GB
#PBS -M user@isee.nagoya-u.ac.jp
#PBS -m be
ulimit -s unlimited
cd $PBS_O_WORKDIR
idl < prog.pro
|
# shell : do not change
# name of queue : do not change
# number of nodes : do not change
# elapsed time : default 24:00:00, max 120:00:00
# number of cores : 1 - 2
# memory size : default 4GB, max 87GB
# send email to user@isee.nagoya-u.ac.jp
# send email at (b) "beginning" and (e) "end"
# Do not change
# Do not change
#
|
(b) SHARED queue
|
#!/bin/bash
#PBS -q SHARED
#PBS –b 1
#PBS -l elapstim_req=96:00:00
#PBS -l cpunum_lhost=26
#PBS -l memsz_lhost=120GB
#PBS -v OMP_STACKSIZE=512m
#PBS -v OMP_NUM_THREADS=13
#PBS -M user@isee.nagoya-u.ac.jp
#PBS -m be
ulimit -s unlimited
cd $PBS_O_WORKDIR
mpirun -n 2 ./a.out < input.txt
|
# shell : do not change
# name of queue : do not change
# number of nodes : do not change
# elapsed time : default 24:00:00, max 120:00:00
# number of cores : 1 - 26
# memory size : default 4GB, max 175GB
# for OpenMP : stack size
# for OpenMP : number of threads
# send email to user@isee.nagoya-u.ac.jp
# send email at (b) "beginning" and (e) "end"
# Do not change
# Do not change
# use 2 processes (x 13 threads = 26 cores)
|
(c) NODEX1 queue
|
#!/bin/bash
#PBS -q NODEX1
#PBS –b 1
#PBS -l elapstim_req=96:00:00
#PBS -v OMP_STACKSIZE=512m
#PBS -v OMP_NUM_THREADS=13
#PBS -M user@isee.nagoya-u.ac.jp
#PBS -m be
ulimit -s unlimited
cd $PBS_O_WORKDIR
mpirun -n 4 ./a.out < input.txt
|
# shell : do not change
# name of queue : do not change
# number of nodes : do not change
# elapsed time : default 24:00:00, max 168:00:00
# for OpenMP : stack size
# for OpenMP : number of threads
# send email to user@isee.nagoya-u.ac.jp
# send email at (b) "beginning" and (e) "end"
# Do not change
# Do not change
# use 4 processes (x 13 threads = 52 cores)
|
(d) NODEX4 queue
|
#!/bin/bash
#PBS -q NODEX4
#PBS –b 4
#PBS -T intmpi
#PBS -l elapstim_req=96:00:00
#PBS -v OMP_STACKSIZE=512m
#PBS -v OMP_NUM_THREADS=26
#PBS -M user@isee.nagoya-u.ac.jp
#PBS -m be
ulimit -s unlimited
cd $PBS_O_WORKDIR
mpirun -n 8 -npp 2 ./a.out < input.txt
|
# shell : do not change
# name of queue : do not change
# number of nodes : 2 - 4
# Do not change
# elapsed time : default 24:00:00, max 120:00:00
# for OpenMP : stack size
# for OpenMP : number of threads
# send email to user@isee.nagoya-u.ac.jp
# send email at (b) "beginning" and (e) "end"
# Do not change
# Do not change
# use 8 processes and 2 processes per node (x 26 threads = 52 cores)
|
==== In Publications ====
When submitting a publication based on results from the CIDAS system,
users are asked to include the following line in the acknowledgements:
The computation was performed on the CIDAS computer system at the Institute for Space-Earth Environmental Research, Nagoya University.
----
* [[public:en:manual:cidas:start|CIDAS system manual (english)]]
* [[public:ja:manual:cidas:start|CIDAS system manual (japanese)]]