# Usage guide for pilot run for Forerunner1(F1)
[toc]
## Revised history
* 10/24 Update for the announcement of Grand Challenge winners.
* 5/31 Update opening schedule and Grand Challenge rules.
* 4/9 Open the following partitions: betatest, ct112, ct448, ct1k for beta test.
* 4/8 Add announcement section
* 4/3 Add SOP for project binding for disk quota.
* 3/27 Modify queuing policy and disk quota limitation.
## Disclaimer
**The main purpose of closed testing is to find bugs and usability issues, and to test the effectiveness of different management strategies. Your use of F1 resources constitutes your acceptance of the "Forerunner 1驗證測試辦法". NCHC reserves the right to dynamically adjust management strategies during the testing period.**
[Forerunner 1驗證測試辦法](https://hackmd.io/@acyang/SkRM82FyR)
## Announcement
* F1停機通知 4/11(四) 09:30~13:30 廠商安排停機四小時, 調整/pkg儲存空間(要解決權限問題), 請用戶暫停送job
## Login information
| Name | IP | Note |
| ------------- | --------------- | ---------------- |
| x64 | 140.110.122.196 | ilgn01 |
| arm64 | 140.110.122.xxx | Text |
| xdata | 140.110.122.xxx | Text |
| interactivate | 140.110.122.xxx | Text |
X11 Forwarding was disable on SSH service on ilgn01.
```
PuTTY X11 proxy: unable to connect to forwarded X server: Network error: Connection refused
(base) 00:17:23 p00acy00@ilgn01:~$
```
## Hardware
<table style="undefined;table-layout: fixed; width: 831px">
<colgroup>
<col style="width: 84px">
<col style="width: 355px">
<col style="width: 392px">
</colgroup>
<thead>
<tr>
<th></th>
<th>Computing node</th>
<th>Visualization node</th>
</tr>
</thead>
<tbody>
<tr>
<td>Nodes</td>
<td>552</td>
<td>6</td>
</tr>
<tr>
<td>CPU</td>
<td colspan="2">Intel Xeon Platinum 8480+<br>2.0G/56C/350W *2</td>
</tr>
<tr>
<td>GPU</td>
<td>NA</td>
<td>NVIDIA A40 48GB PCIe*2</td>
</tr>
<tr>
<td>RAM</td>
<td colspan="2">512GB<br>32GB DDR5-4800 4800RDIMM *16</td>
</tr>
<tr>
<td>SSD(OS)</td>
<td>Samsung PM9A3 NVMe SSD 960GB *1</td>
<td>Samsung PM983 SATA3 SSD 240GB *2</td>
</tr>
<tr>
<td>SSD(DATA)</td>
<td colspan="2">Samsung PM9A3 NVMe SSD 1.92TB *1</td>
</tr>
<tr>
<td>IB</td>
<td colspan="2">NVIDIA ConnectX 7 NDR 200G InfiniBand PCIe *1<br></td>
</tr>
</tbody>
</table>
### Computing
#### x64
##### cpuinfo
```
[p00acy00@ilgn01 ~]$ cat /proc/cpuinfo
...
processor : 111
vendor_id : GenuineIntel
cpu family : 6
model : 143
model name : Intel(R) Xeon(R) Platinum 8480+
stepping : 8
microcode : 0x2b000461
cpu MHz : 2001.000
cache size : 107520 KB
physical id : 1
siblings : 56
core id : 55
cpu cores : 56
apicid : 238
initial apicid : 238
fpu : yes
fpu_exception : yes
cpuid level : 32
wp : yes
flags : fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush dts acpi mmx fxsr sse sse2 ss ht tm pbe syscall nx pdpe1gb rdtscp lm constant_tsc art arch_perfmon pebs bts rep_good nopl xtopology nonstop_tsc cpuid aperfmperf tsc_known_freq pni pclmulqdq dtes64 monitor ds_cpl vmx smx est tm2 ssse3 sdbg fma cx16 xtpr pdcm pcid dca sse4_1 sse4_2 x2apic movbe popcnt tsc_deadline_timer aes xsave avx f16c rdrand lahf_lm abm 3dnowprefetch cpuid_fault epb cat_l3 cat_l2 cdp_l3 invpcid_single intel_ppin cdp_l2 ssbd mba ibrs ibpb stibp ibrs_enhanced tpr_shadow vnmi flexpriority ept vpid ept_ad fsgsbase tsc_adjust bmi1 avx2 smep bmi2 erms invpcid cqm rdt_a avx512f avx512dq rdseed adx smap avx512ifma clflushopt clwb intel_pt avx512cd sha_ni avx512bw avx512vl xsaveopt xsavec xgetbv1 xsaves cqm_llc cqm_occup_llc cqm_mbm_total cqm_mbm_local split_lock_detect avx_vnni avx512_bf16 wbnoinvd dtherm ida arat pln pts avx512vbmi umip pku ospke waitpkg avx512_vbmi2 gfni vaes vpclmulqdq avx512_vnni avx512_bitalg tme avx512_vpopcntdq la57 rdpid bus_lock_detect cldemote movdiri movdir64b enqcmd fsrm md_clear serialize tsxldtrk pconfig arch_lbr amx_bf16 avx512_fp16 amx_tile amx_int8 flush_l1d arch_capabilities
bugs : spectre_v1 spectre_v2 spec_store_bypass swapgs eibrs_pbrsb
bogomips : 3981.31
clflush size : 64
cache_alignment : 64
address sizes : 46 bits physical, 57 bits virtual
power management:
```
[Intel® Xeon® Platinum 8480+ Processor](https://ark.intel.com/content/www/us/en/ark/products/231746/intel-xeon-platinum-8480-processor-105m-cache-2-00-ghz.html)
##### lstopo
##### numactl
### Storage
#### IBM Spectrum Scale (GPFS)
若有MPIIO需求,請把檔案寫在平行檔案系統,不要寫在local disk或是NAS。
1. /home/${USER}
2. /work1/${USER} (beta用戶尚未確認如何開放)
3. /project (尚未開放)
* alpha測試期間取消disk qupta,2024/4/1恢復預設容量上限100GB。
* 建構測試期間,請不要將重要資料保留在F1儲存系統,目前不敢保證SLA。
* 2024/4/1啟用disk qupta,所有用戶都需要在iservice綁定F1檔案系統關聯計畫,詳細作法請參考影片。

[在iservice上設定F1檔案系統綁訂計畫的設定方式(個別用戶)](https://youtu.be/aRVXZaQ_HIk)
* alphatest用戶若有調高disk qupta需求請與各領域負責人聯繫。
* betatest用戶若有調高disk qupta需求,請參考下列影片,由各自計劃管理人操作即可。
[在iservice上調整計畫成員F1的磁碟空間大小(管理者)](https://www.youtube.com/watch?v=lvyhmVdMvA0)
#### NAS shared via NFS
1. /pkg/x86
* 存放軟體開發工具,軟體安裝協力夥伴,請在beta開始協助安裝。
#### Local disk on computing node (/tmp 1.8TB)
```
17:20:23 p00acy00@ilgn01:~$ srun lsblk
NAME MAJ:MIN RM SIZE RO TYPE MOUNTPOINT
sr0 11:0 1 1024M 0 rom
nvme1n1 259:0 0 1.8T 0 disk /tmp
nvme0n1 259:1 0 894.3G 0 disk
├─nvme0n1p1 259:2 0 600M 0 part /boot/efi
├─nvme0n1p2 259:3 0 1G 0 part /boot
└─nvme0n1p3 259:4 0 892.7G 0 part
├─rhel-root 253:0 0 263.1G 0 lvm /
├─rhel-swap 253:1 0 15.6G 0 lvm [SWAP]
├─rhel-home 253:2 0 175.4G 0 lvm /home
└─rhel-var 253:3 0 438.5G 0 lvm /var
```
PS. There is no local scratch on login node.
PPS. 若有MPIIO需求,請把檔案寫在平行檔案系統,不要寫在local disk。
### Networking
NVIDIA® Mellanox® 200G InfiniBand

```
(base) 08:51:59 p00acy00@ilgn01:~$ ucx_info -d | grep Transport
# Transport: self
# Transport: tcp
# Transport: tcp
# Transport: tcp
# Transport: tcp
# Transport: sysv
# Transport: posix
# Transport: dc_mlx5
# Transport: rc_verbs
# Transport: rc_mlx5
# Transport: ud_verbs
# Transport: ud_mlx5
# Transport: dc_mlx5
# Transport: rc_verbs
# Transport: rc_mlx5
# Transport: ud_verbs
# Transport: ud_mlx5
# Transport: cma
# Transport: knem
# Transport: xpmem
```
```
(base) 15:23:09 p00acy00@ilgn01:~$ lspci | grep Ethernet
03:00.0 Ethernet controller: Intel Corporation I210 Gigabit Network Connection (rev 03)
49:00.0 Ethernet controller: Mellanox Technologies MT2894 Family [ConnectX-6 Lx]
49:00.1 Ethernet controller: Mellanox Technologies MT2894 Family [ConnectX-6 Lx]
(base) 15:22:57 p00acy00@ilgn01:~$ lspci | grep Mellanox
49:00.0 Ethernet controller: Mellanox Technologies MT2894 Family [ConnectX-6 Lx]
49:00.1 Ethernet controller: Mellanox Technologies MT2894 Family [ConnectX-6 Lx]
5a:00.0 Infiniband controller: Mellanox Technologies MT2910 Family [ConnectX-7]
(base) 14:48:46 p00acy00@ilgn01:~$ ibv_devinfo
hca_id: mlx5_2
transport: InfiniBand (0)
fw_ver: 28.38.1900
node_guid: 946d:ae03:0097:0012
sys_image_guid: 946d:ae03:0097:0012
vendor_id: 0x02c9
vendor_part_id: 4129
hw_ver: 0x0
board_id: MT_0000000844
phys_port_cnt: 1
port: 1
state: PORT_ACTIVE (4)
max_mtu: 4096 (5)
active_mtu: 4096 (5)
sm_lid: 1
port_lid: 662
port_lmc: 0x00
link_layer: InfiniBand
hca_id: mlx5_bond_0
transport: InfiniBand (0)
fw_ver: 26.38.1900
node_guid: 946d:ae03:00ea:85b8
sys_image_guid: 946d:ae03:00ea:85b8
vendor_id: 0x02c9
vendor_part_id: 4127
hw_ver: 0x0
board_id: MT_0000000546
phys_port_cnt: 1
port: 1
state: PORT_ACTIVE (4)
max_mtu: 4096 (5)
active_mtu: 4096 (5)
sm_lid: 0
port_lid: 0
port_lmc: 0x00
link_layer: Ethernet
```
```
(base) 14:54:40 p00acy00@icpnq134:~$ ibv_devinfo
hca_id: mlx5_0
transport: InfiniBand (0)
fw_ver: 28.38.1900
node_guid: 946d:ae03:0097:0b02
sys_image_guid: 946d:ae03:0097:0b02
vendor_id: 0x02c9
vendor_part_id: 4129
hw_ver: 0x0
board_id: MT_0000000844
phys_port_cnt: 1
port: 1
state: PORT_ACTIVE (4)
max_mtu: 4096 (5)
active_mtu: 4096 (5)
sm_lid: 1
port_lid: 23
port_lmc: 0x00
link_layer: InfiniBand
```
## Software
### OS
```
17:55:31 p00acy00@ilgn01:~$ cat /proc/version
Linux version 4.18.0-425.3.1.el8.x86_64 (mockbuild@x86-vm-08.build.eng.bos.redhat.com) (gcc version 8.5.0 20210514 (Red Hat 8.5.0-15) (GCC)) #1 SMP Fri Sep 30 11:45:06 EDT 2022
```
RHEL 8.5
```
(base) 17:27:41 p00acy00@ilgn01:src$ ulimit -aS
core file size (blocks, -c) 0
data seg size (kbytes, -d) unlimited
scheduling priority (-e) 0
file size (blocks, -f) unlimited
pending signals (-i) 1029203
max locked memory (kbytes, -l) unlimited
max memory size (kbytes, -m) unlimited
open files (-n) 65535
pipe size (512 bytes, -p) 8
POSIX message queues (bytes, -q) 819200
real-time priority (-r) 0
stack size (kbytes, -s) unlimited
cpu time (seconds, -t) unlimited
max user processes (-u) 1029203
virtual memory (kbytes, -v) unlimited
file locks (-x) unlimited
```
### Toolchain
1. GCC 8.5.0(default)
2. Intel oneAPI
- Intel MPI Library Compiler Wrappers
<style type="text/css">
.tg {border-collapse:collapse;border-spacing:0;}
.tg td{border-color:black;border-style:solid;border-width:1px;font-family:Arial, sans-serif;font-size:14px;
overflow:hidden;padding:10px 5px;word-break:normal;}
.tg th{border-color:black;border-style:solid;border-width:1px;font-family:Arial, sans-serif;font-size:14px;
font-weight:normal;overflow:hidden;padding:10px 5px;word-break:normal;}
.tg .tg-s3xf{background-color:#F7F7F7;color:#262626;text-align:center;vertical-align:top}
.tg .tg-baqh{text-align:center;vertical-align:top}
.tg .tg-u97q{background-color:#F7F7F7;border-color:inherit;color:#262626;text-align:center;vertical-align:top}
.tg .tg-c3ow{border-color:inherit;text-align:center;vertical-align:top}
.tg .tg-26z5{background-color:#FFF;border-color:inherit;color:#262626;text-align:center;vertical-align:top}
.tg .tg-lc14{background-color:#FFF;color:#262626;text-align:center;vertical-align:top}
</style>
<table class="tg">
<thead>
<tr>
<th class="tg-c3ow">Compiler Command</th>
<th class="tg-c3ow">Default Compiler</th>
<th class="tg-c3ow">Supported Languages</th>
</tr>
</thead>
<tbody>
<tr>
<td class="tg-baqh" colspan="3">Generic Compilers</td>
</tr>
<tr>
<td class="tg-baqh">mpicc</td>
<td class="tg-baqh">gcc, cc</td>
<td class="tg-baqh">C</td>
</tr>
<tr>
<td class="tg-baqh">mpicxx</td>
<td class="tg-baqh">g++</td>
<td class="tg-baqh">C/C++</td>
</tr>
<tr>
<td class="tg-s3xf">mpifc</td>
<td class="tg-s3xf"><span style="background-color:#F7F7F7">gfortran</span></td>
<td class="tg-s3xf"><span style="background-color:#F7F7F7">Fortran77/Fortran 95</span></td>
</tr>
<tr>
<td class="tg-baqh" colspan="3">GNU Compilers</td>
</tr>
<tr>
<td class="tg-s3xf">mpigcc</td>
<td class="tg-s3xf"><span style="background-color:#F7F7F7">gcc</span></td>
<td class="tg-s3xf"><span style="background-color:#F7F7F7">C</span></td>
</tr>
<tr>
<td class="tg-lc14">mpigxx</td>
<td class="tg-lc14"><span style="background-color:#FFF">g++</span></td>
<td class="tg-lc14"><span style="background-color:#FFF">C/C++</span></td>
</tr>
<tr>
<td class="tg-s3xf">mpif77</td>
<td class="tg-s3xf"><span style="background-color:#F7F7F7">gfortran</span></td>
<td class="tg-s3xf"><span style="background-color:#F7F7F7">Fortran 77</span></td>
</tr>
<tr>
<td class="tg-lc14">mpif90</td>
<td class="tg-lc14"><span style="background-color:#FFF">gfortran</span></td>
<td class="tg-lc14"><span style="background-color:#FFF">Fortran 95</span></td>
</tr>
<tr>
<td class="tg-baqh" colspan="3">Intel® Fortran, C++ Compilers</td>
</tr>
<tr>
<td class="tg-lc14">mpiicc (classical)</td>
<td class="tg-lc14"><span style="background-color:#FFF">icc</span></td>
<td class="tg-lc14"><span style="background-color:#FFF">C</span></td>
</tr>
<tr>
<td class="tg-s3xf">mpiicx (oneapi)</td>
<td class="tg-s3xf"><span style="background-color:#F7F7F7">icx</span></td>
<td class="tg-s3xf"><span style="background-color:#F7F7F7">C</span></td>
</tr>
<tr>
<td class="tg-lc14">mpiicpc (classical)</td>
<td class="tg-lc14"><span style="background-color:#FFF">icpc</span></td>
<td class="tg-lc14"><span style="background-color:#FFF">C++</span></td>
</tr>
<tr>
<td class="tg-s3xf">mpiicpx (oneapi)</td>
<td class="tg-s3xf"><span style="background-color:#F7F7F7">icpx</span></td>
<td class="tg-s3xf"><span style="background-color:#F7F7F7">C++</span></td>
</tr>
<tr>
<td class="tg-26z5">mpiifort (classical)</td>
<td class="tg-26z5"><span style="background-color:#FFF">ifort</span></td>
<td class="tg-26z5"><span style="background-color:#FFF">Fortran77/Fortran 95</span></td>
</tr>
<tr>
<td class="tg-u97q">mpiifx (oneapi)</td>
<td class="tg-u97q"><span style="background-color:#F7F7F7">ifx</span></td>
<td class="tg-u97q"><span style="background-color:#F7F7F7">Fortran77/Fortran 95</span></td>
</tr>
</tbody>
</table>
In intel mpi, the mpirun command uses Hydra as the underlying process manager, which means mpirun invokes the mpiexec.hydra command (Hydra process manager).
[improve-performance-and-stability-with-intel-mpi-library-on-infiniband]([https:https:////](https://www.intel.com/content/www/us/en/developer/articles/technical/improve-performance-and-stability-with-intel-mpi-library-on-infiniband.html))
3. NVHPC
### Modulefile
- default modulefile
```
23:57:29 p00acy00@ilgn01:~$ echo $MODULEPATH
/etc/modulefiles:/usr/share/modulefiles:/pkg/modulefiles/software:/pkg/modulefiles/middleware/compiler:/pkg/x86/modulefiles
00:04:09 p00acy00@ilgn01:~$ module av
-------------------------- /pkg/modulefiles/software ---------------------------
apps/quantumatk/2.30.108 libs/hdf5/1.14.3
--------------------- /pkg/modulefiles/middleware/compiler ---------------------
gcc/8.5.0 gcc/10.4.0 gcc/11.2.0 intel/2024_01_46
----------------------------- /pkg/x86/modulefiles -----------------------------
code-server mvapich nvhpc/23.9
gcc/13.2.0 (D) nvhpc-byo-compiler/23.9 oneapi
gsl/2.7.1 nvhpc-hpcx-cuda12/23.9 openmpi
jupyterlab nvhpc-hpcx/23.9 paraview/5.12.0
miniconda3 nvhpc-nompi/23.9
Where:
D: Default Module
If the avail list is too long consider trying:
"module --default avail" or "ml -d av" to just list the default modules.
"module overview" or "ml ov" to display the number of modules for each name.
Use "module spider" to find all possible modules and extensions.
Use "module keyword key1 key2 ..." to search for all possible modules matching
any of the "keys".
```
- Add extra modulefiles (experimental!!! without guarantee)
```
00:06:12 p00acy00@ilgn01:~$ module use /home/qusim/modulefiles
00:10:38 p00acy00@ilgn01:~$ module av
--------------------------- /home/qusim/modulefiles ----------------------------
mpich/4.2.0-debug
mpich/4.2.0 (D)
openblas/0.3.26
openmpi/4.1.6-usr
openmpi/4.1.6
openmpi/5.0.2-base
openmpi/5.0.2-prrte
openmpi/5.0.2-usr
openmpi/5.0.2
python/3.10.13
sys/intelmpi/2021.9
sys/intelmpi/2024.1 (D)
sys/openmpi/4.1
sys/openmpi/5.0 (D)
valgrind/3.22
xacc/ompi5
-------------------------- /pkg/modulefiles/software ---------------------------
apps/quantumatk/2.30.108 libs/hdf5/1.14.3
--------------------- /pkg/modulefiles/middleware/compiler ---------------------
gcc/8.5.0 gcc/10.4.0 gcc/11.2.0 intel/2024_01_46
----------------------------- /pkg/x86/modulefiles -----------------------------
code-server mvapich nvhpc/23.9
gcc/13.2.0 (D) nvhpc-byo-compiler/23.9 oneapi
gsl/2.7.1 nvhpc-hpcx-cuda12/23.9 openmpi (D)
jupyterlab nvhpc-hpcx/23.9 paraview/5.12.0
miniconda3 nvhpc-nompi/23.9
Where:
D: Default Module
If the avail list is too long consider trying:
"module --default avail" or "ml -d av" to just list the default modules.
"module overview" or "ml ov" to display the number of modules for each name.
Use "module spider" to find all possible modules and extensions.
Use "module keyword key1 key2 ..." to search for all possible modules matching
any of the "keys".
```

## Testing project
For alphatest users, please use GOV113006 to submit job in slurm.
```
...
#SBATCH --account GOV113006
...
```

For beta test users, please use your own project.
## Queuing system
### Report partition summary information:
There are four default partitions in beta test: `betatest`, `ct112`, `ct448`, `ct1k`. As the test progresses, new partitions will be gradually added to the service.
```
(base) 20:55:31 p00acy00@ilgn01:~$ sinfo -s
PARTITION AVAIL TIMELIMIT NODES(A/I/O/T) NODELIST
alphatest inact 4:00:00 528/0/24/552 icpnp[101-156,201-256,301-348],icpnq[101-156,201-256,301-356,401-456,501-556,601-656,701-756]
betatest up infinite 528/0/24/552 icpnp[101-156,201-256,301-348],icpnq[101-156,201-256,301-356,401-456,501-556,601-656,701-756]
ct112 up 2-00:00:00 104/0/8/112 icpnp[101-156,201-256]
ct448 up 2-00:00:00 104/0/8/112 icpnp[101-156,201-256]
ct1k up 2-00:00:00 104/0/8/112 icpnp[101-156,201-256]
```
###
| 佇列名稱 | 可用核心數 | 可執行時間(hour) | 每位用戶可同時執行工作數 | 每位用戶可排隊工作數 | 系統最多可執行工作數 | 使用的計算節點清單 |
|:------------------:|:-----------------:|:-----------------------:|:-------------------------------:|:---------------------------:|:---------------------------:|:-------------------------:|
| development | 1~1120 | 4 | 1 | 1 | 40 | icpn[p3][05-48] |
| ct112 | 1~112 | 96 | 32 | 64 | 128 | icpn[p1-p2][01-56] |
| ct448 | 113~448 | 96 | 16 | 32 | 64 | icpn[p1-p2][01-56] |
| ct1k | 449~1120 | 96 | 8 | 16 | 32 | icpn[p1-p2][01-56] |
| ct2k | 1121~2240 | 96 | 4 | 8 | 16 | icpn[p1-p2][01-56] |
```
(base) 09:12:04 p00acy00@ilgn01:~$ scontrol show partition=ct112
PartitionName=ct112
AllowGroups=ALL AllowAccounts=gov112069,gov113006,mst110392,acd112143,mst112345,mst110307,mst113003,mst111344,mst111296,mst112149,mst111208,mst110349,mst110314,mst112240,mst111461 AllowQos=ALL
AllocNodes=ALL Default=NO QoS=p_ct112
DefaultTime=NONE DisableRootJobs=YES ExclusiveUser=NO GraceTime=0 Hidden=NO
MaxNodes=UNLIMITED MaxTime=2-00:00:00 MinNodes=0 LLN=NO MaxCPUsPerNode=UNLIMITED
Nodes=icpnp[101-156,201-256]
PriorityJobFactor=10 PriorityTier=1 RootOnly=NO ReqResv=NO OverSubscribe=NO
OverTimeLimit=NONE PreemptMode=OFF
State=UP TotalCPUs=12544 TotalNodes=112 SelectTypeParameters=NONE
JobDefaults=(null)
DefMemPerCPU=4308 MaxMemPerCPU=4308
TRES=cpu=12544,mem=54049184M,node=112,billing=12544
```
### Queuing policy
* Durning 3/27~4/5, the maxima execute time will be set to 4hrs.
* Durning 4/9~, the maxima execute time will be set to 2days.
## Application benchmark
### openmpi
```
export UCX_NET_DEVICES=mlx5_0:1
```
By default, UCX tries to use all available devices on the machine, and selects best ones based on performance characteristics (bandwidth, latency, NUMA locality, etc). Running `ucx_info -d` would show all available devices on the system that UCX can utilize.
[Which network devices does UCX use?](https://openucx.readthedocs.io/en/master/faq.html#selecting-networks-and-transports)
```
export UCX_TLS="ud,dc,rc,sm,self"
```
By default, UCX tries to use all available transports, and select best ones according to their performance capabilities and scale.
[Which transports does UCX use?](https://openucx.readthedocs.io/en/master/faq.html#which-transports-does-ucx-use)
| name | description |
| -------- | -------- |
| all | use all the available transports. |
| sm or shm | all shared memory transports. |
| ugni | ugni\_rdma and ugni\_udt. |
| rc | RC (=reliable connection), "accelerated" transports are used if possible. |
| ud | UD (=unreliable datagram), "accelerated" is used if possible. |
| dc | DC - Mellanox scalable offloaded dynamic connection transport |
| rc_x | Same as "rc", but using accelerated transports only |
| rc_v | Same as "rc", but using Verbs-based transports only |
| ud_x | Same as "ud", but using accelerated transports only |
| ud_v | Same as "ud", but using Verbs-based transports only |
| cuda | CUDA (NVIDIA GPU) memory support: cuda\_copy, cuda\_ipc, gdr_copy |
| rocm | ROCm (AMD GPU) memory support: rocm\_copy, rocm\_ipc, rocm_gdr |
| tcp | TCP over SOCK_STREAM sockets |
| self | Loopback transport to communicate within the same process |
[List of main transports and aliases](https://openucx.readthedocs.io/en/master/faq.html#list-of-main-transports-and-aliases)
### HPL


### OSU microbenchmark
### MPIIO
### WRF

### LAMMPS

### Quantum Espresso

## $\alpha$ test schedule (2024/2/1~2024/3/31)
<iframe src="https://calendar.google.com/calendar/embed?src=4a835eb060e51a12aa28cf30a854a804892040ef2e9b41c2452a1820b91e227e%40group.calendar.google.com&ctz=Asia%2FTaipei" style="border: 0" width="800" height="600" frameborder="0" scrolling="no"></iframe>
## $\beta$ test schedule (2024/4/1~2024/6/30)
<iframe src="https://calendar.google.com/calendar/embed?src=7c3ac2c23b089b9f845c511e6a344a5247032125471f849f1ccf253284ade581%40group.calendar.google.com&ctz=Asia%2FTaipei" style="border: 0" width="800" height="600" frameborder="0" scrolling="no"></iframe>
## 測試報告繳交
- report template
- [Forerunner_Benchmark_Report.docx](https://ndrive.narlabs.org.tw/navigate/a/#/s/71C654F56A2B4AB682BBA5A9789A76186BL)
- All participants must submit their test reports by 2024/7/31.
## Grand Challenge for F1 pilot run
- Only those who submit their written report before June 14th will be considered for participation.
- Assessment process will be based solely on the written report.
- The report content includes but is not limited to the following items:
- Methodology
- Background
- Benchmark & Results(Achievement)
- Cost-Effectiveness Comparison
- report template
- [Forerunner_Benchmark_Report.docx](https://ndrive.narlabs.org.tw/navigate/a/#/s/71C654F56A2B4AB682BBA5A9789A76186BL)
- Please send the test report to acyang@narlabs.org.tw
- We will be offering several awards, and winning teams will be invited to share their results at our user conference.
| Award | Number | prize |
|-----------------------|--------|----------------|
| Large-Scale Award | 1 | 50k F1 credits |
| High-Throughput Award | 1 | 30k F1 credits |
| Honorable Mention | | 10k F1 credits |
* The prize valid only from 2024/9~2024/12.
* NCHC reserves the right to modify and interpret the rules of the competition.
| Award | Winner | Topics |
|-----------------------|--------|----------------|
| Large-Scale Award | Chung-Gang Li | 不規則粗糙表面的自然對流誘導之紊流(Turbulent Natural Convection on an Irregular Rough Surface, TNCIRS) | The Department of Mechanical Engineering at National Cheng-Kung University
| Large-Scale Award | Wei-Hsiang Wang | 以沉浸式邊界探索邊界層分離現象 | The Mechanical Engineering department at National Chung Hsing University
| High-Throughput Award | Chin-Lung Kuo | 第一原理計算 (VASP) 來預測材料性質並且搭配古典力場模擬 (LAMMPS) 來觀察動力學現象。 | Department of Materials Science and Engineering, National Taiwan University