Files
gitea-pages/pages/gmerlin6/hardware-and-software-description.md
2023-08-03 13:49:55 +02:00

10 KiB

title, last_updated, sidebar, permalink
title last_updated sidebar permalink
Hardware And Software Description 19 April 2021 merlin6_sidebar /gmerlin6/hardware-and-software.html

Hardware

GPU Computing Nodes

The GPU Merlin6 cluster was initially built from recycled workstations from different groups in the BIO division. From then, little by little it was updated with new nodes from sporadic investments from the same division, and it was never possible a central big investment. Hence, due to this, the Merlin6 GPU computing cluster has a non homogeneus solution, consisting on a big variety of hardware types and components.

On 2018, for the common good, BIO decided to open the cluster to the Merlin users and make it widely accessible for the PSI scientists.

The below table summarizes the hardware setup for the Merlin6 GPU computing nodes:

Merlin6 GPU Computing Nodes
Node Processor Sockets Cores Threads Scratch Memory GPUs GPU Model
merlin-g-001 Intel Core i7-5960X 1 16 2 1.8TB 128GB 2 GTX1080
merlin-g-00[2-5] Intel Xeon E5-2640 2 20 1 1.8TB 128GB 4 GTX1080
merlin-g-006 Intel Xeon E5-2640 2 20 1 800GB 128GB 4 GTX1080Ti
merlin-g-00[7-9] Intel Xeon E5-2640 2 20 1 3.5TB 128GB 4 GTX1080Ti
merlin-g-01[0-3] Intel Xeon Silver 4210R 2 20 1 1.7TB 128GB 4 RTX2080Ti
merlin-g-014 Intel Xeon Gold 6240R 2 48 1 2.9TB 384GB 8 RTX2080Ti
merlin-g-015 Intel(R) Xeon Gold 5318S 2 48 1 2.9TB 384GB 8 RTX A5000

Login Nodes

The login nodes are part of the Merlin6 HPC cluster, and are used to compile and to submit jobs to the different Merlin Slurm clusters (merlin5,merlin6,gmerlin6,etc.). Please refer to the Merlin6 Hardware Documentation for further information.

Storage

The storage is part of the Merlin6 HPC cluster, and is mounted in all the Slurm clusters (merlin5,merlin6,gmerlin6,etc.). Please refer to the Merlin6 Hardware Documentation for further information.

Network

The Merlin6 cluster connectivity is based on the Infiniband FDR and EDR technologies. This allows fast access with very low latencies to the data as well as running extremely efficient MPI-based jobs. To check the network speed (56Gbps for FDR, 100Gbps for EDR) of the different machines, it can be checked by running on each node the following command:

ibstat | grep Rate

Software

In the Merlin6 GPU computing nodes, we try to keep software stack coherency with the main cluster Merlin6.

Due to this, the Merlin6 GPU nodes run: