--- title: Hardware And Software Description #tags: #keywords: last_updated: 19 April 2021 #summary: "" sidebar: merlin6_sidebar permalink: /gmerlin6/hardware-and-software.html --- ## Hardware ### GPU Computing Nodes The GPU Merlin6 cluster was initially built from recycled workstations from different groups in the BIO division. From then, little by little it was updated with new nodes from sporadic investments from the same division, and it was never possible a central big investment. Hence, due to this, the Merlin6 GPU computing cluster has a non homogeneus solution, consisting on a big variety of hardware types and components. On 2018, for the common good, BIO decided to open the cluster to the Merlin users and make it widely accessible for the PSI scientists. The below table summarizes the hardware setup for the Merlin6 GPU computing nodes:
Merlin6 GPU Computing Nodes
Node Processor Sockets Cores Threads Scratch Memory GPUs GPU Model
merlin-g-001 Intel Core i7-5960X 1 16 2 1.8TB 128GB 2 GTX1080
merlin-g-00[2-5] Intel Xeon E5-2640 2 20 1 1.8TB 128GB 4 GTX1080
merlin-g-006 Intel Xeon E5-2640 2 20 1 800GB 128GB 4 GTX1080Ti
merlin-g-00[7-9] Intel Xeon E5-2640 2 20 1 3.5TB 128GB 4 GTX1080Ti
merlin-g-01[0-3] Intel Xeon Silver 4210R 2 20 1 1.7TB 128GB 4 RTX2080Ti
merlin-g-014 Intel Xeon Gold 6240R 2 48 1 2.9TB 384GB 8 RTX2080Ti
merlin-g-015 Intel(R) Xeon Gold 5318S 2 48 1 2.9TB 384GB 8 RTX A5000
### Login Nodes The login nodes are part of the **[Merlin6](/merlin6/introduction.html)** HPC cluster, and are used to compile and to submit jobs to the different ***Merlin Slurm clusters*** (`merlin5`,`merlin6`,`gmerlin6`,etc.). Please refer to the **[Merlin6 Hardware Documentation](/merlin6/hardware-and-software.html)** for further information. ### Storage The storage is part of the **[Merlin6](/merlin6/introduction.html)** HPC cluster, and is mounted in all the ***Slurm clusters*** (`merlin5`,`merlin6`,`gmerlin6`,etc.). Please refer to the **[Merlin6 Hardware Documentation](/merlin6/hardware-and-software.html)** for further information. ### Network The Merlin6 cluster connectivity is based on the [Infiniband FDR and EDR](https://en.wikipedia.org/wiki/InfiniBand) technologies. This allows fast access with very low latencies to the data as well as running extremely efficient MPI-based jobs. To check the network speed (56Gbps for **FDR**, 100Gbps for **EDR**) of the different machines, it can be checked by running on each node the following command: ```bash ibstat | grep Rate ``` ## Software In the Merlin6 GPU computing nodes, we try to keep software stack coherency with the main cluster [Merlin6](/merlin6/index.html). Due to this, the Merlin6 GPU nodes run: * [**RedHat Enterprise Linux 7**](https://access.redhat.com/documentation/en-us/red_hat_enterprise_linux/7/html/7.9_release_notes/index) * [**Slurm**](https://slurm.schedmd.com/), we usually try to keep it up to date with the most recent versions. * [**GPFS v5**](https://www.ibm.com/support/knowledgecenter/en/STXKQY_5.0.2/ibmspectrumscale502_welcome.html) * [**MLNX_OFED LTS v.5.2-2.2.0.0 or newer**](https://www.mellanox.com/products/infiniband-drivers/linux/mlnx_ofed) for all **ConnectX-4** or superior cards.