Files
gitea-pages/pages/merlin7/slurm-configuration.md

1.6 KiB

title, keywords, last_updated, summary, sidebar, permalink
title keywords last_updated summary sidebar permalink
Slurm cluster 'merlin7' configuration, partitions, node definition 24 Mai 2023 This document describes a summary of the Merlin6 configuration. merlin7_sidebar /merlin7/slurm-configuration.html

This documentation shows basic Slurm configuration and options needed to run jobs in the Merlin7 cluster.

Infrastructure

Hardware

The current configuration for the test phase is made up as:

  • 9 nodes for the PSI-Dev development system
    • 8 nodes were meant for baremetal and k8s
    • 1 login node
Node CPU RAM GRES Notes
Compute node 2x AMD EPYC 7713 (x86_64 Milan, 64 Cores, 3.2GHz) 512GB DDR4 3200Mhz 4x NVidia A100 (Ampere, 80GB)
Login node 2x AMD EPYC 7742 (x86_64 Rome, 64 Cores, 3.2GHz) 512GB DRR4 3200Mhz

Storage

  • CephFS only for /home -- 1 TB
  • ClusterStor L300 for /scratch -- 224 TB usable space
  • CephRBD /local -- 100GB

Node IDs

Cray user various identifies to uniquely label each node, details on this can be found on the Crayism page. The table below collates these together for the current configuration:

Node ID Cray XNAME Notes
nid003204 x1500c4s7b0n0 login node, to which psi-dev.cscs.ch points
nid002808 x1007c0s4b0n0
nid002809 x1007c0s4b0n1
nid002812 x1007c0s5b0n0
nid002813 x1007c0s5b0n1
nid002824 x1007c1s0b0n0
nid002825 x1007c1s0b0n1
nid002828 x1007c1s1b0n0
nid002829 x1007c1s1b0n1