From 17053a363f29f32135e04b64321f20256b3be358 Mon Sep 17 00:00:00 2001 From: caubet_m Date: Tue, 10 Feb 2026 10:46:02 +0100 Subject: [PATCH] Clean Merlin6 docs --- docs/css/custom-nav.css | 37 +++++++++++++++ docs/gmerlin6/cluster-introduction.md | 16 +++---- .../hardware-and-software-description.md | 6 ++- docs/merlin6/how-to-use-merlin/storage.md | 2 - docs/merlin6/introduction.md | 26 ++-------- .../accessing-interactive-nodes.md | 1 - .../quick-start-guide/accessing-slurm.md | 13 ----- .../merlin6/quick-start-guide/introduction.md | 20 +------- .../quick-start-guide/requesting-accounts.md | 35 ++++---------- docs/merlin6/slurm-configuration.md | 9 +--- docs/merlin6/slurm-general-docs/monitoring.md | 47 ------------------- .../slurm-general-docs/running-jobs.md | 10 ++-- .../slurm-basic-commands.md | 20 -------- mkdocs.yml | 28 +++++------ 14 files changed, 85 insertions(+), 185 deletions(-) create mode 100644 docs/css/custom-nav.css diff --git a/docs/css/custom-nav.css b/docs/css/custom-nav.css new file mode 100644 index 0000000..a01be70 --- /dev/null +++ b/docs/css/custom-nav.css @@ -0,0 +1,37 @@ +/* Strong high-contrast section headers in MkDocs Material nav */ + +/* Top-level section blocks */ +.md-nav__title { + font-weight: 800; + font-size: 0.65rem; + + color: #ffffff !important; /* white text */ + background-color: #586ACB !important; /* strong dark blue */ + + padding: 6px 10px; + margin-top: 14px; +} + +/* For section-like items (common in newer Material versions) */ +.md-nav--primary .md-nav__item--section > .md-nav__link { + font-weight: 800; + font-size: 0.65rem; + + color: #ffffff !important; /* white text */ + background-color: #586ACB !important; + + padding: 6px 10px; + margin-top: 10px; +} + +/* For nested group headers (covers older/other layouts) */ +.md-nav--primary .md-nav__item--nested > .md-nav__link { + font-weight: 800; + font-size: 0.65rem; + + color: #ffffff !important; + background-color: #586ACB !important; + + padding: 6px 10px; + margin-top: 10px; +} diff --git a/docs/gmerlin6/cluster-introduction.md b/docs/gmerlin6/cluster-introduction.md index 8e83601..409d791 100644 --- a/docs/gmerlin6/cluster-introduction.md +++ b/docs/gmerlin6/cluster-introduction.md @@ -4,23 +4,23 @@ ### Introduction -Merlin6 is a the official PSI Local HPC cluster for development and -mission-critical applications that has been built in 2019. It replaces +Merlin6 was a the official PSI Local HPC cluster for development and +mission-critical applications that has been built in 2019. It replaced the Merlin5 cluster. -Merlin6 is designed to be extensible, so is technically possible to add +Merlin6 was designed to be extensible, so was technically possible to add more compute nodes and cluster storage without significant increase of the costs of the manpower and the operations. -Merlin6 is mostly based on **CPU** resources, but also contains a small amount +Merlin6 was mostly based on **CPU** resources, but also contained a small amount of **GPU**-based resources which are mostly used by the BIO experiments. ### Slurm 'gmerlin6' -THe **GPU nodes** have a dedicated **Slurm** cluster, called **`gmerli6`**. +THe **GPU nodes** have a dedicated **Slurm** cluster, called **`gmerlin6`**. This cluster contains the same shared storage resources (`/data/user`, `/data/project`, `/shared-scracth`, `/afs`, `/psi/home`) -which are present in the other Merlin Slurm clusters (`merlin5`,`merlin6`). The Slurm `gmerlin6` cluster is maintainted +which are present in the Merlin6 Slurm CPU cluster. The Slurm `gmerlin6` cluster is maintainted independently to ease access for the users and keep independent user accounting. ## Merlin6 Architecture @@ -31,8 +31,8 @@ The following image shows the Merlin6 cluster architecture diagram: ![Merlin6 Architecture Diagram](../images/merlinschema3.png) -### Merlin5 + Merlin6 Slurm Cluster Architecture Design +### Merlin6 Slurm Cluster Architecture Design -The following image shows the Slurm architecture design for the Merlin5 & Merlin6 clusters: +The following image shows the Slurm architecture design for the Merlin6 clusters: ![Merlin6 Slurm Architecture Design](../images/merlin-slurm-architecture.png) diff --git a/docs/gmerlin6/hardware-and-software-description.md b/docs/gmerlin6/hardware-and-software-description.md index dbeda84..6bdae12 100644 --- a/docs/gmerlin6/hardware-and-software-description.md +++ b/docs/gmerlin6/hardware-and-software-description.md @@ -10,7 +10,7 @@ Hence, due to this, the Merlin6 GPU computing cluster has a non homogeneus solut On 2018, for the common good, BIO decided to open the cluster to the Merlin users and make it widely accessible for the PSI scientists. -The below table summarizes the hardware setup for the Merlin6 GPU computing nodes: +The below table summarizes the hardware setup for the old Merlin6 GPU computing nodes: @@ -110,6 +110,10 @@ The below table summarizes the hardware setup for the Merlin6 GPU computing node
+### Gwendolen + +Currently only Gwendolen is available on `gmerlin6`. + ### Login Nodes The login nodes are part of the **[Merlin6](../merlin6/introduction.md)** HPC cluster, diff --git a/docs/merlin6/how-to-use-merlin/storage.md b/docs/merlin6/how-to-use-merlin/storage.md index 0f0e58a..df4f6a0 100644 --- a/docs/merlin6/how-to-use-merlin/storage.md +++ b/docs/merlin6/how-to-use-merlin/storage.md @@ -58,7 +58,6 @@ Properties of the directory classes: | ---------------------------------- | ----------------------- | ----------------------- | :-------: | :--------------------------------- |:-------------------------------- | ------ | :----------------------------- | | /psi/home/$username | USR [10GB:11GB] | *Undef* | N/A | Up to x2 when strongly justified. | N/A | yes | Daily snapshots for 1 week | | /data/user/$username | USR [1TB:1.074TB] | USR [1M:1.1M] | 7d | Inmutable. Need a project. | Changeable when justified. | no | Users responsible for backup | -| /data/project/bio/$projectname | GRP [1TB:1.074TB] | GRP [1M:1.1M] | 7d | Subject to project requirements. | Subject to project requirements. | no | Project responsible for backup | | /data/project/general/$projectname | GRP [1TB:1.074TB] | GRP [1M:1.1M] | 7d | Subject to project requirements. | Subject to project requirements. | no | Project responsible for backup | | /scratch | *Undef* | *Undef* | N/A | N/A | N/A | no | N/A | | /shared-scratch | USR [512GB:2TB] | USR [2M:2.5M] | 7d | Up to x2 when strongly justified. | Changeable when justified. | no | N/A | @@ -173,7 +172,6 @@ The properties of the available scratch storage spaces are given in the followin | Cluster | Service | Scratch | Scratch Mountpoint | Shared Scratch | Shared Scratch Mountpoint | Comments | | ------- | -------------- | ------------ | ------------------ | -------------- | ------------------------- | ------------------------------------ | -| merlin5 | computing node | 50GB / SAS | `/scratch` | `N/A` | `N/A` | `merlin-c-[01-64]` | | merlin6 | login node | 100GB / SAS | `/scratch` | 50TB / GPFS | `/shared-scratch` | `merlin-l-0[1,2]` | | merlin6 | computing node | 1.3TB / NVMe | `/scratch` | 50TB / GPFS | `/shared-scratch` | `merlin-c-[001-024,101-124,201-224]` | | merlin6 | login node | 2.0TB / NVMe | `/scratch` | 50TB / GPFS | `/shared-scratch` | `merlin-l-00[1,2]` | diff --git a/docs/merlin6/introduction.md b/docs/merlin6/introduction.md index 0bddc18..52a2e8c 100644 --- a/docs/merlin6/introduction.md +++ b/docs/merlin6/introduction.md @@ -1,25 +1,5 @@ -# Introduction +# About Merlin6 -!!! failure "Deprecated" - Merlin6 is no longer accessible for new users! Please have a look at - [Merlin 7](../merlin7/01-Quick-Start-Guide/introduction.md) instead. +Merlin6 is **no longer accessible for new users** and a smaller footprint of it has been kept for some of the stakeholder which owned nodes and can not fully run on the new cluster! -## Slurm clusters - -Comprised the **`merlin6`** CPU cluster and the -[**`gmerlin6`**](../gmerlin6/cluster-introduction.md) GPU cluster. A legacy -cluster (formally called [**`merlin5`**](../merlin5/cluster-introduction.md) is -kept active with best effort support. - -From July 2019, **`merlin6`** becomes the **default Slurm cluster** and any job -submitted from the login node will be submitted to that cluster if not. - -* Users can keep submitting to the old *`merlin5`* computing nodes by using the option ``--cluster=merlin5``. -* Users submitting to the **`gmerlin6`** GPU cluster need to specify the option ``--cluster=gmerlin6``. - -### Slurm 'merlin6' - -**CPU nodes** are configured in a **Slurm** cluster, called **`merlin6`**, and -this is the ***default Slurm cluster***. Hence, by default, if no Slurm cluster -is specified (with the `--cluster` option), this will be the cluster to which -the jobs will be sent. +Please have a look at [Merlin 7](../merlin7/01-Quick-Start-Guide/introduction.md) instead. diff --git a/docs/merlin6/quick-start-guide/accessing-interactive-nodes.md b/docs/merlin6/quick-start-guide/accessing-interactive-nodes.md index 297c551..e125986 100644 --- a/docs/merlin6/quick-start-guide/accessing-interactive-nodes.md +++ b/docs/merlin6/quick-start-guide/accessing-interactive-nodes.md @@ -46,4 +46,3 @@ The Merlin6 login nodes are the following: | ------------------- | --- | --------- | ------ |:--------:| :-------------------- | ------ | ---------- | :------------------ | | merlin-l-001.psi.ch | yes | yes | 2 x 22 | 2 | Intel Xeon Gold 6152 | 384GB | 1.8TB NVMe | ``/scratch`` | | merlin-l-002.psi.ch | yes | yes | 2 x 22 | 2 | Intel Xeon Gold 6142 | 384GB | 1.8TB NVMe | ``/scratch`` | -| merlin-l-01.psi.ch | yes | - | 2 x 16 | 2 | Intel Xeon E5-2697Av4 | 512GB | 100GB SAS | ``/scratch`` | diff --git a/docs/merlin6/quick-start-guide/accessing-slurm.md b/docs/merlin6/quick-start-guide/accessing-slurm.md index 4a07501..438d4d1 100644 --- a/docs/merlin6/quick-start-guide/accessing-slurm.md +++ b/docs/merlin6/quick-start-guide/accessing-slurm.md @@ -7,7 +7,6 @@ It basically contains the following clusters: * The **Merlin6 Slurm CPU cluster**, which is called [**`merlin6`**](#merlin6-cpu-cluster-access). * The **Merlin6 Slurm GPU cluster**, which is called [**`gmerlin6`**](#merlin6-gpu-cluster-access). -* The *old Merlin5 Slurm CPU cluster*, which is called [**`merlin5`**](#merlin5-cpu-cluster-access), still supported in a best effort basis. ## Accessing the Slurm clusters @@ -35,15 +34,3 @@ For further information about how to use this cluster, please visit: [**Merlin6 The **Merlin6 GPU cluster** (**`gmerlin6`**) is visible from the login nodes. However, to submit jobs to this cluster, one needs to specify the option `--cluster=gmerlin6` when submitting a job or allocation. For further information about how to use this cluster, please visit: [**Merlin6 GPU Slurm Cluster documentation**](../../gmerlin6/slurm-configuration.md). - -### Merlin5 CPU cluster access - -The **Merlin5 CPU cluster** (**`merlin5`**) is visible from the login nodes. However, to submit jobs -to this cluster, one needs to specify the option `--cluster=merlin5` when submitting a job or allocation. - -Using this cluster is in general not recommended, however this is still -available for old users needing extra computational resources or longer jobs. -Have in mind that this cluster is only supported in a **best effort basis**, -and it contains very old hardware and configurations. - -For further information about how to use this cluster, please visit the [**Merlin5 CPU Slurm Cluster documentation**](../../merlin5/slurm-configuration.md). diff --git a/docs/merlin6/quick-start-guide/introduction.md b/docs/merlin6/quick-start-guide/introduction.md index ac410c9..c6c6861 100644 --- a/docs/merlin6/quick-start-guide/introduction.md +++ b/docs/merlin6/quick-start-guide/introduction.md @@ -5,13 +5,6 @@ Historically, the local HPC clusters at PSI were named **Merlin**. Over the years, multiple generations of Merlin have been deployed. -At present, the **Merlin local HPC cluster** contains _two_ generations of it: - -* the old **Merlin5** cluster (`merlin5` Slurm cluster), and -* the newest generation **Merlin6**, which is divided in two Slurm clusters: - * `merlin6` as the Slurm CPU cluster - * `gmerlin6` as the Slurm GPU cluster. - Access to the different Slurm clusters is possible from the [**Merlin login nodes**](accessing-interactive-nodes.md), which can be accessed through the [SSH protocol](accessing-interactive-nodes.md#ssh-access) or the [NoMachine (NX) service](../how-to-use-merlin/nomachine.md). @@ -36,8 +29,8 @@ connected to the central PSI IT infrastructure. #### CPU and GPU Slurm clusters The Merlin6 **computing nodes** are mostly based on **CPU** resources. However, -it also contains a small amount of **GPU**-based resources, which are mostly used -by the BIO Division and by Deep Leaning project. +in the past it also contained a small amount of **GPU**-based resources, which were mostly used +by the BIO Division and by Deep Leaning projects. Today, only Gwendolen is available on `gmerlin6`. These computational resources are split into **two** different **[Slurm](https://slurm.schedmd.com/overview.html)** clusters: @@ -45,12 +38,3 @@ These computational resources are split into **two** different **[Slurm](https:/ * This is the **default Slurm cluster** configured in the login nodes: any job submitted without the option `--cluster` will be submited to this cluster. * The Merlin6 GPU resources are in a dedicated **[Slurm](https://slurm.schedmd.com/overview.html)** cluster called [**`gmerlin6`**](../../gmerlin6/slurm-configuration.md). * Users submitting to the **`gmerlin6`** GPU cluster need to specify the option ``--cluster=gmerlin6``. - -### Merlin5 - -The old Slurm **CPU** _Merlin_ cluster is still active and is maintained in a best effort basis. - -**Merlin5** only contains **computing nodes** resources in a dedicated **[Slurm](https://slurm.schedmd.com/overview.html)** cluster. - -* The Merlin5 CPU cluster is called [**merlin5**](../../merlin5/slurm-configuration.md). - diff --git a/docs/merlin6/quick-start-guide/requesting-accounts.md b/docs/merlin6/quick-start-guide/requesting-accounts.md index 75a3fc8..e9b6195 100644 --- a/docs/merlin6/quick-start-guide/requesting-accounts.md +++ b/docs/merlin6/quick-start-guide/requesting-accounts.md @@ -2,41 +2,26 @@ ## Requesting Access to Merlin6 -Access to Merlin6 is regulated by a PSI user's account being a member of the **`svc-cluster_merlin6`** group. Access to this group will also grant access to older generations of Merlin (`merlin5`). +In the past, access to the public Merlin6 cluster was regulated via the `svc-cluster_merlin6` group, which is no longer in use. +Merlin6 has become a private cluster, and to request access, **users must now be members of one of the Unix groups authorized to use it**, including Gwendolen. -Requesting **Merlin6** access *has to be done* with the corresponding **[Request Linux Group Membership](https://psi.service-now.com/psisp?id=psi_new_sc_cat_item&sys_id=84f2c0c81b04f110679febd9bb4bcbb1)** form, available in the [PSI Service Now Service Catalog](https://psi.service-now.com/psisp). +Requests for Merlin6 access must be submitted using the [Request Linux Group Membership](https://psi.service-now.com/psisp?id=psi_new_sc_cat_item&sys_id=84f2c0c81b04f110679febd9bb4bcbb1) form, available in the [PSI ServiceNow Service Catalog](https://psi.service-now.com/psisp). Access is granted by requesting membership in a Unix group that is permitted to use the cluster. ![Example: Requesting access to Merlin6](../../images/Access/01-request-merlin6-membership.png) -Mandatory customizable fields are the following: +### Mandatory fields -* **`Order Access for user`**, which defaults to the logged in user. However, requesting access for another user it's also possible. -* **`Request membership for group`**, for Merlin6 the **`svc-cluster_merlin6`** must be selected. -* **`Justification`**, please add here a short justification why access to Merlin6 is necessary. +The following fields must be completed: -Once submitted, the Merlin responsible will approve the request as soon as possible (within the next few hours on working days). Once the request is approved, *it may take up to 30 minutes to get the account fully configured*. +* **Order Access for user**: Defaults to the currently logged-in user. Access may also be requested on behalf of another user. +* **Request membership for group**: Select a valid Unix group that has access to Merlin6. +* **Justification**: Provide a brief explanation of why access to this group is required. -## Requesting Access to Merlin5 - -Access to Merlin5 is regulated by a PSI user's account being a member of the **`svc-cluster_merlin5`** group. Access to this group does not grant access to newer generations of Merlin (`merlin6`, `gmerlin6`, and future ones). - -Requesting **Merlin5** access *has to be done* with the corresponding **[Request Linux Group Membership](https://psi.service-now.com/psisp?id=psi_new_sc_cat_item&sys_id=84f2c0c81b04f110679febd9bb4bcbb1)** form, available in the [PSI Service Now Service Catalog](https://psi.service-now.com/psisp). - -![Example: Requesting access to Merlin5](../../images/Access/01-request-merlin5-membership.png) - -Mandatory customizable fields are the following: - -* **`Order Access for user`**, which defaults to the logged in user. However, requesting access for another user it's also possible. -* **`Request membership for group`**, for Merlin5 the **`svc-cluster_merlin5`** must be selected. -* **`Justification`**, please add here a short justification why access to Merlin5 is necessary. - -Once submitted, the Merlin responsible will approve the request as soon as possible (within the next few hours on working days). Once the request is approved, *it may take up to 30 minutes to get the account fully configured*. +Once the request is submitted, the corresponding group administrators will review and approve it as soon as possible (typically within a few working hours). After approval, it may take up to 30 minutes for the account to be fully configured and access to become effective. ## Further documentation -Further information it's also available in the Linux Central Documentation: +Additional information is available in the Linux Central Documentation: * [Unix Group / Group Management for users](https://linux.psi.ch/documentation/services/user-guide/unix_groups.html) * [Unix Group / Group Management for group managers](https://linux.psi.ch/documentation/services/admin-guide/unix_groups.html) - -**Special thanks** to the **Linux Central Team** and **AIT** to make this possible. diff --git a/docs/merlin6/slurm-configuration.md b/docs/merlin6/slurm-configuration.md index 8a52999..cf4b2f3 100644 --- a/docs/merlin6/slurm-configuration.md +++ b/docs/merlin6/slurm-configuration.md @@ -8,9 +8,6 @@ The following table show default and maximum resources that can be used per node | Nodes | Def.#CPUs | Max.#CPUs | #Threads | Max.Mem/CPU | Max.Mem/Node | Max.Swap | Def.#GPUs | Max.#GPUs | |:--------------------:| ---------:| :--------:| :------: | :----------:| :-----------:| :-------:| :-------: | :-------: | -| merlin-c-[001-024] | 1 core | 44 cores | 2 | 352000 | 352000 | 10000 | N/A | N/A | -| merlin-c-[101-124] | 1 core | 44 cores | 2 | 352000 | 352000 | 10000 | N/A | N/A | -| merlin-c-[201-224] | 1 core | 44 cores | 2 | 352000 | 352000 | 10000 | N/A | N/A | | merlin-c-[301-312] | 1 core | 44 cores | 2 | 748800 | 748800 | 10000 | N/A | N/A | | merlin-c-[313-318] | 1 core | 44 cores | 1 | 748800 | 748800 | 10000 | N/A | N/A | | merlin-c-[319-324] | 1 core | 44 cores | 2 | 748800 | 748800 | 10000 | N/A | N/A | @@ -19,7 +16,7 @@ If nothing is specified, by default each core will use up to 8GB of memory. Memo `--mem-per-cpu=` options, and maximum memory allowed is `Max.Mem/Node`. In **`merlin6`**, Memory is considered a Consumable Resource, as well as the CPU. Hence, both resources will account when submitting a job, -and by default resources can not be oversubscribed. This is a main difference with the old **`merlin5`** cluster, when only CPU were accounted, +and by default resources can not be oversubscribed. and memory was by default oversubscribed. @@ -280,7 +277,5 @@ For understanding the Slurm configuration setup in the cluster, sometimes may be The previous configuration files which can be found in the login nodes, correspond exclusively to the **merlin6** cluster configuration files. -Configuration files for the old **merlin5** cluster or for the **gmerlin6** -cluster must be checked directly on any of the **merlin5** or **gmerlin6** -computing nodes (in example, by login in to one of the nodes while a job or an +Configuration files for the **gmerlin6** cluster must be checked directly on the **gmerlin6** computing nodes (in example, by login in to one of the nodes while a job or an active allocation is running). diff --git a/docs/merlin6/slurm-general-docs/monitoring.md b/docs/merlin6/slurm-general-docs/monitoring.md index 7996ccf..e89b771 100644 --- a/docs/merlin6/slurm-general-docs/monitoring.md +++ b/docs/merlin6/slurm-general-docs/monitoring.md @@ -63,45 +63,6 @@ Check in the **man** pages (`man sinfo`) for all possible options for this comma gpu up 7-00:00:00 1-infinite no NO all 8 allocated merlin-g-[001-006,008-009] ``` -### Slurm commander - -The **[Slurm Commander (scom)](https://github.com/CLIP-HPC/SlurmCommander/)** is a simple but very useful open source text-based user interface for -simple and efficient interaction with Slurm. It is developed by the **CLoud Infrastructure Project (CLIP-HPC)** and external contributions. To use it, one can -simply run the following command: - -```bash -scom # merlin6 cluster -SLURM_CLUSTERS=merlin5 scom # merlin5 cluster -SLURM_CLUSTERS=gmerlin6 scom # gmerlin6 cluster -scom -h # Help and extra options -scom -d 14 # Set Job History to 14 days (instead of default 7) -``` - -With this simple interface, users can interact with their jobs, as well as getting information about past and present jobs: - -* Filtering jobs by substring is possible with the `/` key. -* Users can perform multiple actions on their jobs (such like cancelling, - holding or requeing a job), SSH to a node with an already running job, - or getting extended details and statistics of the job itself. - -Also, users can check the status of the cluster, to get statistics and node usage information as well as getting information about node properties. - -The interface also provides a few job templates for different use cases (i.e. MPI, OpenMP, Hybrid, single core). Users can modify these templates, -save it locally to the current directory, and submit the job to the cluster. - -!!! note - Currently, `scom` does not provide live updated information for the [Job History] tab. To update Job History - information, users have to exit the application with the q key. Other tabs will be updated every 5 - seconds (default). On the other hand, the [Job - History] tab contains only information for the **merlin6** CPU cluster - only. Future updates will provide information for other clusters. - -For further information about how to use **scom**, please refer to the **[Slurm Commander Project webpage](https://github.com/CLIP-HPC/SlurmCommander/)** - -!['scom' text-based user interface](../../images/slurm/scom.gif) - ### Job accounting Users can check detailed information of jobs (pending, running, completed, failed, etc.) with the `sacct` command. @@ -267,11 +228,3 @@ support: * Nodes monitoring: * [Merlin6 CPU Nodes Overview](https://hpc-monitor02.psi.ch/d/JmvLR8gZz/merlin6-computing-cpu-nodes?orgId=1&refresh=10s) * [Merlin6 GPU Nodes Overview](https://hpc-monitor02.psi.ch/d/gOo1Z10Wk/merlin6-computing-gpu-nodes?orgId=1&refresh=10s) - -### Merlin5 Monitoring Pages - -* Slurm monitoring: - * [Merlin5 Slurm Live Status](https://hpc-monitor02.psi.ch/d/o8msZJ0Zz/merlin5-slurm-live-status?orgId=1&refresh=10s) - * [Merlin5 Slurm Overview](https://hpc-monitor02.psi.ch/d/eWLEW1AWz/merlin5-slurm-overview?orgId=1&refresh=10s) -* Nodes monitoring: - * [Merlin5 CPU Nodes Overview](https://hpc-monitor02.psi.ch/d/ejTyWJAWk/merlin5-computing-cpu-nodes?orgId=1&refresh=10s) diff --git a/docs/merlin6/slurm-general-docs/running-jobs.md b/docs/merlin6/slurm-general-docs/running-jobs.md index 162e1f7..14ab6fe 100644 --- a/docs/merlin6/slurm-general-docs/running-jobs.md +++ b/docs/merlin6/slurm-general-docs/running-jobs.md @@ -49,10 +49,10 @@ The following settings are the minimum required for running a job in the Merlin * **Clusters:** For running jobs in the different Slurm clusters, users should to add the following option: ```bash - #SBATCH --clusters= # Possible values: merlin5, merlin6, gmerlin6 + #SBATCH --clusters= # Possible values: merlin6, gmerlin6 ``` - Refer to the documentation of each cluster ([**`merlin6`**](../slurm-configuration.md),[**`gmerlin6`**](../../gmerlin6/slurm-configuration.md),[**`merlin5`**](../../merlin5/slurm-configuration.md) for further information. + Refer to the documentation of each cluster ([**`merlin6`**](../slurm-configuration.md),[**`gmerlin6`**](../../gmerlin6/slurm-configuration.md) for further information. * **Partitions:** except when using the *default* partition for each cluster, one needs to specify the partition: @@ -60,7 +60,7 @@ The following settings are the minimum required for running a job in the Merlin #SBATCH --partition= # Check each cluster documentation for possible values ``` - Refer to the documentation of each cluster ([**`merlin6`**](../slurm-configuration.md),[**`gmerlin6`**](../../gmerlin6/slurm-configuration.md),[**`merlin5`**](../../merlin5/slurm-configuration.md) for further information. + Refer to the documentation of each cluster ([**`merlin6`**](../slurm-configuration.md),[**`gmerlin6`**](../../gmerlin6/slurm-configuration.md) for further information. * **[Optional] Disabling shared nodes**: by default, nodes are not exclusive. Hence, multiple users can run in the same node. One can request exclusive node usage with the following option: @@ -74,7 +74,7 @@ The following settings are the minimum required for running a job in the Merlin #SBATCH --time= # Can not exceed the partition `MaxTime` ``` - Refer to the documentation of each cluster ([**`merlin6`**](../slurm-configuration.md),[**`gmerlin6`**](../../gmerlin6/slurm-configuration.md),[**`merlin5`**](../../merlin5/slurm-configuration.md) for further information about partition `MaxTime` values. + Refer to the documentation of each cluster ([**`merlin6`**](../slurm-configuration.md),[**`gmerlin6`**](../../gmerlin6/slurm-configuration.md) for further information about partition `MaxTime` values. * **Output and error files**: by default, Slurm script will generate standard output (`slurm-%j.out`, where `%j` is the job_id) and error (`slurm-%j.err`, where `%j` is the job_id) files in the directory from where the job was submitted. Users can change default name with the following options: @@ -92,7 +92,7 @@ The following settings are the minimum required for running a job in the Merlin #SBATCH --hint=nomultithread # Don't use extra threads with in-core multi-threading. ``` - Refer to the documentation of each cluster ([**`merlin6`**](../slurm-configuration.md),[**`gmerlin6`**](../../gmerlin6/slurm-configuration.md),[**`merlin5`**](../../merlin5/slurm-configuration.md) for further information about node configuration and Hyper-Threading. + Refer to the documentation of each cluster ([**`merlin6`**](../slurm-configuration.md),[**`gmerlin6`**](../../gmerlin6/slurm-configuration.md) for further information about node configuration and Hyper-Threading. Consider that, sometimes, depending on your job requirements, you might need also to setup how many `--ntasks-per-core` or `--cpus-per-task` (even other options) in addition to the `--hint` command. Please, contact us in case of doubts. !!! tip diff --git a/docs/merlin6/slurm-general-docs/slurm-basic-commands.md b/docs/merlin6/slurm-general-docs/slurm-basic-commands.md index 972a6ef..dd5b096 100644 --- a/docs/merlin6/slurm-general-docs/slurm-basic-commands.md +++ b/docs/merlin6/slurm-general-docs/slurm-basic-commands.md @@ -33,23 +33,3 @@ sshare -a # to list shares of associations to a cluster sprio -l # to view the factors that comprise a job's scheduling priority # add '-u ' for filtering user ``` - -## Show information for specific cluster - -By default, any of the above commands shows information of the local cluster which is **merlin6**. - -If you want to see the same information for **merlin5** you have to add the parameter ``--clusters=merlin5``. -If you want to see both clusters at the same time, add the option ``--federation``. - -Examples: - -```bash -sinfo # 'sinfo' local cluster which is 'merlin6' -sinfo --clusters=merlin5 # 'sinfo' non-local cluster 'merlin5' -sinfo --federation # 'sinfo' all clusters which are 'merlin5' & 'merlin6' -squeue # 'squeue' local cluster which is 'merlin6' -squeue --clusters=merlin5 # 'squeue' non-local cluster 'merlin5' -squeue --federation # 'squeue' all clusters which are 'merlin5' & 'merlin6' -``` - ---- diff --git a/mkdocs.yml b/mkdocs.yml index b9433ba..f715e7e 100644 --- a/mkdocs.yml +++ b/mkdocs.yml @@ -11,6 +11,10 @@ edit_uri: _edit/master/docs/ # Copyright copyright: Copyright © 2025 HPC and Emerging Technologies Group/CSD @ Paul Scherrer Institut +# CSS +extra_css: + - css/custom-nav.css + # Configuration theme: name: material @@ -82,7 +86,7 @@ nav: - Home: index.md - News: - news/index.md - - Merlin 7: + - Merlin 7: - Quick Start Guide: - merlin7/01-Quick-Start-Guide/introduction.md - merlin7/01-Quick-Start-Guide/code-of-conduct.md @@ -126,11 +130,9 @@ nav: - Support: - merlin7/99-support/migration-from-merlin6.md - Merlin 6: - - merlin6/introduction.md - - merlin6/hardware-and-software-description.md - - merlin6/slurm-configuration.md + - Introduction: + - merlin6/introduction.md - Quick Start Guide: - - merlin6/quick-start-guide/introduction.md - merlin6/quick-start-guide/code-of-conduct.md - merlin6/quick-start-guide/requesting-accounts.md - merlin6/quick-start-guide/requesting-projects.md @@ -143,7 +145,6 @@ nav: - merlin6/how-to-use-merlin/storage.md - merlin6/how-to-use-merlin/transfer-data.md - merlin6/how-to-use-merlin/archive.md - - merlin6/how-to-use-merlin/nomachine.md - merlin6/how-to-use-merlin/ssh-keys.md - merlin6/how-to-use-merlin/kerberos.md - merlin6/how-to-use-merlin/using-modules.md @@ -157,10 +158,6 @@ nav: - gmerlin6/cluster-introduction.md - gmerlin6/hardware-and-software-description.md - gmerlin6/slurm-configuration.md - - Merlin 5 CPU Cluster: - - merlin5/cluster-introduction.md - - merlin5/hardware-and-software-description.md - - merlin5/slurm-configuration.md - Jupyterhub: - merlin6/jupyterhub/jupyterhub.md - merlin6/jupyterhub/jupyter-examples.md @@ -181,11 +178,6 @@ nav: - merlin6/software-support/openmpi.md - merlin6/software-support/paraview.md - merlin6/software-support/python.md - - Announcements: - - merlin6/98-announcements/downtimes.md - - merlin6/98-announcements/past-downtimes.md - - Support: - - merlin6/99-support/migration-from-merlin5.md - MeG: - meg/index.md - meg/contact.md @@ -198,3 +190,9 @@ nav: - support/faq.md - support/known-problems.md - support/troubleshooting.md + +markdown_extensions: + - pymdownx.superfences: + custom_fences: + - name: mermaid + class: mermaid