diff --git a/docs/css/custom-nav.css b/docs/css/custom-nav.css
new file mode 100644
index 0000000..a01be70
--- /dev/null
+++ b/docs/css/custom-nav.css
@@ -0,0 +1,37 @@
+/* Strong high-contrast section headers in MkDocs Material nav */
+
+/* Top-level section blocks */
+.md-nav__title {
+ font-weight: 800;
+ font-size: 0.65rem;
+
+ color: #ffffff !important; /* white text */
+ background-color: #586ACB !important; /* strong dark blue */
+
+ padding: 6px 10px;
+ margin-top: 14px;
+}
+
+/* For section-like items (common in newer Material versions) */
+.md-nav--primary .md-nav__item--section > .md-nav__link {
+ font-weight: 800;
+ font-size: 0.65rem;
+
+ color: #ffffff !important; /* white text */
+ background-color: #586ACB !important;
+
+ padding: 6px 10px;
+ margin-top: 10px;
+}
+
+/* For nested group headers (covers older/other layouts) */
+.md-nav--primary .md-nav__item--nested > .md-nav__link {
+ font-weight: 800;
+ font-size: 0.65rem;
+
+ color: #ffffff !important;
+ background-color: #586ACB !important;
+
+ padding: 6px 10px;
+ margin-top: 10px;
+}
diff --git a/docs/gmerlin6/cluster-introduction.md b/docs/gmerlin6/cluster-introduction.md
index 8e83601..409d791 100644
--- a/docs/gmerlin6/cluster-introduction.md
+++ b/docs/gmerlin6/cluster-introduction.md
@@ -4,23 +4,23 @@
### Introduction
-Merlin6 is a the official PSI Local HPC cluster for development and
-mission-critical applications that has been built in 2019. It replaces
+Merlin6 was a the official PSI Local HPC cluster for development and
+mission-critical applications that has been built in 2019. It replaced
the Merlin5 cluster.
-Merlin6 is designed to be extensible, so is technically possible to add
+Merlin6 was designed to be extensible, so was technically possible to add
more compute nodes and cluster storage without significant increase of
the costs of the manpower and the operations.
-Merlin6 is mostly based on **CPU** resources, but also contains a small amount
+Merlin6 was mostly based on **CPU** resources, but also contained a small amount
of **GPU**-based resources which are mostly used by the BIO experiments.
### Slurm 'gmerlin6'
-THe **GPU nodes** have a dedicated **Slurm** cluster, called **`gmerli6`**.
+THe **GPU nodes** have a dedicated **Slurm** cluster, called **`gmerlin6`**.
This cluster contains the same shared storage resources (`/data/user`, `/data/project`, `/shared-scracth`, `/afs`, `/psi/home`)
-which are present in the other Merlin Slurm clusters (`merlin5`,`merlin6`). The Slurm `gmerlin6` cluster is maintainted
+which are present in the Merlin6 Slurm CPU cluster. The Slurm `gmerlin6` cluster is maintainted
independently to ease access for the users and keep independent user accounting.
## Merlin6 Architecture
@@ -31,8 +31,8 @@ The following image shows the Merlin6 cluster architecture diagram:

-### Merlin5 + Merlin6 Slurm Cluster Architecture Design
+### Merlin6 Slurm Cluster Architecture Design
-The following image shows the Slurm architecture design for the Merlin5 & Merlin6 clusters:
+The following image shows the Slurm architecture design for the Merlin6 clusters:

diff --git a/docs/gmerlin6/hardware-and-software-description.md b/docs/gmerlin6/hardware-and-software-description.md
index dbeda84..6bdae12 100644
--- a/docs/gmerlin6/hardware-and-software-description.md
+++ b/docs/gmerlin6/hardware-and-software-description.md
@@ -10,7 +10,7 @@ Hence, due to this, the Merlin6 GPU computing cluster has a non homogeneus solut
On 2018, for the common good, BIO decided to open the cluster to the Merlin users and make it widely accessible for the PSI scientists.
-The below table summarizes the hardware setup for the Merlin6 GPU computing nodes:
+The below table summarizes the hardware setup for the old Merlin6 GPU computing nodes:
@@ -110,6 +110,10 @@ The below table summarizes the hardware setup for the Merlin6 GPU computing node
+### Gwendolen
+
+Currently only Gwendolen is available on `gmerlin6`.
+
### Login Nodes
The login nodes are part of the **[Merlin6](../merlin6/introduction.md)** HPC cluster,
diff --git a/docs/merlin6/how-to-use-merlin/storage.md b/docs/merlin6/how-to-use-merlin/storage.md
index 0f0e58a..df4f6a0 100644
--- a/docs/merlin6/how-to-use-merlin/storage.md
+++ b/docs/merlin6/how-to-use-merlin/storage.md
@@ -58,7 +58,6 @@ Properties of the directory classes:
| ---------------------------------- | ----------------------- | ----------------------- | :-------: | :--------------------------------- |:-------------------------------- | ------ | :----------------------------- |
| /psi/home/$username | USR [10GB:11GB] | *Undef* | N/A | Up to x2 when strongly justified. | N/A | yes | Daily snapshots for 1 week |
| /data/user/$username | USR [1TB:1.074TB] | USR [1M:1.1M] | 7d | Inmutable. Need a project. | Changeable when justified. | no | Users responsible for backup |
-| /data/project/bio/$projectname | GRP [1TB:1.074TB] | GRP [1M:1.1M] | 7d | Subject to project requirements. | Subject to project requirements. | no | Project responsible for backup |
| /data/project/general/$projectname | GRP [1TB:1.074TB] | GRP [1M:1.1M] | 7d | Subject to project requirements. | Subject to project requirements. | no | Project responsible for backup |
| /scratch | *Undef* | *Undef* | N/A | N/A | N/A | no | N/A |
| /shared-scratch | USR [512GB:2TB] | USR [2M:2.5M] | 7d | Up to x2 when strongly justified. | Changeable when justified. | no | N/A |
@@ -173,7 +172,6 @@ The properties of the available scratch storage spaces are given in the followin
| Cluster | Service | Scratch | Scratch Mountpoint | Shared Scratch | Shared Scratch Mountpoint | Comments |
| ------- | -------------- | ------------ | ------------------ | -------------- | ------------------------- | ------------------------------------ |
-| merlin5 | computing node | 50GB / SAS | `/scratch` | `N/A` | `N/A` | `merlin-c-[01-64]` |
| merlin6 | login node | 100GB / SAS | `/scratch` | 50TB / GPFS | `/shared-scratch` | `merlin-l-0[1,2]` |
| merlin6 | computing node | 1.3TB / NVMe | `/scratch` | 50TB / GPFS | `/shared-scratch` | `merlin-c-[001-024,101-124,201-224]` |
| merlin6 | login node | 2.0TB / NVMe | `/scratch` | 50TB / GPFS | `/shared-scratch` | `merlin-l-00[1,2]` |
diff --git a/docs/merlin6/introduction.md b/docs/merlin6/introduction.md
index 0bddc18..52a2e8c 100644
--- a/docs/merlin6/introduction.md
+++ b/docs/merlin6/introduction.md
@@ -1,25 +1,5 @@
-# Introduction
+# About Merlin6
-!!! failure "Deprecated"
- Merlin6 is no longer accessible for new users! Please have a look at
- [Merlin 7](../merlin7/01-Quick-Start-Guide/introduction.md) instead.
+Merlin6 is **no longer accessible for new users** and a smaller footprint of it has been kept for some of the stakeholder which owned nodes and can not fully run on the new cluster!
-## Slurm clusters
-
-Comprised the **`merlin6`** CPU cluster and the
-[**`gmerlin6`**](../gmerlin6/cluster-introduction.md) GPU cluster. A legacy
-cluster (formally called [**`merlin5`**](../merlin5/cluster-introduction.md) is
-kept active with best effort support.
-
-From July 2019, **`merlin6`** becomes the **default Slurm cluster** and any job
-submitted from the login node will be submitted to that cluster if not.
-
-* Users can keep submitting to the old *`merlin5`* computing nodes by using the option ``--cluster=merlin5``.
-* Users submitting to the **`gmerlin6`** GPU cluster need to specify the option ``--cluster=gmerlin6``.
-
-### Slurm 'merlin6'
-
-**CPU nodes** are configured in a **Slurm** cluster, called **`merlin6`**, and
-this is the ***default Slurm cluster***. Hence, by default, if no Slurm cluster
-is specified (with the `--cluster` option), this will be the cluster to which
-the jobs will be sent.
+Please have a look at [Merlin 7](../merlin7/01-Quick-Start-Guide/introduction.md) instead.
diff --git a/docs/merlin6/quick-start-guide/accessing-interactive-nodes.md b/docs/merlin6/quick-start-guide/accessing-interactive-nodes.md
index 297c551..e125986 100644
--- a/docs/merlin6/quick-start-guide/accessing-interactive-nodes.md
+++ b/docs/merlin6/quick-start-guide/accessing-interactive-nodes.md
@@ -46,4 +46,3 @@ The Merlin6 login nodes are the following:
| ------------------- | --- | --------- | ------ |:--------:| :-------------------- | ------ | ---------- | :------------------ |
| merlin-l-001.psi.ch | yes | yes | 2 x 22 | 2 | Intel Xeon Gold 6152 | 384GB | 1.8TB NVMe | ``/scratch`` |
| merlin-l-002.psi.ch | yes | yes | 2 x 22 | 2 | Intel Xeon Gold 6142 | 384GB | 1.8TB NVMe | ``/scratch`` |
-| merlin-l-01.psi.ch | yes | - | 2 x 16 | 2 | Intel Xeon E5-2697Av4 | 512GB | 100GB SAS | ``/scratch`` |
diff --git a/docs/merlin6/quick-start-guide/accessing-slurm.md b/docs/merlin6/quick-start-guide/accessing-slurm.md
index 4a07501..438d4d1 100644
--- a/docs/merlin6/quick-start-guide/accessing-slurm.md
+++ b/docs/merlin6/quick-start-guide/accessing-slurm.md
@@ -7,7 +7,6 @@ It basically contains the following clusters:
* The **Merlin6 Slurm CPU cluster**, which is called [**`merlin6`**](#merlin6-cpu-cluster-access).
* The **Merlin6 Slurm GPU cluster**, which is called [**`gmerlin6`**](#merlin6-gpu-cluster-access).
-* The *old Merlin5 Slurm CPU cluster*, which is called [**`merlin5`**](#merlin5-cpu-cluster-access), still supported in a best effort basis.
## Accessing the Slurm clusters
@@ -35,15 +34,3 @@ For further information about how to use this cluster, please visit: [**Merlin6
The **Merlin6 GPU cluster** (**`gmerlin6`**) is visible from the login nodes. However, to submit jobs to this cluster, one needs to specify the option `--cluster=gmerlin6` when submitting a job or allocation.
For further information about how to use this cluster, please visit: [**Merlin6 GPU Slurm Cluster documentation**](../../gmerlin6/slurm-configuration.md).
-
-### Merlin5 CPU cluster access
-
-The **Merlin5 CPU cluster** (**`merlin5`**) is visible from the login nodes. However, to submit jobs
-to this cluster, one needs to specify the option `--cluster=merlin5` when submitting a job or allocation.
-
-Using this cluster is in general not recommended, however this is still
-available for old users needing extra computational resources or longer jobs.
-Have in mind that this cluster is only supported in a **best effort basis**,
-and it contains very old hardware and configurations.
-
-For further information about how to use this cluster, please visit the [**Merlin5 CPU Slurm Cluster documentation**](../../merlin5/slurm-configuration.md).
diff --git a/docs/merlin6/quick-start-guide/introduction.md b/docs/merlin6/quick-start-guide/introduction.md
index ac410c9..c6c6861 100644
--- a/docs/merlin6/quick-start-guide/introduction.md
+++ b/docs/merlin6/quick-start-guide/introduction.md
@@ -5,13 +5,6 @@
Historically, the local HPC clusters at PSI were named **Merlin**. Over the years,
multiple generations of Merlin have been deployed.
-At present, the **Merlin local HPC cluster** contains _two_ generations of it:
-
-* the old **Merlin5** cluster (`merlin5` Slurm cluster), and
-* the newest generation **Merlin6**, which is divided in two Slurm clusters:
- * `merlin6` as the Slurm CPU cluster
- * `gmerlin6` as the Slurm GPU cluster.
-
Access to the different Slurm clusters is possible from the [**Merlin login nodes**](accessing-interactive-nodes.md),
which can be accessed through the [SSH protocol](accessing-interactive-nodes.md#ssh-access) or the [NoMachine (NX) service](../how-to-use-merlin/nomachine.md).
@@ -36,8 +29,8 @@ connected to the central PSI IT infrastructure.
#### CPU and GPU Slurm clusters
The Merlin6 **computing nodes** are mostly based on **CPU** resources. However,
-it also contains a small amount of **GPU**-based resources, which are mostly used
-by the BIO Division and by Deep Leaning project.
+in the past it also contained a small amount of **GPU**-based resources, which were mostly used
+by the BIO Division and by Deep Leaning projects. Today, only Gwendolen is available on `gmerlin6`.
These computational resources are split into **two** different **[Slurm](https://slurm.schedmd.com/overview.html)** clusters:
@@ -45,12 +38,3 @@ These computational resources are split into **two** different **[Slurm](https:/
* This is the **default Slurm cluster** configured in the login nodes: any job submitted without the option `--cluster` will be submited to this cluster.
* The Merlin6 GPU resources are in a dedicated **[Slurm](https://slurm.schedmd.com/overview.html)** cluster called [**`gmerlin6`**](../../gmerlin6/slurm-configuration.md).
* Users submitting to the **`gmerlin6`** GPU cluster need to specify the option ``--cluster=gmerlin6``.
-
-### Merlin5
-
-The old Slurm **CPU** _Merlin_ cluster is still active and is maintained in a best effort basis.
-
-**Merlin5** only contains **computing nodes** resources in a dedicated **[Slurm](https://slurm.schedmd.com/overview.html)** cluster.
-
-* The Merlin5 CPU cluster is called [**merlin5**](../../merlin5/slurm-configuration.md).
-
diff --git a/docs/merlin6/quick-start-guide/requesting-accounts.md b/docs/merlin6/quick-start-guide/requesting-accounts.md
index 75a3fc8..e9b6195 100644
--- a/docs/merlin6/quick-start-guide/requesting-accounts.md
+++ b/docs/merlin6/quick-start-guide/requesting-accounts.md
@@ -2,41 +2,26 @@
## Requesting Access to Merlin6
-Access to Merlin6 is regulated by a PSI user's account being a member of the **`svc-cluster_merlin6`** group. Access to this group will also grant access to older generations of Merlin (`merlin5`).
+In the past, access to the public Merlin6 cluster was regulated via the `svc-cluster_merlin6` group, which is no longer in use.
+Merlin6 has become a private cluster, and to request access, **users must now be members of one of the Unix groups authorized to use it**, including Gwendolen.
-Requesting **Merlin6** access *has to be done* with the corresponding **[Request Linux Group Membership](https://psi.service-now.com/psisp?id=psi_new_sc_cat_item&sys_id=84f2c0c81b04f110679febd9bb4bcbb1)** form, available in the [PSI Service Now Service Catalog](https://psi.service-now.com/psisp).
+Requests for Merlin6 access must be submitted using the [Request Linux Group Membership](https://psi.service-now.com/psisp?id=psi_new_sc_cat_item&sys_id=84f2c0c81b04f110679febd9bb4bcbb1) form, available in the [PSI ServiceNow Service Catalog](https://psi.service-now.com/psisp). Access is granted by requesting membership in a Unix group that is permitted to use the cluster.

-Mandatory customizable fields are the following:
+### Mandatory fields
-* **`Order Access for user`**, which defaults to the logged in user. However, requesting access for another user it's also possible.
-* **`Request membership for group`**, for Merlin6 the **`svc-cluster_merlin6`** must be selected.
-* **`Justification`**, please add here a short justification why access to Merlin6 is necessary.
+The following fields must be completed:
-Once submitted, the Merlin responsible will approve the request as soon as possible (within the next few hours on working days). Once the request is approved, *it may take up to 30 minutes to get the account fully configured*.
+* **Order Access for user**: Defaults to the currently logged-in user. Access may also be requested on behalf of another user.
+* **Request membership for group**: Select a valid Unix group that has access to Merlin6.
+* **Justification**: Provide a brief explanation of why access to this group is required.
-## Requesting Access to Merlin5
-
-Access to Merlin5 is regulated by a PSI user's account being a member of the **`svc-cluster_merlin5`** group. Access to this group does not grant access to newer generations of Merlin (`merlin6`, `gmerlin6`, and future ones).
-
-Requesting **Merlin5** access *has to be done* with the corresponding **[Request Linux Group Membership](https://psi.service-now.com/psisp?id=psi_new_sc_cat_item&sys_id=84f2c0c81b04f110679febd9bb4bcbb1)** form, available in the [PSI Service Now Service Catalog](https://psi.service-now.com/psisp).
-
-
-
-Mandatory customizable fields are the following:
-
-* **`Order Access for user`**, which defaults to the logged in user. However, requesting access for another user it's also possible.
-* **`Request membership for group`**, for Merlin5 the **`svc-cluster_merlin5`** must be selected.
-* **`Justification`**, please add here a short justification why access to Merlin5 is necessary.
-
-Once submitted, the Merlin responsible will approve the request as soon as possible (within the next few hours on working days). Once the request is approved, *it may take up to 30 minutes to get the account fully configured*.
+Once the request is submitted, the corresponding group administrators will review and approve it as soon as possible (typically within a few working hours). After approval, it may take up to 30 minutes for the account to be fully configured and access to become effective.
## Further documentation
-Further information it's also available in the Linux Central Documentation:
+Additional information is available in the Linux Central Documentation:
* [Unix Group / Group Management for users](https://linux.psi.ch/documentation/services/user-guide/unix_groups.html)
* [Unix Group / Group Management for group managers](https://linux.psi.ch/documentation/services/admin-guide/unix_groups.html)
-
-**Special thanks** to the **Linux Central Team** and **AIT** to make this possible.
diff --git a/docs/merlin6/slurm-configuration.md b/docs/merlin6/slurm-configuration.md
index 8a52999..cf4b2f3 100644
--- a/docs/merlin6/slurm-configuration.md
+++ b/docs/merlin6/slurm-configuration.md
@@ -8,9 +8,6 @@ The following table show default and maximum resources that can be used per node
| Nodes | Def.#CPUs | Max.#CPUs | #Threads | Max.Mem/CPU | Max.Mem/Node | Max.Swap | Def.#GPUs | Max.#GPUs |
|:--------------------:| ---------:| :--------:| :------: | :----------:| :-----------:| :-------:| :-------: | :-------: |
-| merlin-c-[001-024] | 1 core | 44 cores | 2 | 352000 | 352000 | 10000 | N/A | N/A |
-| merlin-c-[101-124] | 1 core | 44 cores | 2 | 352000 | 352000 | 10000 | N/A | N/A |
-| merlin-c-[201-224] | 1 core | 44 cores | 2 | 352000 | 352000 | 10000 | N/A | N/A |
| merlin-c-[301-312] | 1 core | 44 cores | 2 | 748800 | 748800 | 10000 | N/A | N/A |
| merlin-c-[313-318] | 1 core | 44 cores | 1 | 748800 | 748800 | 10000 | N/A | N/A |
| merlin-c-[319-324] | 1 core | 44 cores | 2 | 748800 | 748800 | 10000 | N/A | N/A |
@@ -19,7 +16,7 @@ If nothing is specified, by default each core will use up to 8GB of memory. Memo
`--mem-per-cpu=` options, and maximum memory allowed is `Max.Mem/Node`.
In **`merlin6`**, Memory is considered a Consumable Resource, as well as the CPU. Hence, both resources will account when submitting a job,
-and by default resources can not be oversubscribed. This is a main difference with the old **`merlin5`** cluster, when only CPU were accounted,
+and by default resources can not be oversubscribed.
and memory was by default oversubscribed.
@@ -280,7 +277,5 @@ For understanding the Slurm configuration setup in the cluster, sometimes may be
The previous configuration files which can be found in the login nodes,
correspond exclusively to the **merlin6** cluster configuration files.
-Configuration files for the old **merlin5** cluster or for the **gmerlin6**
-cluster must be checked directly on any of the **merlin5** or **gmerlin6**
-computing nodes (in example, by login in to one of the nodes while a job or an
+Configuration files for the **gmerlin6** cluster must be checked directly on the **gmerlin6** computing nodes (in example, by login in to one of the nodes while a job or an
active allocation is running).
diff --git a/docs/merlin6/slurm-general-docs/monitoring.md b/docs/merlin6/slurm-general-docs/monitoring.md
index 7996ccf..e89b771 100644
--- a/docs/merlin6/slurm-general-docs/monitoring.md
+++ b/docs/merlin6/slurm-general-docs/monitoring.md
@@ -63,45 +63,6 @@ Check in the **man** pages (`man sinfo`) for all possible options for this comma
gpu up 7-00:00:00 1-infinite no NO all 8 allocated merlin-g-[001-006,008-009]
```
-### Slurm commander
-
-The **[Slurm Commander (scom)](https://github.com/CLIP-HPC/SlurmCommander/)** is a simple but very useful open source text-based user interface for
-simple and efficient interaction with Slurm. It is developed by the **CLoud Infrastructure Project (CLIP-HPC)** and external contributions. To use it, one can
-simply run the following command:
-
-```bash
-scom # merlin6 cluster
-SLURM_CLUSTERS=merlin5 scom # merlin5 cluster
-SLURM_CLUSTERS=gmerlin6 scom # gmerlin6 cluster
-scom -h # Help and extra options
-scom -d 14 # Set Job History to 14 days (instead of default 7)
-```
-
-With this simple interface, users can interact with their jobs, as well as getting information about past and present jobs:
-
-* Filtering jobs by substring is possible with the `/` key.
-* Users can perform multiple actions on their jobs (such like cancelling,
- holding or requeing a job), SSH to a node with an already running job,
- or getting extended details and statistics of the job itself.
-
-Also, users can check the status of the cluster, to get statistics and node usage information as well as getting information about node properties.
-
-The interface also provides a few job templates for different use cases (i.e. MPI, OpenMP, Hybrid, single core). Users can modify these templates,
-save it locally to the current directory, and submit the job to the cluster.
-
-!!! note
- Currently, `scom` does not provide live updated information for the [Job History] tab. To update Job History
- information, users have to exit the application with the q key. Other tabs will be updated every 5
- seconds (default). On the other hand, the [Job
- History] tab contains only information for the **merlin6** CPU cluster
- only. Future updates will provide information for other clusters.
-
-For further information about how to use **scom**, please refer to the **[Slurm Commander Project webpage](https://github.com/CLIP-HPC/SlurmCommander/)**
-
-
-
### Job accounting
Users can check detailed information of jobs (pending, running, completed, failed, etc.) with the `sacct` command.
@@ -267,11 +228,3 @@ support:
* Nodes monitoring:
* [Merlin6 CPU Nodes Overview](https://hpc-monitor02.psi.ch/d/JmvLR8gZz/merlin6-computing-cpu-nodes?orgId=1&refresh=10s)
* [Merlin6 GPU Nodes Overview](https://hpc-monitor02.psi.ch/d/gOo1Z10Wk/merlin6-computing-gpu-nodes?orgId=1&refresh=10s)
-
-### Merlin5 Monitoring Pages
-
-* Slurm monitoring:
- * [Merlin5 Slurm Live Status](https://hpc-monitor02.psi.ch/d/o8msZJ0Zz/merlin5-slurm-live-status?orgId=1&refresh=10s)
- * [Merlin5 Slurm Overview](https://hpc-monitor02.psi.ch/d/eWLEW1AWz/merlin5-slurm-overview?orgId=1&refresh=10s)
-* Nodes monitoring:
- * [Merlin5 CPU Nodes Overview](https://hpc-monitor02.psi.ch/d/ejTyWJAWk/merlin5-computing-cpu-nodes?orgId=1&refresh=10s)
diff --git a/docs/merlin6/slurm-general-docs/running-jobs.md b/docs/merlin6/slurm-general-docs/running-jobs.md
index 162e1f7..14ab6fe 100644
--- a/docs/merlin6/slurm-general-docs/running-jobs.md
+++ b/docs/merlin6/slurm-general-docs/running-jobs.md
@@ -49,10 +49,10 @@ The following settings are the minimum required for running a job in the Merlin
* **Clusters:** For running jobs in the different Slurm clusters, users should to add the following option:
```bash
- #SBATCH --clusters= # Possible values: merlin5, merlin6, gmerlin6
+ #SBATCH --clusters= # Possible values: merlin6, gmerlin6
```
- Refer to the documentation of each cluster ([**`merlin6`**](../slurm-configuration.md),[**`gmerlin6`**](../../gmerlin6/slurm-configuration.md),[**`merlin5`**](../../merlin5/slurm-configuration.md) for further information.
+ Refer to the documentation of each cluster ([**`merlin6`**](../slurm-configuration.md),[**`gmerlin6`**](../../gmerlin6/slurm-configuration.md) for further information.
* **Partitions:** except when using the *default* partition for each cluster, one needs to specify the partition:
@@ -60,7 +60,7 @@ The following settings are the minimum required for running a job in the Merlin
#SBATCH --partition= # Check each cluster documentation for possible values
```
- Refer to the documentation of each cluster ([**`merlin6`**](../slurm-configuration.md),[**`gmerlin6`**](../../gmerlin6/slurm-configuration.md),[**`merlin5`**](../../merlin5/slurm-configuration.md) for further information.
+ Refer to the documentation of each cluster ([**`merlin6`**](../slurm-configuration.md),[**`gmerlin6`**](../../gmerlin6/slurm-configuration.md) for further information.
* **[Optional] Disabling shared nodes**: by default, nodes are not exclusive. Hence, multiple users can run in the same node. One can request exclusive node usage with the following option:
@@ -74,7 +74,7 @@ The following settings are the minimum required for running a job in the Merlin
#SBATCH --time= # Can not exceed the partition `MaxTime`
```
- Refer to the documentation of each cluster ([**`merlin6`**](../slurm-configuration.md),[**`gmerlin6`**](../../gmerlin6/slurm-configuration.md),[**`merlin5`**](../../merlin5/slurm-configuration.md) for further information about partition `MaxTime` values.
+ Refer to the documentation of each cluster ([**`merlin6`**](../slurm-configuration.md),[**`gmerlin6`**](../../gmerlin6/slurm-configuration.md) for further information about partition `MaxTime` values.
* **Output and error files**: by default, Slurm script will generate standard output (`slurm-%j.out`, where `%j` is the job_id) and error (`slurm-%j.err`, where `%j` is the job_id) files in the directory from where the job was submitted. Users can change default name with the following options:
@@ -92,7 +92,7 @@ The following settings are the minimum required for running a job in the Merlin
#SBATCH --hint=nomultithread # Don't use extra threads with in-core multi-threading.
```
- Refer to the documentation of each cluster ([**`merlin6`**](../slurm-configuration.md),[**`gmerlin6`**](../../gmerlin6/slurm-configuration.md),[**`merlin5`**](../../merlin5/slurm-configuration.md) for further information about node configuration and Hyper-Threading.
+ Refer to the documentation of each cluster ([**`merlin6`**](../slurm-configuration.md),[**`gmerlin6`**](../../gmerlin6/slurm-configuration.md) for further information about node configuration and Hyper-Threading.
Consider that, sometimes, depending on your job requirements, you might need also to setup how many `--ntasks-per-core` or `--cpus-per-task` (even other options) in addition to the `--hint` command. Please, contact us in case of doubts.
!!! tip
diff --git a/docs/merlin6/slurm-general-docs/slurm-basic-commands.md b/docs/merlin6/slurm-general-docs/slurm-basic-commands.md
index 972a6ef..dd5b096 100644
--- a/docs/merlin6/slurm-general-docs/slurm-basic-commands.md
+++ b/docs/merlin6/slurm-general-docs/slurm-basic-commands.md
@@ -33,23 +33,3 @@ sshare -a # to list shares of associations to a cluster
sprio -l # to view the factors that comprise a job's scheduling priority
# add '-u ' for filtering user
```
-
-## Show information for specific cluster
-
-By default, any of the above commands shows information of the local cluster which is **merlin6**.
-
-If you want to see the same information for **merlin5** you have to add the parameter ``--clusters=merlin5``.
-If you want to see both clusters at the same time, add the option ``--federation``.
-
-Examples:
-
-```bash
-sinfo # 'sinfo' local cluster which is 'merlin6'
-sinfo --clusters=merlin5 # 'sinfo' non-local cluster 'merlin5'
-sinfo --federation # 'sinfo' all clusters which are 'merlin5' & 'merlin6'
-squeue # 'squeue' local cluster which is 'merlin6'
-squeue --clusters=merlin5 # 'squeue' non-local cluster 'merlin5'
-squeue --federation # 'squeue' all clusters which are 'merlin5' & 'merlin6'
-```
-
----
diff --git a/mkdocs.yml b/mkdocs.yml
index b9433ba..f715e7e 100644
--- a/mkdocs.yml
+++ b/mkdocs.yml
@@ -11,6 +11,10 @@ edit_uri: _edit/master/docs/
# Copyright
copyright: Copyright © 2025 HPC and Emerging Technologies Group/CSD @ Paul Scherrer Institut
+# CSS
+extra_css:
+ - css/custom-nav.css
+
# Configuration
theme:
name: material
@@ -82,7 +86,7 @@ nav:
- Home: index.md
- News:
- news/index.md
- - Merlin 7:
+ - Merlin 7:
- Quick Start Guide:
- merlin7/01-Quick-Start-Guide/introduction.md
- merlin7/01-Quick-Start-Guide/code-of-conduct.md
@@ -126,11 +130,9 @@ nav:
- Support:
- merlin7/99-support/migration-from-merlin6.md
- Merlin 6:
- - merlin6/introduction.md
- - merlin6/hardware-and-software-description.md
- - merlin6/slurm-configuration.md
+ - Introduction:
+ - merlin6/introduction.md
- Quick Start Guide:
- - merlin6/quick-start-guide/introduction.md
- merlin6/quick-start-guide/code-of-conduct.md
- merlin6/quick-start-guide/requesting-accounts.md
- merlin6/quick-start-guide/requesting-projects.md
@@ -143,7 +145,6 @@ nav:
- merlin6/how-to-use-merlin/storage.md
- merlin6/how-to-use-merlin/transfer-data.md
- merlin6/how-to-use-merlin/archive.md
- - merlin6/how-to-use-merlin/nomachine.md
- merlin6/how-to-use-merlin/ssh-keys.md
- merlin6/how-to-use-merlin/kerberos.md
- merlin6/how-to-use-merlin/using-modules.md
@@ -157,10 +158,6 @@ nav:
- gmerlin6/cluster-introduction.md
- gmerlin6/hardware-and-software-description.md
- gmerlin6/slurm-configuration.md
- - Merlin 5 CPU Cluster:
- - merlin5/cluster-introduction.md
- - merlin5/hardware-and-software-description.md
- - merlin5/slurm-configuration.md
- Jupyterhub:
- merlin6/jupyterhub/jupyterhub.md
- merlin6/jupyterhub/jupyter-examples.md
@@ -181,11 +178,6 @@ nav:
- merlin6/software-support/openmpi.md
- merlin6/software-support/paraview.md
- merlin6/software-support/python.md
- - Announcements:
- - merlin6/98-announcements/downtimes.md
- - merlin6/98-announcements/past-downtimes.md
- - Support:
- - merlin6/99-support/migration-from-merlin5.md
- MeG:
- meg/index.md
- meg/contact.md
@@ -198,3 +190,9 @@ nav:
- support/faq.md
- support/known-problems.md
- support/troubleshooting.md
+
+markdown_extensions:
+ - pymdownx.superfences:
+ custom_fences:
+ - name: mermaid
+ class: mermaid