Clean Merlin6 docs
All checks were successful
Build and deploy documentation / build-and-deploy-docs (push) Successful in 24s

This commit is contained in:
2026-02-10 10:46:02 +01:00
parent 6812bb6bad
commit 17053a363f
14 changed files with 85 additions and 185 deletions

37
docs/css/custom-nav.css Normal file
View File

@@ -0,0 +1,37 @@
/* Strong high-contrast section headers in MkDocs Material nav */
/* Top-level section blocks */
.md-nav__title {
font-weight: 800;
font-size: 0.65rem;
color: #ffffff !important; /* white text */
background-color: #586ACB !important; /* strong dark blue */
padding: 6px 10px;
margin-top: 14px;
}
/* For section-like items (common in newer Material versions) */
.md-nav--primary .md-nav__item--section > .md-nav__link {
font-weight: 800;
font-size: 0.65rem;
color: #ffffff !important; /* white text */
background-color: #586ACB !important;
padding: 6px 10px;
margin-top: 10px;
}
/* For nested group headers (covers older/other layouts) */
.md-nav--primary .md-nav__item--nested > .md-nav__link {
font-weight: 800;
font-size: 0.65rem;
color: #ffffff !important;
background-color: #586ACB !important;
padding: 6px 10px;
margin-top: 10px;
}

View File

@@ -4,23 +4,23 @@
### Introduction
Merlin6 is a the official PSI Local HPC cluster for development and
mission-critical applications that has been built in 2019. It replaces
Merlin6 was a the official PSI Local HPC cluster for development and
mission-critical applications that has been built in 2019. It replaced
the Merlin5 cluster.
Merlin6 is designed to be extensible, so is technically possible to add
Merlin6 was designed to be extensible, so was technically possible to add
more compute nodes and cluster storage without significant increase of
the costs of the manpower and the operations.
Merlin6 is mostly based on **CPU** resources, but also contains a small amount
Merlin6 was mostly based on **CPU** resources, but also contained a small amount
of **GPU**-based resources which are mostly used by the BIO experiments.
### Slurm 'gmerlin6'
THe **GPU nodes** have a dedicated **Slurm** cluster, called **`gmerli6`**.
THe **GPU nodes** have a dedicated **Slurm** cluster, called **`gmerlin6`**.
This cluster contains the same shared storage resources (`/data/user`, `/data/project`, `/shared-scracth`, `/afs`, `/psi/home`)
which are present in the other Merlin Slurm clusters (`merlin5`,`merlin6`). The Slurm `gmerlin6` cluster is maintainted
which are present in the Merlin6 Slurm CPU cluster. The Slurm `gmerlin6` cluster is maintainted
independently to ease access for the users and keep independent user accounting.
## Merlin6 Architecture
@@ -31,8 +31,8 @@ The following image shows the Merlin6 cluster architecture diagram:
![Merlin6 Architecture Diagram](../images/merlinschema3.png)
### Merlin5 + Merlin6 Slurm Cluster Architecture Design
### Merlin6 Slurm Cluster Architecture Design
The following image shows the Slurm architecture design for the Merlin5 & Merlin6 clusters:
The following image shows the Slurm architecture design for the Merlin6 clusters:
![Merlin6 Slurm Architecture Design](../images/merlin-slurm-architecture.png)

View File

@@ -10,7 +10,7 @@ Hence, due to this, the Merlin6 GPU computing cluster has a non homogeneus solut
On 2018, for the common good, BIO decided to open the cluster to the Merlin users and make it widely accessible for the PSI scientists.
The below table summarizes the hardware setup for the Merlin6 GPU computing nodes:
The below table summarizes the hardware setup for the old Merlin6 GPU computing nodes:
<table>
<thead>
@@ -110,6 +110,10 @@ The below table summarizes the hardware setup for the Merlin6 GPU computing node
</tbody>
</table>
### Gwendolen
Currently only Gwendolen is available on `gmerlin6`.
### Login Nodes
The login nodes are part of the **[Merlin6](../merlin6/introduction.md)** HPC cluster,

View File

@@ -58,7 +58,6 @@ Properties of the directory classes:
| ---------------------------------- | ----------------------- | ----------------------- | :-------: | :--------------------------------- |:-------------------------------- | ------ | :----------------------------- |
| /psi/home/$username | USR [10GB:11GB] | *Undef* | N/A | Up to x2 when strongly justified. | N/A | yes | Daily snapshots for 1 week |
| /data/user/$username | USR [1TB:1.074TB] | USR [1M:1.1M] | 7d | Inmutable. Need a project. | Changeable when justified. | no | Users responsible for backup |
| /data/project/bio/$projectname | GRP [1TB:1.074TB] | GRP [1M:1.1M] | 7d | Subject to project requirements. | Subject to project requirements. | no | Project responsible for backup |
| /data/project/general/$projectname | GRP [1TB:1.074TB] | GRP [1M:1.1M] | 7d | Subject to project requirements. | Subject to project requirements. | no | Project responsible for backup |
| /scratch | *Undef* | *Undef* | N/A | N/A | N/A | no | N/A |
| /shared-scratch | USR [512GB:2TB] | USR [2M:2.5M] | 7d | Up to x2 when strongly justified. | Changeable when justified. | no | N/A |
@@ -173,7 +172,6 @@ The properties of the available scratch storage spaces are given in the followin
| Cluster | Service | Scratch | Scratch Mountpoint | Shared Scratch | Shared Scratch Mountpoint | Comments |
| ------- | -------------- | ------------ | ------------------ | -------------- | ------------------------- | ------------------------------------ |
| merlin5 | computing node | 50GB / SAS | `/scratch` | `N/A` | `N/A` | `merlin-c-[01-64]` |
| merlin6 | login node | 100GB / SAS | `/scratch` | 50TB / GPFS | `/shared-scratch` | `merlin-l-0[1,2]` |
| merlin6 | computing node | 1.3TB / NVMe | `/scratch` | 50TB / GPFS | `/shared-scratch` | `merlin-c-[001-024,101-124,201-224]` |
| merlin6 | login node | 2.0TB / NVMe | `/scratch` | 50TB / GPFS | `/shared-scratch` | `merlin-l-00[1,2]` |

View File

@@ -1,25 +1,5 @@
# Introduction
# About Merlin6
!!! failure "Deprecated"
Merlin6 is no longer accessible for new users! Please have a look at
[Merlin 7](../merlin7/01-Quick-Start-Guide/introduction.md) instead.
Merlin6 is **no longer accessible for new users** and a smaller footprint of it has been kept for some of the stakeholder which owned nodes and can not fully run on the new cluster!
## Slurm clusters
Comprised the **`merlin6`** CPU cluster and the
[**`gmerlin6`**](../gmerlin6/cluster-introduction.md) GPU cluster. A legacy
cluster (formally called [**`merlin5`**](../merlin5/cluster-introduction.md) is
kept active with best effort support.
From July 2019, **`merlin6`** becomes the **default Slurm cluster** and any job
submitted from the login node will be submitted to that cluster if not.
* Users can keep submitting to the old *`merlin5`* computing nodes by using the option ``--cluster=merlin5``.
* Users submitting to the **`gmerlin6`** GPU cluster need to specify the option ``--cluster=gmerlin6``.
### Slurm 'merlin6'
**CPU nodes** are configured in a **Slurm** cluster, called **`merlin6`**, and
this is the ***default Slurm cluster***. Hence, by default, if no Slurm cluster
is specified (with the `--cluster` option), this will be the cluster to which
the jobs will be sent.
Please have a look at [Merlin 7](../merlin7/01-Quick-Start-Guide/introduction.md) instead.

View File

@@ -46,4 +46,3 @@ The Merlin6 login nodes are the following:
| ------------------- | --- | --------- | ------ |:--------:| :-------------------- | ------ | ---------- | :------------------ |
| merlin-l-001.psi.ch | yes | yes | 2 x 22 | 2 | Intel Xeon Gold 6152 | 384GB | 1.8TB NVMe | ``/scratch`` |
| merlin-l-002.psi.ch | yes | yes | 2 x 22 | 2 | Intel Xeon Gold 6142 | 384GB | 1.8TB NVMe | ``/scratch`` |
| merlin-l-01.psi.ch | yes | - | 2 x 16 | 2 | Intel Xeon E5-2697Av4 | 512GB | 100GB SAS | ``/scratch`` |

View File

@@ -7,7 +7,6 @@ It basically contains the following clusters:
* The **Merlin6 Slurm CPU cluster**, which is called [**`merlin6`**](#merlin6-cpu-cluster-access).
* The **Merlin6 Slurm GPU cluster**, which is called [**`gmerlin6`**](#merlin6-gpu-cluster-access).
* The *old Merlin5 Slurm CPU cluster*, which is called [**`merlin5`**](#merlin5-cpu-cluster-access), still supported in a best effort basis.
## Accessing the Slurm clusters
@@ -35,15 +34,3 @@ For further information about how to use this cluster, please visit: [**Merlin6
The **Merlin6 GPU cluster** (**`gmerlin6`**) is visible from the login nodes. However, to submit jobs to this cluster, one needs to specify the option `--cluster=gmerlin6` when submitting a job or allocation.
For further information about how to use this cluster, please visit: [**Merlin6 GPU Slurm Cluster documentation**](../../gmerlin6/slurm-configuration.md).
### Merlin5 CPU cluster access
The **Merlin5 CPU cluster** (**`merlin5`**) is visible from the login nodes. However, to submit jobs
to this cluster, one needs to specify the option `--cluster=merlin5` when submitting a job or allocation.
Using this cluster is in general not recommended, however this is still
available for old users needing extra computational resources or longer jobs.
Have in mind that this cluster is only supported in a **best effort basis**,
and it contains very old hardware and configurations.
For further information about how to use this cluster, please visit the [**Merlin5 CPU Slurm Cluster documentation**](../../merlin5/slurm-configuration.md).

View File

@@ -5,13 +5,6 @@
Historically, the local HPC clusters at PSI were named **Merlin**. Over the years,
multiple generations of Merlin have been deployed.
At present, the **Merlin local HPC cluster** contains _two_ generations of it:
* the old **Merlin5** cluster (`merlin5` Slurm cluster), and
* the newest generation **Merlin6**, which is divided in two Slurm clusters:
* `merlin6` as the Slurm CPU cluster
* `gmerlin6` as the Slurm GPU cluster.
Access to the different Slurm clusters is possible from the [**Merlin login nodes**](accessing-interactive-nodes.md),
which can be accessed through the [SSH protocol](accessing-interactive-nodes.md#ssh-access) or the [NoMachine (NX) service](../how-to-use-merlin/nomachine.md).
@@ -36,8 +29,8 @@ connected to the central PSI IT infrastructure.
#### CPU and GPU Slurm clusters
The Merlin6 **computing nodes** are mostly based on **CPU** resources. However,
it also contains a small amount of **GPU**-based resources, which are mostly used
by the BIO Division and by Deep Leaning project.
in the past it also contained a small amount of **GPU**-based resources, which were mostly used
by the BIO Division and by Deep Leaning projects. Today, only Gwendolen is available on `gmerlin6`.
These computational resources are split into **two** different **[Slurm](https://slurm.schedmd.com/overview.html)** clusters:
@@ -45,12 +38,3 @@ These computational resources are split into **two** different **[Slurm](https:/
* This is the **default Slurm cluster** configured in the login nodes: any job submitted without the option `--cluster` will be submited to this cluster.
* The Merlin6 GPU resources are in a dedicated **[Slurm](https://slurm.schedmd.com/overview.html)** cluster called [**`gmerlin6`**](../../gmerlin6/slurm-configuration.md).
* Users submitting to the **`gmerlin6`** GPU cluster need to specify the option ``--cluster=gmerlin6``.
### Merlin5
The old Slurm **CPU** _Merlin_ cluster is still active and is maintained in a best effort basis.
**Merlin5** only contains **computing nodes** resources in a dedicated **[Slurm](https://slurm.schedmd.com/overview.html)** cluster.
* The Merlin5 CPU cluster is called [**merlin5**](../../merlin5/slurm-configuration.md).

View File

@@ -2,41 +2,26 @@
## Requesting Access to Merlin6
Access to Merlin6 is regulated by a PSI user's account being a member of the **`svc-cluster_merlin6`** group. Access to this group will also grant access to older generations of Merlin (`merlin5`).
In the past, access to the public Merlin6 cluster was regulated via the `svc-cluster_merlin6` group, which is no longer in use.
Merlin6 has become a private cluster, and to request access, **users must now be members of one of the Unix groups authorized to use it**, including Gwendolen.
Requesting **Merlin6** access *has to be done* with the corresponding **[Request Linux Group Membership](https://psi.service-now.com/psisp?id=psi_new_sc_cat_item&sys_id=84f2c0c81b04f110679febd9bb4bcbb1)** form, available in the [PSI Service Now Service Catalog](https://psi.service-now.com/psisp).
Requests for Merlin6 access must be submitted using the [Request Linux Group Membership](https://psi.service-now.com/psisp?id=psi_new_sc_cat_item&sys_id=84f2c0c81b04f110679febd9bb4bcbb1) form, available in the [PSI ServiceNow Service Catalog](https://psi.service-now.com/psisp). Access is granted by requesting membership in a Unix group that is permitted to use the cluster.
![Example: Requesting access to Merlin6](../../images/Access/01-request-merlin6-membership.png)
Mandatory customizable fields are the following:
### Mandatory fields
* **`Order Access for user`**, which defaults to the logged in user. However, requesting access for another user it's also possible.
* **`Request membership for group`**, for Merlin6 the **`svc-cluster_merlin6`** must be selected.
* **`Justification`**, please add here a short justification why access to Merlin6 is necessary.
The following fields must be completed:
Once submitted, the Merlin responsible will approve the request as soon as possible (within the next few hours on working days). Once the request is approved, *it may take up to 30 minutes to get the account fully configured*.
* **Order Access for user**: Defaults to the currently logged-in user. Access may also be requested on behalf of another user.
* **Request membership for group**: Select a valid Unix group that has access to Merlin6.
* **Justification**: Provide a brief explanation of why access to this group is required.
## Requesting Access to Merlin5
Access to Merlin5 is regulated by a PSI user's account being a member of the **`svc-cluster_merlin5`** group. Access to this group does not grant access to newer generations of Merlin (`merlin6`, `gmerlin6`, and future ones).
Requesting **Merlin5** access *has to be done* with the corresponding **[Request Linux Group Membership](https://psi.service-now.com/psisp?id=psi_new_sc_cat_item&sys_id=84f2c0c81b04f110679febd9bb4bcbb1)** form, available in the [PSI Service Now Service Catalog](https://psi.service-now.com/psisp).
![Example: Requesting access to Merlin5](../../images/Access/01-request-merlin5-membership.png)
Mandatory customizable fields are the following:
* **`Order Access for user`**, which defaults to the logged in user. However, requesting access for another user it's also possible.
* **`Request membership for group`**, for Merlin5 the **`svc-cluster_merlin5`** must be selected.
* **`Justification`**, please add here a short justification why access to Merlin5 is necessary.
Once submitted, the Merlin responsible will approve the request as soon as possible (within the next few hours on working days). Once the request is approved, *it may take up to 30 minutes to get the account fully configured*.
Once the request is submitted, the corresponding group administrators will review and approve it as soon as possible (typically within a few working hours). After approval, it may take up to 30 minutes for the account to be fully configured and access to become effective.
## Further documentation
Further information it's also available in the Linux Central Documentation:
Additional information is available in the Linux Central Documentation:
* [Unix Group / Group Management for users](https://linux.psi.ch/documentation/services/user-guide/unix_groups.html)
* [Unix Group / Group Management for group managers](https://linux.psi.ch/documentation/services/admin-guide/unix_groups.html)
**Special thanks** to the **Linux Central Team** and **AIT** to make this possible.

View File

@@ -8,9 +8,6 @@ The following table show default and maximum resources that can be used per node
| Nodes | Def.#CPUs | Max.#CPUs | #Threads | Max.Mem/CPU | Max.Mem/Node | Max.Swap | Def.#GPUs | Max.#GPUs |
|:--------------------:| ---------:| :--------:| :------: | :----------:| :-----------:| :-------:| :-------: | :-------: |
| merlin-c-[001-024] | 1 core | 44 cores | 2 | 352000 | 352000 | 10000 | N/A | N/A |
| merlin-c-[101-124] | 1 core | 44 cores | 2 | 352000 | 352000 | 10000 | N/A | N/A |
| merlin-c-[201-224] | 1 core | 44 cores | 2 | 352000 | 352000 | 10000 | N/A | N/A |
| merlin-c-[301-312] | 1 core | 44 cores | 2 | 748800 | 748800 | 10000 | N/A | N/A |
| merlin-c-[313-318] | 1 core | 44 cores | 1 | 748800 | 748800 | 10000 | N/A | N/A |
| merlin-c-[319-324] | 1 core | 44 cores | 2 | 748800 | 748800 | 10000 | N/A | N/A |
@@ -19,7 +16,7 @@ If nothing is specified, by default each core will use up to 8GB of memory. Memo
`--mem-per-cpu=<mem_in_MB>` options, and maximum memory allowed is `Max.Mem/Node`.
In **`merlin6`**, Memory is considered a Consumable Resource, as well as the CPU. Hence, both resources will account when submitting a job,
and by default resources can not be oversubscribed. This is a main difference with the old **`merlin5`** cluster, when only CPU were accounted,
and by default resources can not be oversubscribed.
and memory was by default oversubscribed.
@@ -280,7 +277,5 @@ For understanding the Slurm configuration setup in the cluster, sometimes may be
The previous configuration files which can be found in the login nodes,
correspond exclusively to the **merlin6** cluster configuration files.
Configuration files for the old **merlin5** cluster or for the **gmerlin6**
cluster must be checked directly on any of the **merlin5** or **gmerlin6**
computing nodes (in example, by login in to one of the nodes while a job or an
Configuration files for the **gmerlin6** cluster must be checked directly on the **gmerlin6** computing nodes (in example, by login in to one of the nodes while a job or an
active allocation is running).

View File

@@ -63,45 +63,6 @@ Check in the **man** pages (`man sinfo`) for all possible options for this comma
gpu up 7-00:00:00 1-infinite no NO all 8 allocated merlin-g-[001-006,008-009]
```
### Slurm commander
The **[Slurm Commander (scom)](https://github.com/CLIP-HPC/SlurmCommander/)** is a simple but very useful open source text-based user interface for
simple and efficient interaction with Slurm. It is developed by the **CLoud Infrastructure Project (CLIP-HPC)** and external contributions. To use it, one can
simply run the following command:
```bash
scom # merlin6 cluster
SLURM_CLUSTERS=merlin5 scom # merlin5 cluster
SLURM_CLUSTERS=gmerlin6 scom # gmerlin6 cluster
scom -h # Help and extra options
scom -d 14 # Set Job History to 14 days (instead of default 7)
```
With this simple interface, users can interact with their jobs, as well as getting information about past and present jobs:
* Filtering jobs by substring is possible with the `/` key.
* Users can perform multiple actions on their jobs (such like cancelling,
holding or requeing a job), SSH to a node with an already running job,
or getting extended details and statistics of the job itself.
Also, users can check the status of the cluster, to get statistics and node usage information as well as getting information about node properties.
The interface also provides a few job templates for different use cases (i.e. MPI, OpenMP, Hybrid, single core). Users can modify these templates,
save it locally to the current directory, and submit the job to the cluster.
!!! note
Currently, `scom` does not provide live updated information for the <span
style="color:darkorange;">[Job History]</span> tab. To update Job History
information, users have to exit the application with the <span
style="color:darkorange;">q</span> key. Other tabs will be updated every 5
seconds (default). On the other hand, the <span style="color:darkorange;">[Job
History]</span> tab contains only information for the **merlin6** CPU cluster
only. Future updates will provide information for other clusters.
For further information about how to use **scom**, please refer to the **[Slurm Commander Project webpage](https://github.com/CLIP-HPC/SlurmCommander/)**
!['scom' text-based user interface](../../images/slurm/scom.gif)
### Job accounting
Users can check detailed information of jobs (pending, running, completed, failed, etc.) with the `sacct` command.
@@ -267,11 +228,3 @@ support:
* Nodes monitoring:
* [Merlin6 CPU Nodes Overview](https://hpc-monitor02.psi.ch/d/JmvLR8gZz/merlin6-computing-cpu-nodes?orgId=1&refresh=10s)
* [Merlin6 GPU Nodes Overview](https://hpc-monitor02.psi.ch/d/gOo1Z10Wk/merlin6-computing-gpu-nodes?orgId=1&refresh=10s)
### Merlin5 Monitoring Pages
* Slurm monitoring:
* [Merlin5 Slurm Live Status](https://hpc-monitor02.psi.ch/d/o8msZJ0Zz/merlin5-slurm-live-status?orgId=1&refresh=10s)
* [Merlin5 Slurm Overview](https://hpc-monitor02.psi.ch/d/eWLEW1AWz/merlin5-slurm-overview?orgId=1&refresh=10s)
* Nodes monitoring:
* [Merlin5 CPU Nodes Overview](https://hpc-monitor02.psi.ch/d/ejTyWJAWk/merlin5-computing-cpu-nodes?orgId=1&refresh=10s)

View File

@@ -49,10 +49,10 @@ The following settings are the minimum required for running a job in the Merlin
* **Clusters:** For running jobs in the different Slurm clusters, users should to add the following option:
```bash
#SBATCH --clusters=<cluster_name> # Possible values: merlin5, merlin6, gmerlin6
#SBATCH --clusters=<cluster_name> # Possible values: merlin6, gmerlin6
```
Refer to the documentation of each cluster ([**`merlin6`**](../slurm-configuration.md),[**`gmerlin6`**](../../gmerlin6/slurm-configuration.md),[**`merlin5`**](../../merlin5/slurm-configuration.md) for further information.
Refer to the documentation of each cluster ([**`merlin6`**](../slurm-configuration.md),[**`gmerlin6`**](../../gmerlin6/slurm-configuration.md) for further information.
* **Partitions:** except when using the *default* partition for each cluster, one needs to specify the partition:
@@ -60,7 +60,7 @@ The following settings are the minimum required for running a job in the Merlin
#SBATCH --partition=<partition_name> # Check each cluster documentation for possible values
```
Refer to the documentation of each cluster ([**`merlin6`**](../slurm-configuration.md),[**`gmerlin6`**](../../gmerlin6/slurm-configuration.md),[**`merlin5`**](../../merlin5/slurm-configuration.md) for further information.
Refer to the documentation of each cluster ([**`merlin6`**](../slurm-configuration.md),[**`gmerlin6`**](../../gmerlin6/slurm-configuration.md) for further information.
* **[Optional] Disabling shared nodes**: by default, nodes are not exclusive. Hence, multiple users can run in the same node. One can request exclusive node usage with the following option:
@@ -74,7 +74,7 @@ The following settings are the minimum required for running a job in the Merlin
#SBATCH --time=<D-HH:MM:SS> # Can not exceed the partition `MaxTime`
```
Refer to the documentation of each cluster ([**`merlin6`**](../slurm-configuration.md),[**`gmerlin6`**](../../gmerlin6/slurm-configuration.md),[**`merlin5`**](../../merlin5/slurm-configuration.md) for further information about partition `MaxTime` values.
Refer to the documentation of each cluster ([**`merlin6`**](../slurm-configuration.md),[**`gmerlin6`**](../../gmerlin6/slurm-configuration.md) for further information about partition `MaxTime` values.
* **Output and error files**: by default, Slurm script will generate standard output (`slurm-%j.out`, where `%j` is the job_id) and error (`slurm-%j.err`, where `%j` is the job_id) files in the directory from where the job was submitted. Users can change default name with the following options:
@@ -92,7 +92,7 @@ The following settings are the minimum required for running a job in the Merlin
#SBATCH --hint=nomultithread # Don't use extra threads with in-core multi-threading.
```
Refer to the documentation of each cluster ([**`merlin6`**](../slurm-configuration.md),[**`gmerlin6`**](../../gmerlin6/slurm-configuration.md),[**`merlin5`**](../../merlin5/slurm-configuration.md) for further information about node configuration and Hyper-Threading.
Refer to the documentation of each cluster ([**`merlin6`**](../slurm-configuration.md),[**`gmerlin6`**](../../gmerlin6/slurm-configuration.md) for further information about node configuration and Hyper-Threading.
Consider that, sometimes, depending on your job requirements, you might need also to setup how many `--ntasks-per-core` or `--cpus-per-task` (even other options) in addition to the `--hint` command. Please, contact us in case of doubts.
!!! tip

View File

@@ -33,23 +33,3 @@ sshare -a # to list shares of associations to a cluster
sprio -l # to view the factors that comprise a job's scheduling priority
# add '-u <username>' for filtering user
```
## Show information for specific cluster
By default, any of the above commands shows information of the local cluster which is **merlin6**.
If you want to see the same information for **merlin5** you have to add the parameter ``--clusters=merlin5``.
If you want to see both clusters at the same time, add the option ``--federation``.
Examples:
```bash
sinfo # 'sinfo' local cluster which is 'merlin6'
sinfo --clusters=merlin5 # 'sinfo' non-local cluster 'merlin5'
sinfo --federation # 'sinfo' all clusters which are 'merlin5' & 'merlin6'
squeue # 'squeue' local cluster which is 'merlin6'
squeue --clusters=merlin5 # 'squeue' non-local cluster 'merlin5'
squeue --federation # 'squeue' all clusters which are 'merlin5' & 'merlin6'
```
---

View File

@@ -11,6 +11,10 @@ edit_uri: _edit/master/docs/
# Copyright
copyright: Copyright &copy; 2025 HPC and Emerging Technologies Group/CSD @ Paul Scherrer Institut
# CSS
extra_css:
- css/custom-nav.css
# Configuration
theme:
name: material
@@ -126,11 +130,9 @@ nav:
- Support:
- merlin7/99-support/migration-from-merlin6.md
- Merlin 6:
- merlin6/introduction.md
- merlin6/hardware-and-software-description.md
- merlin6/slurm-configuration.md
- Introduction:
- merlin6/introduction.md
- Quick Start Guide:
- merlin6/quick-start-guide/introduction.md
- merlin6/quick-start-guide/code-of-conduct.md
- merlin6/quick-start-guide/requesting-accounts.md
- merlin6/quick-start-guide/requesting-projects.md
@@ -143,7 +145,6 @@ nav:
- merlin6/how-to-use-merlin/storage.md
- merlin6/how-to-use-merlin/transfer-data.md
- merlin6/how-to-use-merlin/archive.md
- merlin6/how-to-use-merlin/nomachine.md
- merlin6/how-to-use-merlin/ssh-keys.md
- merlin6/how-to-use-merlin/kerberos.md
- merlin6/how-to-use-merlin/using-modules.md
@@ -157,10 +158,6 @@ nav:
- gmerlin6/cluster-introduction.md
- gmerlin6/hardware-and-software-description.md
- gmerlin6/slurm-configuration.md
- Merlin 5 CPU Cluster:
- merlin5/cluster-introduction.md
- merlin5/hardware-and-software-description.md
- merlin5/slurm-configuration.md
- Jupyterhub:
- merlin6/jupyterhub/jupyterhub.md
- merlin6/jupyterhub/jupyter-examples.md
@@ -181,11 +178,6 @@ nav:
- merlin6/software-support/openmpi.md
- merlin6/software-support/paraview.md
- merlin6/software-support/python.md
- Announcements:
- merlin6/98-announcements/downtimes.md
- merlin6/98-announcements/past-downtimes.md
- Support:
- merlin6/99-support/migration-from-merlin5.md
- MeG:
- meg/index.md
- meg/contact.md
@@ -198,3 +190,9 @@ nav:
- support/faq.md
- support/known-problems.md
- support/troubleshooting.md
markdown_extensions:
- pymdownx.superfences:
custom_fences:
- name: mermaid
class: mermaid