Clean Merlin6 docs
All checks were successful
Build and deploy documentation / build-and-deploy-docs (push) Successful in 24s
All checks were successful
Build and deploy documentation / build-and-deploy-docs (push) Successful in 24s
This commit is contained in:
37
docs/css/custom-nav.css
Normal file
37
docs/css/custom-nav.css
Normal file
@@ -0,0 +1,37 @@
|
||||
/* Strong high-contrast section headers in MkDocs Material nav */
|
||||
|
||||
/* Top-level section blocks */
|
||||
.md-nav__title {
|
||||
font-weight: 800;
|
||||
font-size: 0.65rem;
|
||||
|
||||
color: #ffffff !important; /* white text */
|
||||
background-color: #586ACB !important; /* strong dark blue */
|
||||
|
||||
padding: 6px 10px;
|
||||
margin-top: 14px;
|
||||
}
|
||||
|
||||
/* For section-like items (common in newer Material versions) */
|
||||
.md-nav--primary .md-nav__item--section > .md-nav__link {
|
||||
font-weight: 800;
|
||||
font-size: 0.65rem;
|
||||
|
||||
color: #ffffff !important; /* white text */
|
||||
background-color: #586ACB !important;
|
||||
|
||||
padding: 6px 10px;
|
||||
margin-top: 10px;
|
||||
}
|
||||
|
||||
/* For nested group headers (covers older/other layouts) */
|
||||
.md-nav--primary .md-nav__item--nested > .md-nav__link {
|
||||
font-weight: 800;
|
||||
font-size: 0.65rem;
|
||||
|
||||
color: #ffffff !important;
|
||||
background-color: #586ACB !important;
|
||||
|
||||
padding: 6px 10px;
|
||||
margin-top: 10px;
|
||||
}
|
||||
@@ -4,23 +4,23 @@
|
||||
|
||||
### Introduction
|
||||
|
||||
Merlin6 is a the official PSI Local HPC cluster for development and
|
||||
mission-critical applications that has been built in 2019. It replaces
|
||||
Merlin6 was a the official PSI Local HPC cluster for development and
|
||||
mission-critical applications that has been built in 2019. It replaced
|
||||
the Merlin5 cluster.
|
||||
|
||||
Merlin6 is designed to be extensible, so is technically possible to add
|
||||
Merlin6 was designed to be extensible, so was technically possible to add
|
||||
more compute nodes and cluster storage without significant increase of
|
||||
the costs of the manpower and the operations.
|
||||
|
||||
Merlin6 is mostly based on **CPU** resources, but also contains a small amount
|
||||
Merlin6 was mostly based on **CPU** resources, but also contained a small amount
|
||||
of **GPU**-based resources which are mostly used by the BIO experiments.
|
||||
|
||||
### Slurm 'gmerlin6'
|
||||
|
||||
THe **GPU nodes** have a dedicated **Slurm** cluster, called **`gmerli6`**.
|
||||
THe **GPU nodes** have a dedicated **Slurm** cluster, called **`gmerlin6`**.
|
||||
|
||||
This cluster contains the same shared storage resources (`/data/user`, `/data/project`, `/shared-scracth`, `/afs`, `/psi/home`)
|
||||
which are present in the other Merlin Slurm clusters (`merlin5`,`merlin6`). The Slurm `gmerlin6` cluster is maintainted
|
||||
which are present in the Merlin6 Slurm CPU cluster. The Slurm `gmerlin6` cluster is maintainted
|
||||
independently to ease access for the users and keep independent user accounting.
|
||||
|
||||
## Merlin6 Architecture
|
||||
@@ -31,8 +31,8 @@ The following image shows the Merlin6 cluster architecture diagram:
|
||||
|
||||

|
||||
|
||||
### Merlin5 + Merlin6 Slurm Cluster Architecture Design
|
||||
### Merlin6 Slurm Cluster Architecture Design
|
||||
|
||||
The following image shows the Slurm architecture design for the Merlin5 & Merlin6 clusters:
|
||||
The following image shows the Slurm architecture design for the Merlin6 clusters:
|
||||
|
||||

|
||||
|
||||
@@ -10,7 +10,7 @@ Hence, due to this, the Merlin6 GPU computing cluster has a non homogeneus solut
|
||||
|
||||
On 2018, for the common good, BIO decided to open the cluster to the Merlin users and make it widely accessible for the PSI scientists.
|
||||
|
||||
The below table summarizes the hardware setup for the Merlin6 GPU computing nodes:
|
||||
The below table summarizes the hardware setup for the old Merlin6 GPU computing nodes:
|
||||
|
||||
<table>
|
||||
<thead>
|
||||
@@ -110,6 +110,10 @@ The below table summarizes the hardware setup for the Merlin6 GPU computing node
|
||||
</tbody>
|
||||
</table>
|
||||
|
||||
### Gwendolen
|
||||
|
||||
Currently only Gwendolen is available on `gmerlin6`.
|
||||
|
||||
### Login Nodes
|
||||
|
||||
The login nodes are part of the **[Merlin6](../merlin6/introduction.md)** HPC cluster,
|
||||
|
||||
@@ -58,7 +58,6 @@ Properties of the directory classes:
|
||||
| ---------------------------------- | ----------------------- | ----------------------- | :-------: | :--------------------------------- |:-------------------------------- | ------ | :----------------------------- |
|
||||
| /psi/home/$username | USR [10GB:11GB] | *Undef* | N/A | Up to x2 when strongly justified. | N/A | yes | Daily snapshots for 1 week |
|
||||
| /data/user/$username | USR [1TB:1.074TB] | USR [1M:1.1M] | 7d | Inmutable. Need a project. | Changeable when justified. | no | Users responsible for backup |
|
||||
| /data/project/bio/$projectname | GRP [1TB:1.074TB] | GRP [1M:1.1M] | 7d | Subject to project requirements. | Subject to project requirements. | no | Project responsible for backup |
|
||||
| /data/project/general/$projectname | GRP [1TB:1.074TB] | GRP [1M:1.1M] | 7d | Subject to project requirements. | Subject to project requirements. | no | Project responsible for backup |
|
||||
| /scratch | *Undef* | *Undef* | N/A | N/A | N/A | no | N/A |
|
||||
| /shared-scratch | USR [512GB:2TB] | USR [2M:2.5M] | 7d | Up to x2 when strongly justified. | Changeable when justified. | no | N/A |
|
||||
@@ -173,7 +172,6 @@ The properties of the available scratch storage spaces are given in the followin
|
||||
|
||||
| Cluster | Service | Scratch | Scratch Mountpoint | Shared Scratch | Shared Scratch Mountpoint | Comments |
|
||||
| ------- | -------------- | ------------ | ------------------ | -------------- | ------------------------- | ------------------------------------ |
|
||||
| merlin5 | computing node | 50GB / SAS | `/scratch` | `N/A` | `N/A` | `merlin-c-[01-64]` |
|
||||
| merlin6 | login node | 100GB / SAS | `/scratch` | 50TB / GPFS | `/shared-scratch` | `merlin-l-0[1,2]` |
|
||||
| merlin6 | computing node | 1.3TB / NVMe | `/scratch` | 50TB / GPFS | `/shared-scratch` | `merlin-c-[001-024,101-124,201-224]` |
|
||||
| merlin6 | login node | 2.0TB / NVMe | `/scratch` | 50TB / GPFS | `/shared-scratch` | `merlin-l-00[1,2]` |
|
||||
|
||||
@@ -1,25 +1,5 @@
|
||||
# Introduction
|
||||
# About Merlin6
|
||||
|
||||
!!! failure "Deprecated"
|
||||
Merlin6 is no longer accessible for new users! Please have a look at
|
||||
[Merlin 7](../merlin7/01-Quick-Start-Guide/introduction.md) instead.
|
||||
Merlin6 is **no longer accessible for new users** and a smaller footprint of it has been kept for some of the stakeholder which owned nodes and can not fully run on the new cluster!
|
||||
|
||||
## Slurm clusters
|
||||
|
||||
Comprised the **`merlin6`** CPU cluster and the
|
||||
[**`gmerlin6`**](../gmerlin6/cluster-introduction.md) GPU cluster. A legacy
|
||||
cluster (formally called [**`merlin5`**](../merlin5/cluster-introduction.md) is
|
||||
kept active with best effort support.
|
||||
|
||||
From July 2019, **`merlin6`** becomes the **default Slurm cluster** and any job
|
||||
submitted from the login node will be submitted to that cluster if not.
|
||||
|
||||
* Users can keep submitting to the old *`merlin5`* computing nodes by using the option ``--cluster=merlin5``.
|
||||
* Users submitting to the **`gmerlin6`** GPU cluster need to specify the option ``--cluster=gmerlin6``.
|
||||
|
||||
### Slurm 'merlin6'
|
||||
|
||||
**CPU nodes** are configured in a **Slurm** cluster, called **`merlin6`**, and
|
||||
this is the ***default Slurm cluster***. Hence, by default, if no Slurm cluster
|
||||
is specified (with the `--cluster` option), this will be the cluster to which
|
||||
the jobs will be sent.
|
||||
Please have a look at [Merlin 7](../merlin7/01-Quick-Start-Guide/introduction.md) instead.
|
||||
|
||||
@@ -46,4 +46,3 @@ The Merlin6 login nodes are the following:
|
||||
| ------------------- | --- | --------- | ------ |:--------:| :-------------------- | ------ | ---------- | :------------------ |
|
||||
| merlin-l-001.psi.ch | yes | yes | 2 x 22 | 2 | Intel Xeon Gold 6152 | 384GB | 1.8TB NVMe | ``/scratch`` |
|
||||
| merlin-l-002.psi.ch | yes | yes | 2 x 22 | 2 | Intel Xeon Gold 6142 | 384GB | 1.8TB NVMe | ``/scratch`` |
|
||||
| merlin-l-01.psi.ch | yes | - | 2 x 16 | 2 | Intel Xeon E5-2697Av4 | 512GB | 100GB SAS | ``/scratch`` |
|
||||
|
||||
@@ -7,7 +7,6 @@ It basically contains the following clusters:
|
||||
|
||||
* The **Merlin6 Slurm CPU cluster**, which is called [**`merlin6`**](#merlin6-cpu-cluster-access).
|
||||
* The **Merlin6 Slurm GPU cluster**, which is called [**`gmerlin6`**](#merlin6-gpu-cluster-access).
|
||||
* The *old Merlin5 Slurm CPU cluster*, which is called [**`merlin5`**](#merlin5-cpu-cluster-access), still supported in a best effort basis.
|
||||
|
||||
## Accessing the Slurm clusters
|
||||
|
||||
@@ -35,15 +34,3 @@ For further information about how to use this cluster, please visit: [**Merlin6
|
||||
The **Merlin6 GPU cluster** (**`gmerlin6`**) is visible from the login nodes. However, to submit jobs to this cluster, one needs to specify the option `--cluster=gmerlin6` when submitting a job or allocation.
|
||||
|
||||
For further information about how to use this cluster, please visit: [**Merlin6 GPU Slurm Cluster documentation**](../../gmerlin6/slurm-configuration.md).
|
||||
|
||||
### Merlin5 CPU cluster access
|
||||
|
||||
The **Merlin5 CPU cluster** (**`merlin5`**) is visible from the login nodes. However, to submit jobs
|
||||
to this cluster, one needs to specify the option `--cluster=merlin5` when submitting a job or allocation.
|
||||
|
||||
Using this cluster is in general not recommended, however this is still
|
||||
available for old users needing extra computational resources or longer jobs.
|
||||
Have in mind that this cluster is only supported in a **best effort basis**,
|
||||
and it contains very old hardware and configurations.
|
||||
|
||||
For further information about how to use this cluster, please visit the [**Merlin5 CPU Slurm Cluster documentation**](../../merlin5/slurm-configuration.md).
|
||||
|
||||
@@ -5,13 +5,6 @@
|
||||
Historically, the local HPC clusters at PSI were named **Merlin**. Over the years,
|
||||
multiple generations of Merlin have been deployed.
|
||||
|
||||
At present, the **Merlin local HPC cluster** contains _two_ generations of it:
|
||||
|
||||
* the old **Merlin5** cluster (`merlin5` Slurm cluster), and
|
||||
* the newest generation **Merlin6**, which is divided in two Slurm clusters:
|
||||
* `merlin6` as the Slurm CPU cluster
|
||||
* `gmerlin6` as the Slurm GPU cluster.
|
||||
|
||||
Access to the different Slurm clusters is possible from the [**Merlin login nodes**](accessing-interactive-nodes.md),
|
||||
which can be accessed through the [SSH protocol](accessing-interactive-nodes.md#ssh-access) or the [NoMachine (NX) service](../how-to-use-merlin/nomachine.md).
|
||||
|
||||
@@ -36,8 +29,8 @@ connected to the central PSI IT infrastructure.
|
||||
#### CPU and GPU Slurm clusters
|
||||
|
||||
The Merlin6 **computing nodes** are mostly based on **CPU** resources. However,
|
||||
it also contains a small amount of **GPU**-based resources, which are mostly used
|
||||
by the BIO Division and by Deep Leaning project.
|
||||
in the past it also contained a small amount of **GPU**-based resources, which were mostly used
|
||||
by the BIO Division and by Deep Leaning projects. Today, only Gwendolen is available on `gmerlin6`.
|
||||
|
||||
These computational resources are split into **two** different **[Slurm](https://slurm.schedmd.com/overview.html)** clusters:
|
||||
|
||||
@@ -45,12 +38,3 @@ These computational resources are split into **two** different **[Slurm](https:/
|
||||
* This is the **default Slurm cluster** configured in the login nodes: any job submitted without the option `--cluster` will be submited to this cluster.
|
||||
* The Merlin6 GPU resources are in a dedicated **[Slurm](https://slurm.schedmd.com/overview.html)** cluster called [**`gmerlin6`**](../../gmerlin6/slurm-configuration.md).
|
||||
* Users submitting to the **`gmerlin6`** GPU cluster need to specify the option ``--cluster=gmerlin6``.
|
||||
|
||||
### Merlin5
|
||||
|
||||
The old Slurm **CPU** _Merlin_ cluster is still active and is maintained in a best effort basis.
|
||||
|
||||
**Merlin5** only contains **computing nodes** resources in a dedicated **[Slurm](https://slurm.schedmd.com/overview.html)** cluster.
|
||||
|
||||
* The Merlin5 CPU cluster is called [**merlin5**](../../merlin5/slurm-configuration.md).
|
||||
|
||||
|
||||
@@ -2,41 +2,26 @@
|
||||
|
||||
## Requesting Access to Merlin6
|
||||
|
||||
Access to Merlin6 is regulated by a PSI user's account being a member of the **`svc-cluster_merlin6`** group. Access to this group will also grant access to older generations of Merlin (`merlin5`).
|
||||
In the past, access to the public Merlin6 cluster was regulated via the `svc-cluster_merlin6` group, which is no longer in use.
|
||||
Merlin6 has become a private cluster, and to request access, **users must now be members of one of the Unix groups authorized to use it**, including Gwendolen.
|
||||
|
||||
Requesting **Merlin6** access *has to be done* with the corresponding **[Request Linux Group Membership](https://psi.service-now.com/psisp?id=psi_new_sc_cat_item&sys_id=84f2c0c81b04f110679febd9bb4bcbb1)** form, available in the [PSI Service Now Service Catalog](https://psi.service-now.com/psisp).
|
||||
Requests for Merlin6 access must be submitted using the [Request Linux Group Membership](https://psi.service-now.com/psisp?id=psi_new_sc_cat_item&sys_id=84f2c0c81b04f110679febd9bb4bcbb1) form, available in the [PSI ServiceNow Service Catalog](https://psi.service-now.com/psisp). Access is granted by requesting membership in a Unix group that is permitted to use the cluster.
|
||||
|
||||

|
||||
|
||||
Mandatory customizable fields are the following:
|
||||
### Mandatory fields
|
||||
|
||||
* **`Order Access for user`**, which defaults to the logged in user. However, requesting access for another user it's also possible.
|
||||
* **`Request membership for group`**, for Merlin6 the **`svc-cluster_merlin6`** must be selected.
|
||||
* **`Justification`**, please add here a short justification why access to Merlin6 is necessary.
|
||||
The following fields must be completed:
|
||||
|
||||
Once submitted, the Merlin responsible will approve the request as soon as possible (within the next few hours on working days). Once the request is approved, *it may take up to 30 minutes to get the account fully configured*.
|
||||
* **Order Access for user**: Defaults to the currently logged-in user. Access may also be requested on behalf of another user.
|
||||
* **Request membership for group**: Select a valid Unix group that has access to Merlin6.
|
||||
* **Justification**: Provide a brief explanation of why access to this group is required.
|
||||
|
||||
## Requesting Access to Merlin5
|
||||
|
||||
Access to Merlin5 is regulated by a PSI user's account being a member of the **`svc-cluster_merlin5`** group. Access to this group does not grant access to newer generations of Merlin (`merlin6`, `gmerlin6`, and future ones).
|
||||
|
||||
Requesting **Merlin5** access *has to be done* with the corresponding **[Request Linux Group Membership](https://psi.service-now.com/psisp?id=psi_new_sc_cat_item&sys_id=84f2c0c81b04f110679febd9bb4bcbb1)** form, available in the [PSI Service Now Service Catalog](https://psi.service-now.com/psisp).
|
||||
|
||||

|
||||
|
||||
Mandatory customizable fields are the following:
|
||||
|
||||
* **`Order Access for user`**, which defaults to the logged in user. However, requesting access for another user it's also possible.
|
||||
* **`Request membership for group`**, for Merlin5 the **`svc-cluster_merlin5`** must be selected.
|
||||
* **`Justification`**, please add here a short justification why access to Merlin5 is necessary.
|
||||
|
||||
Once submitted, the Merlin responsible will approve the request as soon as possible (within the next few hours on working days). Once the request is approved, *it may take up to 30 minutes to get the account fully configured*.
|
||||
Once the request is submitted, the corresponding group administrators will review and approve it as soon as possible (typically within a few working hours). After approval, it may take up to 30 minutes for the account to be fully configured and access to become effective.
|
||||
|
||||
## Further documentation
|
||||
|
||||
Further information it's also available in the Linux Central Documentation:
|
||||
Additional information is available in the Linux Central Documentation:
|
||||
|
||||
* [Unix Group / Group Management for users](https://linux.psi.ch/documentation/services/user-guide/unix_groups.html)
|
||||
* [Unix Group / Group Management for group managers](https://linux.psi.ch/documentation/services/admin-guide/unix_groups.html)
|
||||
|
||||
**Special thanks** to the **Linux Central Team** and **AIT** to make this possible.
|
||||
|
||||
@@ -8,9 +8,6 @@ The following table show default and maximum resources that can be used per node
|
||||
|
||||
| Nodes | Def.#CPUs | Max.#CPUs | #Threads | Max.Mem/CPU | Max.Mem/Node | Max.Swap | Def.#GPUs | Max.#GPUs |
|
||||
|:--------------------:| ---------:| :--------:| :------: | :----------:| :-----------:| :-------:| :-------: | :-------: |
|
||||
| merlin-c-[001-024] | 1 core | 44 cores | 2 | 352000 | 352000 | 10000 | N/A | N/A |
|
||||
| merlin-c-[101-124] | 1 core | 44 cores | 2 | 352000 | 352000 | 10000 | N/A | N/A |
|
||||
| merlin-c-[201-224] | 1 core | 44 cores | 2 | 352000 | 352000 | 10000 | N/A | N/A |
|
||||
| merlin-c-[301-312] | 1 core | 44 cores | 2 | 748800 | 748800 | 10000 | N/A | N/A |
|
||||
| merlin-c-[313-318] | 1 core | 44 cores | 1 | 748800 | 748800 | 10000 | N/A | N/A |
|
||||
| merlin-c-[319-324] | 1 core | 44 cores | 2 | 748800 | 748800 | 10000 | N/A | N/A |
|
||||
@@ -19,7 +16,7 @@ If nothing is specified, by default each core will use up to 8GB of memory. Memo
|
||||
`--mem-per-cpu=<mem_in_MB>` options, and maximum memory allowed is `Max.Mem/Node`.
|
||||
|
||||
In **`merlin6`**, Memory is considered a Consumable Resource, as well as the CPU. Hence, both resources will account when submitting a job,
|
||||
and by default resources can not be oversubscribed. This is a main difference with the old **`merlin5`** cluster, when only CPU were accounted,
|
||||
and by default resources can not be oversubscribed.
|
||||
|
||||
and memory was by default oversubscribed.
|
||||
|
||||
@@ -280,7 +277,5 @@ For understanding the Slurm configuration setup in the cluster, sometimes may be
|
||||
The previous configuration files which can be found in the login nodes,
|
||||
correspond exclusively to the **merlin6** cluster configuration files.
|
||||
|
||||
Configuration files for the old **merlin5** cluster or for the **gmerlin6**
|
||||
cluster must be checked directly on any of the **merlin5** or **gmerlin6**
|
||||
computing nodes (in example, by login in to one of the nodes while a job or an
|
||||
Configuration files for the **gmerlin6** cluster must be checked directly on the **gmerlin6** computing nodes (in example, by login in to one of the nodes while a job or an
|
||||
active allocation is running).
|
||||
|
||||
@@ -63,45 +63,6 @@ Check in the **man** pages (`man sinfo`) for all possible options for this comma
|
||||
gpu up 7-00:00:00 1-infinite no NO all 8 allocated merlin-g-[001-006,008-009]
|
||||
```
|
||||
|
||||
### Slurm commander
|
||||
|
||||
The **[Slurm Commander (scom)](https://github.com/CLIP-HPC/SlurmCommander/)** is a simple but very useful open source text-based user interface for
|
||||
simple and efficient interaction with Slurm. It is developed by the **CLoud Infrastructure Project (CLIP-HPC)** and external contributions. To use it, one can
|
||||
simply run the following command:
|
||||
|
||||
```bash
|
||||
scom # merlin6 cluster
|
||||
SLURM_CLUSTERS=merlin5 scom # merlin5 cluster
|
||||
SLURM_CLUSTERS=gmerlin6 scom # gmerlin6 cluster
|
||||
scom -h # Help and extra options
|
||||
scom -d 14 # Set Job History to 14 days (instead of default 7)
|
||||
```
|
||||
|
||||
With this simple interface, users can interact with their jobs, as well as getting information about past and present jobs:
|
||||
|
||||
* Filtering jobs by substring is possible with the `/` key.
|
||||
* Users can perform multiple actions on their jobs (such like cancelling,
|
||||
holding or requeing a job), SSH to a node with an already running job,
|
||||
or getting extended details and statistics of the job itself.
|
||||
|
||||
Also, users can check the status of the cluster, to get statistics and node usage information as well as getting information about node properties.
|
||||
|
||||
The interface also provides a few job templates for different use cases (i.e. MPI, OpenMP, Hybrid, single core). Users can modify these templates,
|
||||
save it locally to the current directory, and submit the job to the cluster.
|
||||
|
||||
!!! note
|
||||
Currently, `scom` does not provide live updated information for the <span
|
||||
style="color:darkorange;">[Job History]</span> tab. To update Job History
|
||||
information, users have to exit the application with the <span
|
||||
style="color:darkorange;">q</span> key. Other tabs will be updated every 5
|
||||
seconds (default). On the other hand, the <span style="color:darkorange;">[Job
|
||||
History]</span> tab contains only information for the **merlin6** CPU cluster
|
||||
only. Future updates will provide information for other clusters.
|
||||
|
||||
For further information about how to use **scom**, please refer to the **[Slurm Commander Project webpage](https://github.com/CLIP-HPC/SlurmCommander/)**
|
||||
|
||||

|
||||
|
||||
### Job accounting
|
||||
|
||||
Users can check detailed information of jobs (pending, running, completed, failed, etc.) with the `sacct` command.
|
||||
@@ -267,11 +228,3 @@ support:
|
||||
* Nodes monitoring:
|
||||
* [Merlin6 CPU Nodes Overview](https://hpc-monitor02.psi.ch/d/JmvLR8gZz/merlin6-computing-cpu-nodes?orgId=1&refresh=10s)
|
||||
* [Merlin6 GPU Nodes Overview](https://hpc-monitor02.psi.ch/d/gOo1Z10Wk/merlin6-computing-gpu-nodes?orgId=1&refresh=10s)
|
||||
|
||||
### Merlin5 Monitoring Pages
|
||||
|
||||
* Slurm monitoring:
|
||||
* [Merlin5 Slurm Live Status](https://hpc-monitor02.psi.ch/d/o8msZJ0Zz/merlin5-slurm-live-status?orgId=1&refresh=10s)
|
||||
* [Merlin5 Slurm Overview](https://hpc-monitor02.psi.ch/d/eWLEW1AWz/merlin5-slurm-overview?orgId=1&refresh=10s)
|
||||
* Nodes monitoring:
|
||||
* [Merlin5 CPU Nodes Overview](https://hpc-monitor02.psi.ch/d/ejTyWJAWk/merlin5-computing-cpu-nodes?orgId=1&refresh=10s)
|
||||
|
||||
@@ -49,10 +49,10 @@ The following settings are the minimum required for running a job in the Merlin
|
||||
* **Clusters:** For running jobs in the different Slurm clusters, users should to add the following option:
|
||||
|
||||
```bash
|
||||
#SBATCH --clusters=<cluster_name> # Possible values: merlin5, merlin6, gmerlin6
|
||||
#SBATCH --clusters=<cluster_name> # Possible values: merlin6, gmerlin6
|
||||
```
|
||||
|
||||
Refer to the documentation of each cluster ([**`merlin6`**](../slurm-configuration.md),[**`gmerlin6`**](../../gmerlin6/slurm-configuration.md),[**`merlin5`**](../../merlin5/slurm-configuration.md) for further information.
|
||||
Refer to the documentation of each cluster ([**`merlin6`**](../slurm-configuration.md),[**`gmerlin6`**](../../gmerlin6/slurm-configuration.md) for further information.
|
||||
|
||||
* **Partitions:** except when using the *default* partition for each cluster, one needs to specify the partition:
|
||||
|
||||
@@ -60,7 +60,7 @@ The following settings are the minimum required for running a job in the Merlin
|
||||
#SBATCH --partition=<partition_name> # Check each cluster documentation for possible values
|
||||
```
|
||||
|
||||
Refer to the documentation of each cluster ([**`merlin6`**](../slurm-configuration.md),[**`gmerlin6`**](../../gmerlin6/slurm-configuration.md),[**`merlin5`**](../../merlin5/slurm-configuration.md) for further information.
|
||||
Refer to the documentation of each cluster ([**`merlin6`**](../slurm-configuration.md),[**`gmerlin6`**](../../gmerlin6/slurm-configuration.md) for further information.
|
||||
|
||||
* **[Optional] Disabling shared nodes**: by default, nodes are not exclusive. Hence, multiple users can run in the same node. One can request exclusive node usage with the following option:
|
||||
|
||||
@@ -74,7 +74,7 @@ The following settings are the minimum required for running a job in the Merlin
|
||||
#SBATCH --time=<D-HH:MM:SS> # Can not exceed the partition `MaxTime`
|
||||
```
|
||||
|
||||
Refer to the documentation of each cluster ([**`merlin6`**](../slurm-configuration.md),[**`gmerlin6`**](../../gmerlin6/slurm-configuration.md),[**`merlin5`**](../../merlin5/slurm-configuration.md) for further information about partition `MaxTime` values.
|
||||
Refer to the documentation of each cluster ([**`merlin6`**](../slurm-configuration.md),[**`gmerlin6`**](../../gmerlin6/slurm-configuration.md) for further information about partition `MaxTime` values.
|
||||
|
||||
* **Output and error files**: by default, Slurm script will generate standard output (`slurm-%j.out`, where `%j` is the job_id) and error (`slurm-%j.err`, where `%j` is the job_id) files in the directory from where the job was submitted. Users can change default name with the following options:
|
||||
|
||||
@@ -92,7 +92,7 @@ The following settings are the minimum required for running a job in the Merlin
|
||||
#SBATCH --hint=nomultithread # Don't use extra threads with in-core multi-threading.
|
||||
```
|
||||
|
||||
Refer to the documentation of each cluster ([**`merlin6`**](../slurm-configuration.md),[**`gmerlin6`**](../../gmerlin6/slurm-configuration.md),[**`merlin5`**](../../merlin5/slurm-configuration.md) for further information about node configuration and Hyper-Threading.
|
||||
Refer to the documentation of each cluster ([**`merlin6`**](../slurm-configuration.md),[**`gmerlin6`**](../../gmerlin6/slurm-configuration.md) for further information about node configuration and Hyper-Threading.
|
||||
Consider that, sometimes, depending on your job requirements, you might need also to setup how many `--ntasks-per-core` or `--cpus-per-task` (even other options) in addition to the `--hint` command. Please, contact us in case of doubts.
|
||||
|
||||
!!! tip
|
||||
|
||||
@@ -33,23 +33,3 @@ sshare -a # to list shares of associations to a cluster
|
||||
sprio -l # to view the factors that comprise a job's scheduling priority
|
||||
# add '-u <username>' for filtering user
|
||||
```
|
||||
|
||||
## Show information for specific cluster
|
||||
|
||||
By default, any of the above commands shows information of the local cluster which is **merlin6**.
|
||||
|
||||
If you want to see the same information for **merlin5** you have to add the parameter ``--clusters=merlin5``.
|
||||
If you want to see both clusters at the same time, add the option ``--federation``.
|
||||
|
||||
Examples:
|
||||
|
||||
```bash
|
||||
sinfo # 'sinfo' local cluster which is 'merlin6'
|
||||
sinfo --clusters=merlin5 # 'sinfo' non-local cluster 'merlin5'
|
||||
sinfo --federation # 'sinfo' all clusters which are 'merlin5' & 'merlin6'
|
||||
squeue # 'squeue' local cluster which is 'merlin6'
|
||||
squeue --clusters=merlin5 # 'squeue' non-local cluster 'merlin5'
|
||||
squeue --federation # 'squeue' all clusters which are 'merlin5' & 'merlin6'
|
||||
```
|
||||
|
||||
---
|
||||
|
||||
26
mkdocs.yml
26
mkdocs.yml
@@ -11,6 +11,10 @@ edit_uri: _edit/master/docs/
|
||||
# Copyright
|
||||
copyright: Copyright © 2025 HPC and Emerging Technologies Group/CSD @ Paul Scherrer Institut
|
||||
|
||||
# CSS
|
||||
extra_css:
|
||||
- css/custom-nav.css
|
||||
|
||||
# Configuration
|
||||
theme:
|
||||
name: material
|
||||
@@ -126,11 +130,9 @@ nav:
|
||||
- Support:
|
||||
- merlin7/99-support/migration-from-merlin6.md
|
||||
- Merlin 6:
|
||||
- merlin6/introduction.md
|
||||
- merlin6/hardware-and-software-description.md
|
||||
- merlin6/slurm-configuration.md
|
||||
- Introduction:
|
||||
- merlin6/introduction.md
|
||||
- Quick Start Guide:
|
||||
- merlin6/quick-start-guide/introduction.md
|
||||
- merlin6/quick-start-guide/code-of-conduct.md
|
||||
- merlin6/quick-start-guide/requesting-accounts.md
|
||||
- merlin6/quick-start-guide/requesting-projects.md
|
||||
@@ -143,7 +145,6 @@ nav:
|
||||
- merlin6/how-to-use-merlin/storage.md
|
||||
- merlin6/how-to-use-merlin/transfer-data.md
|
||||
- merlin6/how-to-use-merlin/archive.md
|
||||
- merlin6/how-to-use-merlin/nomachine.md
|
||||
- merlin6/how-to-use-merlin/ssh-keys.md
|
||||
- merlin6/how-to-use-merlin/kerberos.md
|
||||
- merlin6/how-to-use-merlin/using-modules.md
|
||||
@@ -157,10 +158,6 @@ nav:
|
||||
- gmerlin6/cluster-introduction.md
|
||||
- gmerlin6/hardware-and-software-description.md
|
||||
- gmerlin6/slurm-configuration.md
|
||||
- Merlin 5 CPU Cluster:
|
||||
- merlin5/cluster-introduction.md
|
||||
- merlin5/hardware-and-software-description.md
|
||||
- merlin5/slurm-configuration.md
|
||||
- Jupyterhub:
|
||||
- merlin6/jupyterhub/jupyterhub.md
|
||||
- merlin6/jupyterhub/jupyter-examples.md
|
||||
@@ -181,11 +178,6 @@ nav:
|
||||
- merlin6/software-support/openmpi.md
|
||||
- merlin6/software-support/paraview.md
|
||||
- merlin6/software-support/python.md
|
||||
- Announcements:
|
||||
- merlin6/98-announcements/downtimes.md
|
||||
- merlin6/98-announcements/past-downtimes.md
|
||||
- Support:
|
||||
- merlin6/99-support/migration-from-merlin5.md
|
||||
- MeG:
|
||||
- meg/index.md
|
||||
- meg/contact.md
|
||||
@@ -198,3 +190,9 @@ nav:
|
||||
- support/faq.md
|
||||
- support/known-problems.md
|
||||
- support/troubleshooting.md
|
||||
|
||||
markdown_extensions:
|
||||
- pymdownx.superfences:
|
||||
custom_fences:
|
||||
- name: mermaid
|
||||
class: mermaid
|
||||
|
||||
Reference in New Issue
Block a user