Updated interactive-jobs

This commit is contained in:
2019-11-06 08:39:15 +01:00
parent a88253ffb3
commit 19c50a3813

View File

@ -56,13 +56,12 @@ users are able to execute interactive command(s). Once finished (``exit`` or ``C
the allocation is released. **``salloc``** is a blocking command, it is, command will be blocked the allocation is released. **``salloc``** is a blocking command, it is, command will be blocked
until the requested resources are allocated. until the requested resources are allocated.
When running **``salloc``**, once the resources get allocated, *by default* the user will get When running **``salloc``**, once the resources are allocated, *by default* the user will get
a ***new shell on one of the allocated resources*** (if a user has requested few nodes, it will a ***new shell on one of the allocated resources*** (if a user has requested few nodes, it will
prompt a new shell in the first allocated node). This is thanks to the default command prompt a new shell on the first allocated node). This is thanks to the default ``srun`` command
``srun -n1 -N1 --mem-per-cpu=0 --gres=gpu:0 --pty --preserve-env --mpi=none $SHELL`` which will run ``srun -n1 -N1 --mem-per-cpu=0 --gres=gpu:0 --pty --preserve-env --mpi=none $SHELL`` which will run
in the background (users do not need to specify any ``srun`` command). However, this behaviour can in the background (users do not need to specify it). However, this behaviour can
be changed by running a different command after the **``salloc``** command, and someone can replace be changed by running a different command after the **``salloc``** command. In example:
the default ``srun`` command. In example:
```bash ```bash
# Typical 'salloc' call # Typical 'salloc' call
@ -71,7 +70,7 @@ the default ``srun`` command. In example:
salloc --clusters=merlin6 -N 2 -n 2 salloc --clusters=merlin6 -N 2 -n 2
# Custom 'salloc' call # Custom 'salloc' call
# - $SHELL will open a shell locally on the machine running 'salloc' (login node shell) # - $SHELL will open a local shell on the login node from where ``salloc`` is running
salloc --clusters=merlin6 -N 2 -n 2 $SHELL salloc --clusters=merlin6 -N 2 -n 2 $SHELL
``` ```