6.8 KiB
title, keywords, last_updated, summary, sidebar, permalink
title | keywords | last_updated | summary | sidebar | permalink |
---|---|---|---|---|---|
ANSYS / Fluent | software, ansys, fluent, slurm, interactive, rsm, batch job | 07 September 2022 | This document describes how to run ANSYS/Fluent in the Merlin6 cluster | merlin6_sidebar | /merlin6/ansys-fluent.html |
This document describes the different ways for running ANSYS/Fluent
ANSYS/Fluent
Is always recommended to check which parameters are available in Fluent and adapt the below example according to your needs.
For that, run fluent -help
for getting a list of options. However, as when running Fluent one must specify one of the
following flags:
- 2d: This is a 2D solver with single point precision.
- 3d: This is a 3D solver with single point precision.
- 2dpp: This is a 2D solver with double point precision.
- 3dpp: This is a 3D solver with double point precision.
Running Fluent jobs
PModules
Is strongly recommended the use of the latest ANSYS software available in PModules.
module use unstable
module load Pmodules/1.1.6
module use overlay_merlin
module load ANSYS/2022R1
Interactive: RSM from remote PSI Workstations
Is possible to run Fluent through RSM from remote PSI (Linux or Windows) Workstation having a local installation of ANSYS Fluent and RSM client. For that, please refer to the [ANSYS RSM](/merlin6/ansys-rsm.html) in the Merlin documentation for further information of how to setup a RSM client for submitting jobs to Merlin.
Non-interactive: sbatch
Running jobs with sbatch
is always the recommended method. This makes the use of the resources more efficient.
For running it as a job, one needs to run in no graphical mode (-g
option).
Serial example
This example shows a very basic serial job.
#!/bin/bash
#SBATCH --job-name=Fluent # Job Name
#SBATCH --partition=hourly # Using 'daily' will grant higher priority than 'general'
#SBATCH --time=0-01:00:00 # Time needed for running the job. Must match with 'partition' limits.
#SBATCH --cpus-per-task=1 # Double if hyperthreading enabled
#SBATCH --hint=nomultithread # Disable Hyperthreading
#SBATCH --error=slurm-%j.err # Define your error file
module use unstable
module load ANSYS/2020R1-1
# [Optional:BEGIN] Specify your license server if this is not 'lic-ansys.psi.ch'
LICENSE_SERVER=<your_license_server>
export ANSYSLMD_LICENSE_FILE=1055@$LICENSE_SERVER
export ANSYSLI_SERVERS=2325@$LICENSE_SERVER
# [Optional:END]
JOURNAL_FILE=/data/user/caubet_m/Fluent/myjournal.in
fluent 3ddp -g -i ${JOURNAL_FILE}
One can enable hypertheading by defining --hint=multithread
, --cpus-per-task=2
and --ntasks-per-core=2
.
However, this is in general not recommended, unless one can ensure that can be beneficial.
MPI-based example
An example for running Fluent using a Slurm batch script is the following:
#!/bin/bash
#SBATCH --job-name=Fluent # Job Name
#SBATCH --partition=hourly # Using 'daily' will grant higher priority than 'general'
#SBATCH --time=0-01:00:00 # Time needed for running the job. Must match with 'partition' limits.
#SBATCH --nodes=1 # Number of nodes
#SBATCH --ntasks=44 # Number of tasks
#SBATCH --cpus-per-task=1 # Double if hyperthreading enabled
#SBATCH --ntasks-per-core=1 # Run one task per core
#SBATCH --hint=nomultithread # Disable Hyperthreading
#SBATCH --error=slurm-%j.err # Define a file for standard error messages
##SBATCH --exclusive # Uncomment if you want exclusive usage of the nodes
module use unstable
module load ANSYS/2020R1-1
# [Optional:BEGIN] Specify your license server if this is not 'lic-ansys.psi.ch'
LICENSE_SERVER=<your_license_server>
export ANSYSLMD_LICENSE_FILE=1055@$LICENSE_SERVER
export ANSYSLI_SERVERS=2325@$LICENSE_SERVER
# [Optional:END]
JOURNAL_FILE=/data/user/caubet_m/Fluent/myjournal.in
fluent 3ddp -g -t ${SLURM_NTASKS} -i ${JOURNAL_FILE}
In the above example, one can increase the number of nodes and/or ntasks if needed. One can remove
--nodes
for running on multiple nodes, but may lead to communication overhead. In general, no
hyperthreading is recommended for MPI based jobs. Also, one can combine it with --exclusive
when necessary.
Interactive: salloc
Running Fluent interactively is strongly not recommended and one should whenever possible use sbatch
.
However, sometimes interactive runs are needed. For jobs requiring only few CPUs (in example, 2 CPUs) and for a short period of time, one can use the login nodes.
Otherwise, one must use the Slurm batch system using allocations:
- For short jobs requiring more CPUs, one can use the Merlin shortest partitions (
hourly
). - For longer jobs, one can use longer partitions, however, interactive access is not always possible (depending on the usage of the cluster).
Please refer to the documentation Running Interactive Jobs for firther information about different ways for running interactive jobs in the Merlin6 cluster.
Requirements
SSH Keys
Running Fluent interactively requires the use of SSH Keys. This is the way of communication between the GUI and the different nodes. For doing that, one must have
a passphrase protected SSH Key. If the user does not have SSH Keys yet (simply run ls $HOME/.ssh/
to check whether id_rsa
files exist or not). For
deploying SSH Keys for running Fluent interactively, one should follow this documentation: Configuring SSH Keys
List of hosts
For running Fluent using Slurm computing nodes, one needs to get the list of the reserved nodes. For getting that list, once you have the allocation, one can run the following command:
scontrol show hostname
This list must be included in the settings as the list of hosts where to run Fluent. Alternatively, one can give that list as parameter (-cnf
option) when running fluent
,
as follows:
[Running Fluent with 'salloc' example]
(base) [caubet_m@merlin-l-001 caubet_m]$ salloc --nodes=2 --ntasks=88 --hint=nomultithread --time=0-01:00:00 --partition=test $SHELL
salloc: Pending job allocation 135030174
salloc: job 135030174 queued and waiting for resources
salloc: job 135030174 has been allocated resources
salloc: Granted job allocation 135030174
(base) [caubet_m@merlin-l-001 caubet_m]$ module use unstable
(
base) [caubet_m@merlin-l-001 caubet_m]$ module load ANSYS/2020R1-1
module load: unstable module has been loaded -- ANSYS/2020R1-1
(base) [caubet_m@merlin-l-001 caubet_m]$ fluent 3ddp -t$SLURM_NPROCS -cnf=$(scontrol show hostname | tr '\n' ',')
(base) [caubet_m@merlin-l-001 caubet_m]$ exit
exit
salloc: Relinquishing job allocation 135030174
salloc: Job allocation 135030174 has been revoked.