7. Add Cluster Configuration Files¶
(Optional, but recommended)
The following apps do not require a cluster configuration file:
- Dashboard App
- Files App
- Shell App (but only launched from within the Files App)
The following apps require a cluster configuration file:
- Shell App (connect to a cluster login node from the Dashboard App)
- Active Jobs App (view a list of active jobs for the various clusters)
- My Jobs App (submit jobs to various clusters)
Create default directory for config files:
sudo mkdir -p /etc/ood/config/clusters.d
Add a cluster configuration file for each HPC cluster you want to provide access to. These are YAML files and must have the
.yml
file extension (e.g.,cluster1.yml
,cluster2.yml
, …)
7.1. Example Configs¶
7.1.1. Minimal Configuration¶
A YAML cluster configuration file for an HPC cluster with only a login node but no resource manager looks like:
# /etc/ood/config/clusters.d/cluster1.yml
---
v2:
metadata:
title: "Cluster 1"
login:
host: "cluster1.my_center.edu"
Where host
is the SSH server host for the given cluster.
Note
The Active Jobs App and the My Jobs App won’t be able to list or submit jobs without a resource manager.
7.1.2. Torque/PBS¶
A YAML cluster configuration file for a Torque/PBS resource manager on an HPC cluster looks like:
# /etc/ood/config/clusters.d/cluster1.yml
---
v2:
metadata:
title: "Cluster 1"
login:
host: "cluster1.my_center.edu"
job:
adapter: "torque"
host: "cluster1-batch.my_center.edu"
lib: "/path/to/torque/lib"
bin: "/path/to/torque/bin"
Where the adapter
is set to torque
for the job feature in the cluster
configuration file.
Warning
The corresponding cluster’s batch server will need to be configured with the
Open OnDemand server as a valid submit_host
to allow the My Jobs App
to submit jobs to it.
7.1.3. Slurm¶
A YAML cluster configuration file for a Slurm resource manager on an HPC cluster looks like:
# /etc/ood/config/clusters.d/cluster1.yml
---
v2:
metadata:
title: "Cluster 1"
login:
host: "cluster1.my_center.edu"
job:
adapter: "slurm"
cluster: "cluster1"
bin: "/path/to/slurm/bin"
Where the adapter
is set to slurm
for the job feature in the cluster
configuration file.
Note
If you do not have a multi-cluster Slurm setup you can remove the cluster:
"cluster1"
line from the above configuration file.
Warning
The Open OnDemand server will need the appropriate MUNGE keys (see Slurm Quick Start Administrator Guide) for the various clusters to be able to status and submit batch jobs.
7.1.4. LSF¶
A YAML cluster configuration file for an LSF resource manager on an HPC cluster looks like:
# /etc/ood/config/clusters.d/cluster1.yml
---
v2:
metadata:
title: "Cluster 1"
login:
host: "cluster1.my_center.edu"
job:
adapter: "lsf"
bindir: "/path/to/lsf/bin"
libdir: "/path/to/lsf/lib"
envdir: "/path/to/lsf/conf"
serverdir: "/path/to/lsf/etc"
Where the adapter
is set to lsf
for the job feature in the cluster
configuration file.
Warning
Verified for only LSF 8.3 and support for LSF MultiCluster is not yet implemented.