7. Add Cluster Configuration Files

(Optional, but recommended)

The following apps do not require a cluster configuration file:

The following apps require a cluster configuration file:

  1. Create default directory for config files:

    sudo mkdir -p /etc/ood/config/clusters.d
    
  2. Add a cluster configuration file for each HPC cluster you want to provide access to. These are YAML files and must have the .yml file extension (e.g., cluster1.yml, cluster2.yml, …)

7.1. Example Configs

7.1.1. Minimal Configuration

A YAML cluster configuration file for an HPC cluster with only a login node but no resource manager looks like:

# /etc/ood/config/clusters.d/cluster1.yml
---
v2:
  metadata:
    title: "Cluster 1"
  login:
    host: "cluster1.my_center.edu"

Where host is the SSH server host for the given cluster.

Note

The Active Jobs App and the Job Composer App won’t be able to list or submit jobs without a resource manager.

7.1.2. Torque

A YAML cluster configuration file for a Torque/PBS resource manager on an HPC cluster looks like:

# /etc/ood/config/clusters.d/cluster1.yml
---
v2:
  metadata:
    title: "Cluster 1"
  login:
    host: "cluster1.my_center.edu"
  job:
    adapter: "torque"
    host: "cluster1-batch.my_center.edu"
    lib: "/path/to/torque/lib"
    bin: "/path/to/torque/bin"

with the following configuration options:

adapter
This is set to torque.
host
The host of the Torque batch server.
lib
The path to the Torque client libraries.
bin
The path to the Torque client binaries.

Warning

The corresponding cluster’s batch server will need to be configured with the Open OnDemand server as a valid submit_host to allow the Job Composer App to submit jobs to it.

7.1.3. Slurm

A YAML cluster configuration file for a Slurm resource manager on an HPC cluster looks like:

# /etc/ood/config/clusters.d/cluster1.yml
---
v2:
  metadata:
    title: "Cluster 1"
  login:
    host: "cluster1.my_center.edu"
  job:
    adapter: "slurm"
    cluster: "cluster1"
    bin: "/path/to/slurm/bin"
    conf: "/path/to/slurm.conf"

with the following configuration options:

adapter
This is set to slurm.
cluster
The Slurm cluster name.
bin
The path to the Slurm client installation binaries.
conf
The path to the Slurm configuration file for this cluster. Optional

Note

If you do not have a multi-cluster Slurm setup you can remove the cluster: "cluster1" line from the above configuration file.

Warning

The Open OnDemand server will need the appropriate MUNGE keys (see Slurm Quick Start Administrator Guide) for the various clusters to be able to status and submit batch jobs.

7.1.4. LSF

A YAML cluster configuration file for an LSF resource manager on an HPC cluster looks like:

# /etc/ood/config/clusters.d/cluster1.yml
---
v2:
  metadata:
    title: "Cluster 1"
  login:
    host: "cluster1.my_center.edu"
  job:
    adapter: "lsf"
    bindir: "/path/to/lsf/bin"
    libdir: "/path/to/lsf/lib"
    envdir: "/path/to/lsf/conf"
    serverdir: "/path/to/lsf/etc"

with the following configuration options:

adapter
This is set to lsf.
bindir
The path to the LSF client bin/ directory.
libdir
The path to the LSF client lib/ directory.
envdir
The path to the LSF client conf/ directory.
serverdir
The path to the LSF client etc/ directory.

Warning

Verified for only LSF 8.3 and support for LSF MultiCluster is not yet implemented.

7.1.5. PBS Professional

A YAML cluster configuration file for a PBS Professional resource manager on an HPC cluster looks like:

# /etc/ood/config/clusters.d/cluster1.yml
---
v2:
  metadata:
    title: "Cluster 1"
  login:
    host: "cluster1.my_center.edu"
  job:
    adapter: "pbspro"
    host: "cluster1-batch.my_center.edu"
    exec: "/path/to/pbspro"

with the following configuration options:

adapter
This is set to pbspro.
host
The host of the PBS Pro batch server.
exec
The installation path for the PBS Pro binaries and libraries on the OnDemand host.