Skip to content


Before proceeding with this document, please review the installation types documentation to understand the difference between air-gapped and connected installations.

Control-plane and clusters

As part of the installation process you will install:

  • A control-plane managing cluster
  • One or more clusters

Both the control plane and clusters require Kubernetes. Typically the control plane and first cluster are installed on the same Kubernetes cluster but this is not a must.


In OpenShift environments, adding a cluster connecting to a remote control plane currently requires the assistance of customer support.

Hardware Requirements

See Cluster prerequisites hardware requirements.

Run:ai Software

You should receive a file: runai-gcr-secret.yaml from Run:ai Customer Support. The file provides access to the Run:ai Container registry.

You should receive a single file runai-<version>.tar from Run:ai customer support

Run:ai Software Prerequisites

Operating System

OpenShift has specific operating system requirements that can be found in the RedHat documentation.


Run:ai supports OpenShift. OpenShift Versions supported are detailed here.

  • OpenShift must be configured with a trusted certificate. Run:ai installation relies on OpenShift to create certificates for subdomains.
  • OpenShift must have a configured identity provider (Idp).
  • If your network is air-gapped, you will need to provide the Run:ai control-plane and cluster with information about the local certificate authority.

(Air-gapped only) Local Certificate Authority

In Air-gapped environments, you must prepare the public key of your local certificate authority as described here. It will need to be installed in Kubernetes for the installation to succeed.

NVIDIA Prerequisites

See Run:ai Cluster prerequisites installing NVIDIA dependencies in OpenShift.

The Run:ai control plane, when installed without a Run:ai cluster, does not require the NVIDIA prerequisites.

Information on how to download the GPU Operator for air-gapped installation can be found in the NVIDIA GPU Operator pre-requisites.

(Optional) Inference Prerequisites

See Run:ai Cluster prerequisites Inference requirements.

The Run:ai control plane, when installed without a Run:ai cluster, does not require the Inference prerequisites.


Run:ai requires Helm 3.10 or later. To install Helm, see If you are installing an air-gapped version of Run:ai, The Run:ai tar file contains the helm binary.

Installer Machine

The machine running the installation script (typically the Kubernetes master) must have:

  • At least 50GB of free space.
  • Docker installed.


  • (Airgapped installation only) Private Docker Registry. Run:ai assumes the existence of a Docker registry for images. Most likely installed within the organization. The installation requires the network address and port for the registry (referenced below as <REGISTRY_URL>).

Pre-install Script

Once you believe that the Run:ai prerequisites are met, we highly recommend installing and running the Run:ai pre-install diagnostics script. The tool:

  • Tests the below requirements as well as additional failure points related to Kubernetes, NVIDIA, storage, and networking.
  • Looks at additional components installed and analyzes their relevancy to a successful Run:ai installation.

To use the script download the latest version of the script and run:

chmod +x preinstall-diagnostics-<platform>

If the script fails, or if the script succeeds but the Kubernetes system contains components other than Run:ai, locate the file runai-preinstall-diagnostics.txt in the current directory and send it to Run:ai technical support.

For more information on the script including additional command-line flags, see here.