Namespaces
You don’t have to configure namespaces prior to install. Domino will create three namespaces in the cluster during installation, according to the following specifications:
Namespace | Contains |
---|---|
| Durable Domino application, metadata, and platform services required for platform operation. |
| Ephemeral Domino execution pods launched by user actions in the application. |
| Domino installation metadata and secrets. |
Node pools
The OpenShift cluster must have worker nodes with the following specifications and distinct node labels. It includes optional pools:
-
OpenShift Container Storage (OCS). This pool runs the storage nodes as part of the OCS Operator which is part of the OpenShift Data Foundation (ODF) Operator.
-
GPU. Nodes in this pool contain Nvidia GPUs.
Pool | Min-Max | vCPU | Memory | Disk | Labels |
---|---|---|---|---|---|
| 4-6 | 8 | 32G | 128G |
|
| 1-20 | 8 | 32G | 128G |
|
Optional: | 0-5 | 8 | 32G | 128G |
|
Optional: | 3-3 | 8 | 32G | 128G |
|
Generally, the platform
worker nodes need an aggregate minimum of 24 CPUs and 96G of memory.
Domino recommends that you spread the resources across multiple nodes with proper failure isolation (for example, availability zones).
We recommend deploying to at least three availability zones (AZs) for high availability and tolerance. You must create a machineset per AZ per node pool, as shown in this AWS MachineSet Example.
Node autoscaling
For clusters on top of an elastic cloud provider like AWS, you must create ClusterAutoscaler, MachineAutoscaler and MachineHealthCheck resources to achieve node autoscaling.
GPU support
In order to run GPU workloads in OpenShift, the following must be installed:
-
Node Feature Discovery (NFD) Operator
-
NFD Instance
-
Nvidia GPU Operator
-
ClusterPolicy
-
GPU Enabled MachineSet
You can use the GPU Operator on OpenShift guide.
To confirm that you are able to schedule GPU workloads, you can create a pod that requires a GPU node:
spec:
resources:
limits:
nvidia.com/gpu: 1
Storage
See the storage requirements for your infrastructure.
We recommend using the Openshift Data Foundation (ODF) Operator to handle the storage.
In order to create a storage cluster for ODF, the following must be installed:
-
OCS Dedicated MachineSet (optional but recommended)
-
ODF Operator
-
StorageSystem
-
StorageCluster
You can use the ODF CLI Install guide.
Confirm the following storageclasses
are created:
-
ocs-storagecluster-ceph-rbd
-
ocs-storagecluster-cephfs
Domain
Domino must be configured to serve from a specific FQDN. To serve Domino securely over HTTPS, you also need an SSL certificate that covers the chosen name.
Important
|
A Domino install can’t be hosted on a subdomain of another Domino install.
For example, if you have Domino deployed at data-science.example.com , you can’t deploy another instance of Domino at acme.data-science.example.com .
|
Network plugin
Domino relies on Kubernetes network policies to manage secure communication between pods in the cluster. By default, OpenShift uses the Cluster Network Operator to deploy the OpenShift SDN default CNI network provider plugin, which supports network policies and hence should just work.
Ingress
Domino uses the NGNIX ingress controller maintained by the Kubernetes project instead of (but does not replace) the OpenShift implemented HAProxy-based ingress controller and deploys the ingress controller as a node port service.
By default, the ingress listens on node ports 443
(HTTPS) and 80
(HTTP).
Load balancer
A load balancer must be set up to use your DNS name. For example, in AWS, you must set up the DNS so it points a CNAME at an Elastic Load Balancer.
After you complete the installation process, you must configure the load balancer to balance across the platform nodes at the ports specified by your ingress.
Container registry
Domino deploys its own container image registry instead of using the OpenShift built in container image registry.
During installation, the OpenShift cluster image configuration is modified to trust the Domino certificate authority (CA).
This is done to ensure that OpenShift can run pods using Domino’s custom built images.
In the images.config.openshift.io/cluster
resource, you can find a reference to a ConfigMap that contains the Domino CA.
spec:
additionalTrustedCA:
name: domino-deployment-registry-config
If you opt to use the ddlctl bootstrap
subcommand to install the platform operator and Domino, make sure to pass the --openshift
flag.
For more information on bootstrapping, see Bootstrap a Domino Cluster.
-
Get the
$FLEETCOMMAND_AGENT_TAG
for your target release from the releases page. -
Use environment variables to set some values used by the
ddlctl
CLI. This simplifies the commands you’ll run while installing Domino components:unset HISTFILE export QUAY_USERNAME=<`quay.io` username provided by Domino> export QUAY_PASSWORD=<`quay.io` password provided by Domino> export FLEETCOMMAND_AGENT_TAG=<Tag that corresponds to the version of Domino deployed>
-
Generate an OpenShift configuration file:
ddlctl create config --agent-version $FLEETCOMMAND_AGENT_TAG --preset openshift
ImportantChanging the defaults in the generated configuration can affect the deployment. If you must adjust its parameters, contact a Domino representative.
You can proceed with the installation as outlined in Install Domino.