Questo contenuto non è disponibile nella lingua selezionata.
Chapter 15. Installing IBM Cloud Bare Metal (Classic)
15.1. Prerequisites Copia collegamentoCollegamento copiato negli appunti!
You can use installer-provisioned installation to install OpenShift Container Platform on IBM Cloud® Bare Metal (Classic) nodes. This document describes the prerequisites and procedures when installing OpenShift Container Platform on IBM Cloud nodes.
Red Hat supports IPMI and PXE on the provisioning network only. Red Hat has not tested Red Fish, virtual media, or other complementary technologies such as Secure Boot on IBM Cloud deployments. A provisioning network is required.
Installer-provisioned installation of OpenShift Container Platform requires:
- One node with Red Hat Enterprise Linux CoreOS (RHCOS) 8.x installed, for running the provisioner
- Three control plane nodes
- One routable network
- One provisioning network
Before starting an installer-provisioned installation of OpenShift Container Platform on IBM Cloud Bare Metal (Classic), address the following prerequisites and requirements.
15.1.1. Setting up IBM Cloud Bare Metal (Classic) infrastructure Copia collegamentoCollegamento copiato negli appunti!
To deploy an OpenShift Container Platform cluster on IBM Cloud® Bare Metal (Classic) infrastructure, you must first provision the IBM Cloud nodes.
Red Hat supports IPMI and PXE on the
provisioning
provisioning
You can customize IBM Cloud nodes using the IBM Cloud API. When creating IBM Cloud nodes, you must consider the following requirements.
Use one data center per cluster
All nodes in the OpenShift Container Platform cluster must run in the same IBM Cloud data center.
Create public and private VLANs
Create all nodes with a single public VLAN and a single private VLAN.
Ensure subnets have sufficient IP addresses
IBM Cloud public VLAN subnets use a
/28
baremetal
IBM Cloud private VLAN subnets use a
/26
provisioning
provisioning
| IP addresses | Prefix |
|---|---|
| 32 |
|
| 64 |
|
| 128 |
|
| 256 |
|
Configuring NICs
OpenShift Container Platform deploys with two networks:
-
: The
provisioningnetwork is a non-routable network used for provisioning the underlying operating system on each node that is a part of the OpenShift Container Platform cluster.provisioning -
: The
baremetalnetwork is a routable network. You can use any NIC order to interface with thebaremetalnetwork, provided it is not the NIC specified in thebaremetalconfiguration setting or the NIC associated to a node’sprovisioningNetworkInterfaceconfiguration setting for thebootMACAddressnetwork.provisioning
While the cluster nodes can contain more than two NICs, the installation process only focuses on the first two NICs. For example:
| NIC | Network | VLAN |
|---|---|---|
| NIC1 |
| <provisioning_vlan> |
| NIC2 |
| <baremetal_vlan> |
In the previous example, NIC1 on all control plane and worker nodes connects to the non-routable network (
provisioning
baremetal
| PXE | Boot order |
|---|---|
| NIC1 PXE-enabled
| 1 |
| NIC2
| 2 |
Ensure PXE is enabled on the NIC used for the
provisioning
Configuring canonical names
Clients access the OpenShift Container Platform cluster nodes over the
baremetal
<cluster_name>.<domain>
For example:
test-cluster.example.com
Creating DNS entries
You must create DNS
A
| Usage | Host Name | IP |
|---|---|---|
| API | api.<cluster_name>.<domain> | <ip> |
| Ingress LB (apps) | *.apps.<cluster_name>.<domain> | <ip> |
Control plane and worker nodes already have DNS entries after provisioning.
The following table provides an example of fully qualified domain names. The API and Nameserver addresses begin with canonical name extensions. The host names of the control plane and worker nodes are examples, so you can use any host naming convention you prefer.
| Usage | Host Name | IP |
|---|---|---|
| API | api.<cluster_name>.<domain> | <ip> |
| Ingress LB (apps) | *.apps.<cluster_name>.<domain> | <ip> |
| Provisioner node | provisioner.<cluster_name>.<domain> | <ip> |
| Master-0 | openshift-master-0.<cluster_name>.<domain> | <ip> |
| Master-1 | openshift-master-1.<cluster_name>.<domain> | <ip> |
| Master-2 | openshift-master-2.<cluster_name>.<domain> | <ip> |
| Worker-0 | openshift-worker-0.<cluster_name>.<domain> | <ip> |
| Worker-1 | openshift-worker-1.<cluster_name>.<domain> | <ip> |
| Worker-n | openshift-worker-n.<cluster_name>.<domain> | <ip> |
OpenShift Container Platform includes functionality that uses cluster membership information to generate
A
After provisioning the IBM Cloud nodes, you must create a DNS entry for the
api.<cluster_name>.<domain>
api.<cluster_name>.<domain>
Network Time Protocol (NTP)
Each OpenShift Container Platform node in the cluster must have access to an NTP server. OpenShift Container Platform nodes use NTP to synchronize their clocks. For example, cluster nodes use SSL certificates that require validation, which might fail if the date and time between the nodes are not in sync.
Define a consistent clock date and time format in each cluster node’s BIOS settings, or installation might fail.
Configure a DHCP server
IBM Cloud Bare Metal (Classic) does not run DHCP on the public or private VLANs. After provisioning IBM Cloud nodes, you must set up a DHCP server for the public VLAN, which corresponds to OpenShift Container Platform’s
baremetal
The IP addresses allocated to each node do not need to match the IP addresses allocated by the IBM Cloud Bare Metal (Classic) provisioning system.
See the "Configuring the public subnet" section for details.
Ensure BMC access privileges
The "Remote management" page for each node on the dashboard contains the node’s intelligent platform management interface (IPMI) credentials. The default IPMI privileges prevent the user from making certain boot target changes. You must change the privilege level to
OPERATOR
In the
install-config.yaml
privilegelevel
ipmi://<IP>:<port>?privilegelevel=OPERATOR
Alternatively, contact IBM Cloud support and request that they increase the IPMI privileges to
ADMINISTRATOR
Create bare metal servers
Create bare metal servers in the IBM Cloud dashboard by navigating to Create resource
Alternatively, you can create bare metal servers with the
ibmcloud
$ ibmcloud sl hardware create --hostname <SERVERNAME> \
--domain <DOMAIN> \
--size <SIZE> \
--os <OS-TYPE> \
--datacenter <DC-NAME> \
--port-speed <SPEED> \
--billing <BILLING>
See Installing the stand-alone IBM Cloud CLI for details on installing the IBM Cloud CLI.
IBM Cloud servers might take 3-5 hours to become available.
15.2. Setting up the environment for an OpenShift Container Platform installation Copia collegamentoCollegamento copiato negli appunti!
15.2.1. Preparing the provisioner node on IBM Cloud Bare Metal (Classic) infrastructure Copia collegamentoCollegamento copiato negli appunti!
Perform the following steps to prepare the provisioner node.
Procedure
-
Log in to the provisioner node via .
ssh Create a non-root user (
) and provide that user withkniprivileges:sudo# useradd kni# passwd kni# echo "kni ALL=(root) NOPASSWD:ALL" | tee -a /etc/sudoers.d/kni# chmod 0440 /etc/sudoers.d/kniCreate an
key for the new user:ssh# su - kni -c "ssh-keygen -f /home/kni/.ssh/id_rsa -N ''"Log in as the new user on the provisioner node:
# su - kniUse Red Hat Subscription Manager to register the provisioner node:
$ sudo subscription-manager register --username=<user> --password=<pass> --auto-attach$ sudo subscription-manager repos --enable=rhel-8-for-x86_64-appstream-rpms \ --enable=rhel-8-for-x86_64-baseos-rpmsNoteFor more information about Red Hat Subscription Manager, see Using and Configuring Red Hat Subscription Manager.
Install the following packages:
$ sudo dnf install -y libvirt qemu-kvm mkisofs python3-devel jq ipmitoolModify the user to add the
group to the newly created user:libvirt$ sudo usermod --append --groups libvirt kniStart
:firewalld$ sudo systemctl start firewalldEnable
:firewalld$ sudo systemctl enable firewalldStart the
service:http$ sudo firewall-cmd --zone=public --add-service=http --permanent$ sudo firewall-cmd --reloadStart and enable the
service:libvirtd$ sudo systemctl enable libvirtd --nowSet the ID of the provisioner node:
$ PRVN_HOST_ID=<ID>You can view the ID with the following
command:ibmcloud$ ibmcloud sl hardware listSet the ID of the public subnet:
$ PUBLICSUBNETID=<ID>You can view the ID with the following
command:ibmcloud$ ibmcloud sl subnet listSet the ID of the private subnet:
$ PRIVSUBNETID=<ID>You can view the ID with the following
command:ibmcloud$ ibmcloud sl subnet listSet the provisioner node public IP address:
$ PRVN_PUB_IP=$(ibmcloud sl hardware detail $PRVN_HOST_ID --output JSON | jq .primaryIpAddress -r)Set the CIDR for the public network:
$ PUBLICCIDR=$(ibmcloud sl subnet detail $PUBLICSUBNETID --output JSON | jq .cidr)Set the IP address and CIDR for the public network:
$ PUB_IP_CIDR=$PRVN_PUB_IP/$PUBLICCIDRSet the gateway for the public network:
$ PUB_GATEWAY=$(ibmcloud sl subnet detail $PUBLICSUBNETID --output JSON | jq .gateway -r)Set the private IP address of the provisioner node:
$ PRVN_PRIV_IP=$(ibmcloud sl hardware detail $PRVN_HOST_ID --output JSON | \ jq .primaryBackendIpAddress -r)Set the CIDR for the private network:
$ PRIVCIDR=$(ibmcloud sl subnet detail $PRIVSUBNETID --output JSON | jq .cidr)Set the IP address and CIDR for the private network:
$ PRIV_IP_CIDR=$PRVN_PRIV_IP/$PRIVCIDRSet the gateway for the private network:
$ PRIV_GATEWAY=$(ibmcloud sl subnet detail $PRIVSUBNETID --output JSON | jq .gateway -r)Set up the bridges for the
andbaremetalnetworks:provisioning$ sudo nohup bash -c " nmcli --get-values UUID con show | xargs -n 1 nmcli con delete nmcli connection add ifname provisioning type bridge con-name provisioning nmcli con add type bridge-slave ifname eth1 master provisioning nmcli connection add ifname baremetal type bridge con-name baremetal nmcli con add type bridge-slave ifname eth2 master baremetal nmcli connection modify baremetal ipv4.addresses $PUB_IP_CIDR ipv4.method manual ipv4.gateway $PUB_GATEWAY nmcli connection modify provisioning ipv4.addresses 172.22.0.1/24,$PRIV_IP_CIDR ipv4.method manual nmcli connection modify provisioning +ipv4.routes \"10.0.0.0/8 $PRIV_GATEWAY\" nmcli con down baremetal nmcli con up baremetal nmcli con down provisioning nmcli con up provisioning init 6 "NoteFor
andeth1, substitute the appropriate interface name, as needed.eth2If required, SSH back into the
node:provisioner# ssh kni@provisioner.<cluster-name>.<domain>Verify the connection bridges have been properly created:
$ sudo nmcli con showExample output
NAME UUID TYPE DEVICE baremetal 4d5133a5-8351-4bb9-bfd4-3af264801530 bridge baremetal provisioning 43942805-017f-4d7d-a2c2-7cb3324482ed bridge provisioning virbr0 d9bca40f-eee1-410b-8879-a2d4bb0465e7 bridge virbr0 bridge-slave-eth1 76a8ed50-c7e5-4999-b4f6-6d9014dd0812 ethernet eth1 bridge-slave-eth2 f31c3353-54b7-48de-893a-02d2b34c4736 ethernet eth2Create a
file:pull-secret.txt$ vim pull-secret.txtIn a web browser, navigate to Install on Bare Metal with user-provisioned infrastructure. In step 1, click Download pull secret. Paste the contents into the
file and save the contents in thepull-secret.txtuser’s home directory.kni
15.2.2. Configuring the public subnet Copia collegamentoCollegamento copiato negli appunti!
All of the OpenShift Container Platform cluster nodes must be on the public subnet. IBM Cloud® Bare Metal (Classic) does not provide a DHCP server on the subnet. Set it up separately on the provisioner node.
You must reset the BASH variables defined when preparing the provisioner node. Rebooting the provisioner node after preparing it will delete the BASH variables previously set.
Procedure
Install
:dnsmasq$ sudo dnf install dnsmasqOpen the
configuration file:dnsmasq$ sudo vi /etc/dnsmasq.confAdd the following configuration to the
configuration file:dnsmasqinterface=baremetal except-interface=lo bind-dynamic log-dhcp dhcp-range=<ip_addr>,<ip_addr>,<pub_cidr>1 dhcp-option=baremetal,121,0.0.0.0/0,<pub_gateway>,<prvn_priv_ip>,<prvn_pub_ip>2 dhcp-hostsfile=/var/lib/dnsmasq/dnsmasq.hostsfile- 1
- Set the DHCP range. Replace both instances of
<ip_addr>with one unused IP address from the public subnet so that thedhcp-rangefor thebaremetalnetwork begins and ends with the same the IP address. Replace<pub_cidr>with the CIDR of the public subnet. - 2
- Set the DHCP option. Replace
<pub_gateway>with the IP address of the gateway for thebaremetalnetwork. Replace<prvn_priv_ip>with the IP address of the provisioner node’s private IP address on theprovisioningnetwork. Replace<prvn_pub_ip>with the IP address of the provisioner node’s public IP address on thebaremetalnetwork.
To retrieve the value for
, execute:<pub_cidr>$ ibmcloud sl subnet detail <publicsubnetid> --output JSON | jq .cidrReplace
with the ID of the public subnet.<publicsubnetid>To retrieve the value for
, execute:<pub_gateway>$ ibmcloud sl subnet detail <publicsubnetid> --output JSON | jq .gateway -rReplace
with the ID of the public subnet.<publicsubnetid>To retrieve the value for
, execute:<prvn_priv_ip>$ ibmcloud sl hardware detail <id> --output JSON | \ jq .primaryBackendIpAddress -rReplace
with the ID of the provisioner node.<id>To retrieve the value for
, execute:<prvn_pub_ip>$ ibmcloud sl hardware detail <id> --output JSON | jq .primaryIpAddress -rReplace
with the ID of the provisioner node.<id>Obtain the list of hardware for the cluster:
$ ibmcloud sl hardware listObtain the MAC addresses and IP addresses for each node:
$ ibmcloud sl hardware detail <id> --output JSON | \ jq '.networkComponents[] | \ "\(.primaryIpAddress) \(.macAddress)"' | grep -v nullReplace
with the ID of the node.<id>Example output
"10.196.130.144 00:e0:ed:6a:ca:b4" "141.125.65.215 00:e0:ed:6a:ca:b5"Make a note of the MAC address and IP address of the public network. Make a separate note of the MAC address of the private network, which you will use later in the
file. Repeat this procedure for each node until you have all the public MAC and IP addresses for the publicinstall-config.yamlnetwork, and the MAC addresses of the privatebaremetalnetwork.provisioningAdd the MAC and IP address pair of the public
network for each node into thebaremetalfile:dnsmasq.hostsfile$ sudo vim /var/lib/dnsmasq/dnsmasq.hostsfileExample input
00:e0:ed:6a:ca:b5,141.125.65.215,master-0 <mac>,<ip>,master-1 <mac>,<ip>,master-2 <mac>,<ip>,worker-0 <mac>,<ip>,worker-1 ...Replace
with the public MAC address and public IP address of the corresponding node name.<mac>,<ip>Start
:dnsmasq$ sudo systemctl start dnsmasqEnable
so that it starts when booting the node:dnsmasq$ sudo systemctl enable dnsmasqVerify
is running:dnsmasq$ sudo systemctl status dnsmasqExample output
● dnsmasq.service - DNS caching server. Loaded: loaded (/usr/lib/systemd/system/dnsmasq.service; enabled; vendor preset: disabled) Active: active (running) since Tue 2021-10-05 05:04:14 CDT; 49s ago Main PID: 3101 (dnsmasq) Tasks: 1 (limit: 204038) Memory: 732.0K CGroup: /system.slice/dnsmasq.service └─3101 /usr/sbin/dnsmasq -kOpen ports
and53with UDP protocol:67$ sudo firewall-cmd --add-port 53/udp --permanent$ sudo firewall-cmd --add-port 67/udp --permanentAdd
to the external zone with masquerade:provisioning$ sudo firewall-cmd --change-zone=provisioning --zone=external --permanentThis step ensures network address translation for IPMI calls to the management subnet.
Reload the
configuration:firewalld$ sudo firewall-cmd --reload
15.2.3. Retrieving the OpenShift Container Platform installer Copia collegamentoCollegamento copiato negli appunti!
Use the
stable-4.x
$ export VERSION=stable-4.11
$ export RELEASE_ARCH=<architecture>
$ export RELEASE_IMAGE=$(curl -s https://mirror.openshift.com/pub/openshift-v4/$RELEASE_ARCH/clients/ocp/$VERSION/release.txt | grep 'Pull From: quay.io' | awk -F ' ' '{print $3}')
15.2.4. Extracting the OpenShift Container Platform installer Copia collegamentoCollegamento copiato negli appunti!
After retrieving the installer, the next step is to extract it.
Procedure
Set the environment variables:
$ export cmd=openshift-baremetal-install$ export pullsecret_file=~/pull-secret.txt$ export extract_dir=$(pwd)Get the
binary:oc$ curl -s https://mirror.openshift.com/pub/openshift-v4/clients/ocp/$VERSION/openshift-client-linux.tar.gz | tar zxvf - ocExtract the installer:
$ sudo cp oc /usr/local/bin$ oc adm release extract --registry-config "${pullsecret_file}" --command=$cmd --to "${extract_dir}" ${RELEASE_IMAGE}$ sudo cp openshift-baremetal-install /usr/local/bin
15.2.5. Configuring the install-config.yaml file Copia collegamentoCollegamento copiato negli appunti!
The
install-config.yaml
install-config.yaml
Procedure
Configure
. Change the appropriate variables to match the environment, includinginstall-config.yamlandpullSecret.sshKeyapiVersion: v1 baseDomain: <domain> metadata: name: <cluster_name> networking: machineNetwork: - cidr: <public-cidr> networkType: OVNKubernetes compute: - name: worker replicas: 2 controlPlane: name: master replicas: 3 platform: baremetal: {} platform: baremetal: apiVIP: <api_ip> ingressVIP: <wildcard_ip> provisioningNetworkInterface: <NIC1> provisioningNetworkCIDR: <CIDR> hosts: - name: openshift-master-0 role: master bmc: address: ipmi://10.196.130.145?privilegelevel=OPERATOR1 username: root password: <password> bootMACAddress: 00:e0:ed:6a:ca:b42 rootDeviceHints: deviceName: "/dev/sda" - name: openshift-worker-0 role: worker bmc: address: ipmi://<out-of-band-ip>?privilegelevel=OPERATOR3 username: <user> password: <password> bootMACAddress: <NIC1_mac_address>4 rootDeviceHints: deviceName: "/dev/sda" pullSecret: '<pull_secret>' sshKey: '<ssh_pub_key>'NoteYou can use the
command-line utility to retrieve the password.ibmcloud$ ibmcloud sl hardware detail <id> --output JSON | \ jq '"(.networkManagementIpAddress) (.remoteManagementAccounts[0].password)"'Replace
with the ID of the node.<id>Create a directory to store the cluster configuration:
$ mkdir ~/clusterconfigsCopy the
file into the directory:install-config.yaml$ cp install-config.yaml ~/clusterconfigEnsure all bare metal nodes are powered off prior to installing the OpenShift Container Platform cluster:
$ ipmitool -I lanplus -U <user> -P <password> -H <management_server_ip> power offRemove old bootstrap resources if any are left over from a previous deployment attempt:
for i in $(sudo virsh list | tail -n +3 | grep bootstrap | awk {'print $2'}); do sudo virsh destroy $i; sudo virsh undefine $i; sudo virsh vol-delete $i --pool $i; sudo virsh vol-delete $i.ign --pool $i; sudo virsh pool-destroy $i; sudo virsh pool-undefine $i; done
15.2.6. Additional install-config parameters Copia collegamentoCollegamento copiato negli appunti!
See the following tables for the required parameters, the
hosts
bmc
install-config.yaml
| Parameters | Default | Description |
|---|---|---|
|
| The domain name for the cluster. For example,
| |
|
|
| The boot mode for a node. Options are
|
|
| The static IP address for the bootstrap VM. You must set this value when deploying a cluster with static IP addresses when there is no DHCP server on the bare-metal network. | |
|
| The static IP address of the gateway for the bootstrap VM. You must set this value when deploying a cluster with static IP addresses when there is no DHCP server on the bare-metal network. | |
|
| The
| |
|
| The
| |
| The name to be given to the OpenShift Container Platform cluster. For example,
| |
| The public CIDR (Classless Inter-Domain Routing) of the external network. For example,
| |
| The OpenShift Container Platform cluster requires a name be provided for worker (or compute) nodes even if there are zero nodes. | |
| Replicas sets the number of worker (or compute) nodes in the OpenShift Container Platform cluster. | |
| The OpenShift Container Platform cluster requires a name for control plane (master) nodes. | |
| Replicas sets the number of control plane (master) nodes included as part of the OpenShift Container Platform cluster. | |
|
| The name of the network interface on nodes connected to the provisioning network. For OpenShift Container Platform 4.9 and later releases, use the
| |
|
| The default configuration used for machine pools without a platform configuration. | |
|
| (Optional) The virtual IP address for Kubernetes API communication. This setting must either be provided in the
| |
|
|
|
|
|
| (Optional) The virtual IP address for ingress traffic. This setting must either be provided in the
|
| Parameters | Default | Description |
|---|---|---|
|
|
| Defines the IP range for nodes on the provisioning network. |
|
|
| The CIDR for the network to use for provisioning. This option is required when not using the default address range on the provisioning network. |
|
| The third IP address of the
| The IP address within the cluster where the provisioning services run. Defaults to the third IP address of the provisioning subnet. For example,
|
|
| The second IP address of the
| The IP address on the bootstrap VM where the provisioning services run while the installer is deploying the control plane (master) nodes. Defaults to the second IP address of the provisioning subnet. For example,
|
|
|
| The name of the bare-metal bridge of the hypervisor attached to the bare-metal network. |
|
|
| The name of the provisioning bridge on the
|
|
| Defines the host architecture for your cluster. Valid values are
| |
|
| The default configuration used for machine pools without a platform configuration. | |
|
| A URL to override the default operating system image for the bootstrap node. The URL must contain a SHA-256 hash of the image. For example:
| |
|
| The
| |
|
| Set this parameter to the appropriate HTTP proxy used within your environment. | |
|
| Set this parameter to the appropriate HTTPS proxy used within your environment. | |
|
| Set this parameter to the appropriate list of exclusions for proxy usage within your environment. |
Hosts
The
hosts
| Name | Default | Description |
|---|---|---|
|
| The name of the
| |
|
| The role of the bare metal node. Either
| |
|
| Connection details for the baseboard management controller. See the BMC addressing section for additional details. | |
|
| The MAC address of the NIC that the host uses for the provisioning network. Ironic retrieves the IP address using the
Note You must provide a valid MAC address from the host if you disabled the provisioning network. | |
|
| Set this optional parameter to configure the network interface of a host. See "(Optional) Configuring host network interfaces" for additional details. |
15.2.7. Root device hints Copia collegamentoCollegamento copiato negli appunti!
The
rootDeviceHints
| Subfield | Description |
|---|---|
|
| A string containing a Linux device name like
|
|
| A string containing a SCSI bus address like
|
|
| A string containing a vendor-specific device identifier. The hint can be a substring of the actual value. |
|
| A string containing the name of the vendor or manufacturer of the device. The hint can be a sub-string of the actual value. |
|
| A string containing the device serial number. The hint must match the actual value exactly. |
|
| An integer representing the minimum size of the device in gigabytes. |
|
| A string containing the unique storage identifier. The hint must match the actual value exactly. |
|
| A string containing the unique storage identifier with the vendor extension appended. The hint must match the actual value exactly. |
|
| A string containing the unique vendor storage identifier. The hint must match the actual value exactly. |
|
| A boolean indicating whether the device should be a rotating disk (true) or not (false). |
Example usage
- name: master-0
role: master
bmc:
address: ipmi://10.10.0.3:6203
username: admin
password: redhat
bootMACAddress: de:ad:be:ef:00:40
rootDeviceHints:
deviceName: "/dev/sda"
15.2.8. Creating the OpenShift Container Platform manifests Copia collegamentoCollegamento copiato negli appunti!
Create the OpenShift Container Platform manifests.
$ ./openshift-baremetal-install --dir ~/clusterconfigs create manifestsINFO Consuming Install Config from target directory WARNING Making control-plane schedulable by setting MastersSchedulable to true for Scheduler cluster settings WARNING Discarding the OpenShift Manifest that was provided in the target directory because its dependencies are dirty and it needs to be regenerated
15.2.9. Deploying the cluster via the OpenShift Container Platform installer Copia collegamentoCollegamento copiato negli appunti!
Run the OpenShift Container Platform installer:
$ ./openshift-baremetal-install --dir ~/clusterconfigs --log-level debug create cluster
15.2.10. Following the installation Copia collegamentoCollegamento copiato negli appunti!
During the deployment process, you can check the installation’s overall status by issuing the
tail
.openshift_install.log
$ tail -f /path/to/install-dir/.openshift_install.log