Questo contenuto non è disponibile nella lingua selezionata.
Chapter 7. Troubleshooting
7.1. Troubleshooting installations Copia collegamentoCollegamento copiato negli appunti!
7.1.1. Determining where installation issues occur Copia collegamentoCollegamento copiato negli appunti!
When troubleshooting OpenShift Container Platform installation issues, you can monitor installation logs to determine at which stage issues occur. Then, retrieve diagnostic data relevant to that stage.
OpenShift Container Platform installation proceeds through the following stages:
- Ignition configuration files are created.
- The bootstrap machine boots and starts hosting the remote resources required for the control plane machines to boot.
- The control plane machines fetch the remote resources from the bootstrap machine and finish booting.
- The control plane machines use the bootstrap machine to form an etcd cluster.
- The bootstrap machine starts a temporary Kubernetes control plane using the new etcd cluster.
- The temporary control plane schedules the production control plane to the control plane machines.
- The temporary control plane shuts down and passes control to the production control plane.
- The bootstrap machine adds OpenShift Container Platform components into the production control plane.
- The installation program shuts down the bootstrap machine.
- The control plane sets up the worker nodes.
- The control plane installs additional services in the form of a set of Operators.
- The cluster downloads and configures remaining components needed for the day-to-day operation, including the creation of worker machines in supported environments.
7.1.2. User-provisioned infrastructure installation considerations Copia collegamentoCollegamento copiato negli appunti!
The default installation method uses installer-provisioned infrastructure. With installer-provisioned infrastructure clusters, OpenShift Container Platform manages all aspects of the cluster, including the operating system itself. If possible, use this feature to avoid having to provision and maintain the cluster infrastructure.
You can alternatively install OpenShift Container Platform 4.11 on infrastructure that you provide. If you use this installation method, follow user-provisioned infrastructure installation documentation carefully. Additionally, review the following considerations before the installation:
- Check the Red Hat Enterprise Linux (RHEL) Ecosystem to determine the level of Red Hat Enterprise Linux CoreOS (RHCOS) support provided for your chosen server hardware or virtualization technology.
- Many virtualization and cloud environments require agents to be installed on guest operating systems. Ensure that these agents are installed as a containerized workload deployed through a daemon set.
Install cloud provider integration if you want to enable features such as dynamic storage, on-demand service routing, node hostname to Kubernetes hostname resolution, and cluster autoscaling.
NoteIt is not possible to enable cloud provider integration in OpenShift Container Platform environments that mix resources from different cloud providers, or that span multiple physical or virtual platforms. The node life cycle controller will not allow nodes that are external to the existing provider to be added to a cluster, and it is not possible to specify more than one cloud provider integration.
- A provider-specific Machine API implementation is required if you want to use machine sets or autoscaling to automatically provision OpenShift Container Platform cluster nodes.
- Check whether your chosen cloud provider offers a method to inject Ignition configuration files into hosts as part of their initial deployment. If they do not, you will need to host Ignition configuration files by using an HTTP server. The steps taken to troubleshoot Ignition configuration file issues will differ depending on which of these two methods is deployed.
- Storage needs to be manually provisioned if you want to leverage optional framework components such as the embedded container registry, Elasticsearch, or Prometheus. Default storage classes are not defined in user-provisioned infrastructure installations unless explicitly configured.
- A load balancer is required to distribute API requests across all control plane nodes in highly available OpenShift Container Platform environments. You can use any TCP-based load balancing solution that meets OpenShift Container Platform DNS routing and port requirements.
7.1.3. Checking a load balancer configuration before OpenShift Container Platform installation Copia collegamentoCollegamento copiato negli appunti!
Check your load balancer configuration prior to starting an OpenShift Container Platform installation.
Prerequisites
- You have configured an external load balancer of your choosing, in preparation for an OpenShift Container Platform installation. The following example is based on a Red Hat Enterprise Linux (RHEL) host using HAProxy to provide load balancing services to a cluster.
- You have configured DNS in preparation for an OpenShift Container Platform installation.
- You have SSH access to your load balancer.
Procedure
Check that the
systemd service is active:haproxy$ ssh <user_name>@<load_balancer> systemctl status haproxyVerify that the load balancer is listening on the required ports. The following example references ports
,80,443, and6443.22623For HAProxy instances running on Red Hat Enterprise Linux (RHEL) 6, verify port status by using the
command:netstat$ ssh <user_name>@<load_balancer> netstat -nltupe | grep -E ':80|:443|:6443|:22623'For HAProxy instances running on Red Hat Enterprise Linux (RHEL) 7 or 8, verify port status by using the
command:ss$ ssh <user_name>@<load_balancer> ss -nltupe | grep -E ':80|:443|:6443|:22623'NoteRed Hat recommends the
command instead ofssin Red Hat Enterprise Linux (RHEL) 7 or later.netstatis provided by the iproute package. For more information on thesscommand, see the Red Hat Enterprise Linux (RHEL) 7 Performance Tuning Guide.ss
Check that the wildcard DNS record resolves to the load balancer:
$ dig <wildcard_fqdn> @<dns_server>
7.1.4. Specifying OpenShift Container Platform installer log levels Copia collegamentoCollegamento copiato negli appunti!
By default, the OpenShift Container Platform installer log level is set to
info
openshift-install
debug
Prerequisites
- You have access to the installation host.
Procedure
Set the installation log level to
when initiating the installation:debug$ ./openshift-install --dir <installation_directory> wait-for bootstrap-complete --log-level debug1 - 1
- Possible log levels include
info,warn,error,anddebug.
7.1.5. Troubleshooting openshift-install command issues Copia collegamentoCollegamento copiato negli appunti!
If you experience issues running the
openshift-install
The installation has been initiated within 24 hours of Ignition configuration file creation. The Ignition files are created when the following command is run:
$ ./openshift-install create ignition-configs --dir=./install_dir-
The file is in the same directory as the installer. If an alternative installation path is declared by using the
install-config.yamloption, verify that the./openshift-install --dirfile exists within that directory.install-config.yaml
7.1.6. Monitoring installation progress Copia collegamentoCollegamento copiato negli appunti!
You can monitor high-level installation, bootstrap, and control plane logs as an OpenShift Container Platform installation progresses. This provides greater visibility into how an installation progresses and helps identify the stage at which an installation failure occurs.
Prerequisites
-
You have access to the cluster as a user with the cluster role.
cluster-admin -
You have installed the OpenShift CLI ().
oc - You have SSH access to your hosts.
You have the fully qualified domain names of the bootstrap and control plane nodes.
NoteThe initial
password can be found inkubeadminon the installation host.<install_directory>/auth/kubeadmin-password
Procedure
Watch the installation log as the installation progresses:
$ tail -f ~/<installation_directory>/.openshift_install.logMonitor the
journald unit log on the bootstrap node, after it has booted. This provides visibility into the bootstrapping of the first control plane. Replacebootkube.servicewith the bootstrap node’s fully qualified domain name:<bootstrap_fqdn>$ ssh core@<bootstrap_fqdn> journalctl -b -f -u bootkube.serviceNoteThe
log on the bootstrap node outputs etcdbootkube.serviceerrors, indicating that the bootstrap server is unable to connect to etcd on control plane nodes. After etcd has started on each control plane node and the nodes have joined the cluster, the errors should stop.connection refusedMonitor
journald unit logs on control plane nodes, after they have booted. This provides visibility into control plane node agent activity.kubelet.serviceMonitor the logs using
:oc$ oc adm node-logs --role=master -u kubeletIf the API is not functional, review the logs using SSH instead. Replace
with appropriate values:<master-node>.<cluster_name>.<base_domain>$ ssh core@<master-node>.<cluster_name>.<base_domain> journalctl -b -f -u kubelet.service
Monitor
journald unit logs on control plane nodes, after they have booted. This provides visibility into control plane node CRI-O container runtime activity.crio.serviceMonitor the logs using
:oc$ oc adm node-logs --role=master -u crioIf the API is not functional, review the logs using SSH instead. Replace
with appropriate values:<master-node>.<cluster_name>.<base_domain>$ ssh core@master-N.cluster_name.sub_domain.domain journalctl -b -f -u crio.service
7.1.7. Gathering bootstrap node diagnostic data Copia collegamentoCollegamento copiato negli appunti!
When experiencing bootstrap-related issues, you can gather
bootkube.service
journald
Prerequisites
- You have SSH access to your bootstrap node.
- You have the fully qualified domain name of the bootstrap node.
- If you are hosting Ignition configuration files by using an HTTP server, you must have the HTTP server’s fully qualified domain name and the port number. You must also have SSH access to the HTTP host.
Procedure
- If you have access to the bootstrap node’s console, monitor the console until the node reaches the login prompt.
Verify the Ignition file configuration.
If you are hosting Ignition configuration files by using an HTTP server.
Verify the bootstrap node Ignition file URL. Replace
with HTTP server’s fully qualified domain name:<http_server_fqdn>$ curl -I http://<http_server_fqdn>:<port>/bootstrap.ign1 - 1
- The
-Ioption returns the header only. If the Ignition file is available on the specified URL, the command returns200 OKstatus. If it is not available, the command returns404 file not found.
To verify that the Ignition file was received by the bootstrap node, query the HTTP server logs on the serving host. For example, if you are using an Apache web server to serve Ignition files, enter the following command:
$ grep -is 'bootstrap.ign' /var/log/httpd/access_logIf the bootstrap Ignition file is received, the associated
log message will include aHTTP GETsuccess status, indicating that the request succeeded.200 OK- If the Ignition file was not received, check that the Ignition files exist and that they have the appropriate file and web server permissions on the serving host directly.
If you are using a cloud provider mechanism to inject Ignition configuration files into hosts as part of their initial deployment.
- Review the bootstrap node’s console to determine if the mechanism is injecting the bootstrap node Ignition file correctly.
- Verify the availability of the bootstrap node’s assigned storage device.
- Verify that the bootstrap node has been assigned an IP address from the DHCP server.
Collect
journald unit logs from the bootstrap node. Replacebootkube.servicewith the bootstrap node’s fully qualified domain name:<bootstrap_fqdn>$ ssh core@<bootstrap_fqdn> journalctl -b -f -u bootkube.serviceNoteThe
log on the bootstrap node outputs etcdbootkube.serviceerrors, indicating that the bootstrap server is unable to connect to etcd on control plane nodes. After etcd has started on each control plane node and the nodes have joined the cluster, the errors should stop.connection refusedCollect logs from the bootstrap node containers.
Collect the logs using
on the bootstrap node. Replacepodmanwith the bootstrap node’s fully qualified domain name:<bootstrap_fqdn>$ ssh core@<bootstrap_fqdn> 'for pod in $(sudo podman ps -a -q); do sudo podman logs $pod; done'
If the bootstrap process fails, verify the following.
-
You can resolve from the installation host.
api.<cluster_name>.<base_domain> - The load balancer proxies port 6443 connections to bootstrap and control plane nodes. Ensure that the proxy configuration meets OpenShift Container Platform installation requirements.
-
You can resolve
7.1.8. Investigating control plane node installation issues Copia collegamentoCollegamento copiato negli appunti!
If you experience control plane node installation issues, determine the control plane node OpenShift Container Platform software defined network (SDN), and network Operator status. Collect
kubelet.service
crio.service
Prerequisites
-
You have access to the cluster as a user with the role.
cluster-admin -
You have installed the OpenShift CLI ().
oc - You have SSH access to your hosts.
- You have the fully qualified domain names of the bootstrap and control plane nodes.
If you are hosting Ignition configuration files by using an HTTP server, you must have the HTTP server’s fully qualified domain name and the port number. You must also have SSH access to the HTTP host.
NoteThe initial
password can be found inkubeadminon the installation host.<install_directory>/auth/kubeadmin-password
Procedure
- If you have access to the console for the control plane node, monitor the console until the node reaches the login prompt. During the installation, Ignition log messages are output to the console.
Verify Ignition file configuration.
If you are hosting Ignition configuration files by using an HTTP server.
Verify the control plane node Ignition file URL. Replace
with HTTP server’s fully qualified domain name:<http_server_fqdn>$ curl -I http://<http_server_fqdn>:<port>/master.ign1 - 1
- The
-Ioption returns the header only. If the Ignition file is available on the specified URL, the command returns200 OKstatus. If it is not available, the command returns404 file not found.
To verify that the Ignition file was received by the control plane node query the HTTP server logs on the serving host. For example, if you are using an Apache web server to serve Ignition files:
$ grep -is 'master.ign' /var/log/httpd/access_logIf the master Ignition file is received, the associated
log message will include aHTTP GETsuccess status, indicating that the request succeeded.200 OK- If the Ignition file was not received, check that it exists on the serving host directly. Ensure that the appropriate file and web server permissions are in place.
If you are using a cloud provider mechanism to inject Ignition configuration files into hosts as part of their initial deployment.
- Review the console for the control plane node to determine if the mechanism is injecting the control plane node Ignition file correctly.
- Check the availability of the storage device assigned to the control plane node.
- Verify that the control plane node has been assigned an IP address from the DHCP server.
Determine control plane node status.
Query control plane node status:
$ oc get nodesIf one of the control plane nodes does not reach a
status, retrieve a detailed node description:Ready$ oc describe node <master_node>NoteIt is not possible to run
commands if an installation issue prevents the OpenShift Container Platform API from running or if the kubelet is not running yet on each node:oc
Determine OpenShift Container Platform SDN status.
Review
,sdn-controller, andsdndaemon set status, in theovsnamespace:openshift-sdn$ oc get daemonsets -n openshift-sdnIf those resources are listed as
, review pods in theNot foundnamespace:openshift-sdn$ oc get pods -n openshift-sdnReview logs relating to failed OpenShift Container Platform SDN pods in the
namespace:openshift-sdn$ oc logs <sdn_pod> -n openshift-sdn
Determine cluster network configuration status.
Review whether the cluster’s network configuration exists:
$ oc get network.config.openshift.io cluster -o yamlIf the installer failed to create the network configuration, generate the Kubernetes manifests again and review message output:
$ ./openshift-install create manifestsReview the pod status in the
namespace to determine whether the Cluster Network Operator (CNO) is running:openshift-network-operator$ oc get pods -n openshift-network-operatorGather network Operator pod logs from the
namespace:openshift-network-operator$ oc logs pod/<network_operator_pod_name> -n openshift-network-operator
Monitor
journald unit logs on control plane nodes, after they have booted. This provides visibility into control plane node agent activity.kubelet.serviceRetrieve the logs using
:oc$ oc adm node-logs --role=master -u kubeletIf the API is not functional, review the logs using SSH instead. Replace
with appropriate values:<master-node>.<cluster_name>.<base_domain>$ ssh core@<master-node>.<cluster_name>.<base_domain> journalctl -b -f -u kubelet.serviceNoteOpenShift Container Platform 4.11 cluster nodes running Red Hat Enterprise Linux CoreOS (RHCOS) are immutable and rely on Operators to apply cluster changes. Accessing cluster nodes by using SSH is not recommended. Before attempting to collect diagnostic data over SSH, review whether the data collected by running
and otheroc adm must gathercommands is sufficient instead. However, if the OpenShift Container Platform API is not available, or the kubelet is not properly functioning on the target node,ocoperations will be impacted. In such situations, it is possible to access nodes usingoc.ssh core@<node>.<cluster_name>.<base_domain>
Retrieve
journald unit logs on control plane nodes, after they have booted. This provides visibility into control plane node CRI-O container runtime activity.crio.serviceRetrieve the logs using
:oc$ oc adm node-logs --role=master -u crioIf the API is not functional, review the logs using SSH instead:
$ ssh core@<master-node>.<cluster_name>.<base_domain> journalctl -b -f -u crio.service
Collect logs from specific subdirectories under
on control plane nodes./var/log/Retrieve a list of logs contained within a
subdirectory. The following example lists files in/var/log/on all control plane nodes:/var/log/openshift-apiserver/$ oc adm node-logs --role=master --path=openshift-apiserverInspect a specific log within a
subdirectory. The following example outputs/var/log/contents from all control plane nodes:/var/log/openshift-apiserver/audit.log$ oc adm node-logs --role=master --path=openshift-apiserver/audit.logIf the API is not functional, review the logs on each node using SSH instead. The following example tails
:/var/log/openshift-apiserver/audit.log$ ssh core@<master-node>.<cluster_name>.<base_domain> sudo tail -f /var/log/openshift-apiserver/audit.log
Review control plane node container logs using SSH.
List the containers:
$ ssh core@<master-node>.<cluster_name>.<base_domain> sudo crictl ps -aRetrieve a container’s logs using
:crictl$ ssh core@<master-node>.<cluster_name>.<base_domain> sudo crictl logs -f <container_id>
If you experience control plane node configuration issues, verify that the MCO, MCO endpoint, and DNS record are functioning. The Machine Config Operator (MCO) manages operating system configuration during the installation procedure. Also verify system clock accuracy and certificate validity.
Test whether the MCO endpoint is available. Replace
with appropriate values:<cluster_name>$ curl https://api-int.<cluster_name>:22623/config/master- If the endpoint is unresponsive, verify load balancer configuration. Ensure that the endpoint is configured to run on port 22623.
Verify that the MCO endpoint’s DNS record is configured and resolves to the load balancer.
Run a DNS lookup for the defined MCO endpoint name:
$ dig api-int.<cluster_name> @<dns_server>Run a reverse lookup to the assigned MCO IP address on the load balancer:
$ dig -x <load_balancer_mco_ip_address> @<dns_server>
Verify that the MCO is functioning from the bootstrap node directly. Replace
with the bootstrap node’s fully qualified domain name:<bootstrap_fqdn>$ ssh core@<bootstrap_fqdn> curl https://api-int.<cluster_name>:22623/config/masterSystem clock time must be synchronized between bootstrap, master, and worker nodes. Check each node’s system clock reference time and time synchronization statistics:
$ ssh core@<node>.<cluster_name>.<base_domain> chronyc trackingReview certificate validity:
$ openssl s_client -connect api-int.<cluster_name>:22623 | openssl x509 -noout -text
7.1.9. Investigating etcd installation issues Copia collegamentoCollegamento copiato negli appunti!
If you experience etcd issues during installation, you can check etcd pod status and collect etcd pod logs. You can also verify etcd DNS records and check DNS availability on control plane nodes.
Prerequisites
-
You have access to the cluster as a user with the role.
cluster-admin -
You have installed the OpenShift CLI ().
oc - You have SSH access to your hosts.
- You have the fully qualified domain names of the control plane nodes.
Procedure
Check the status of etcd pods.
Review the status of pods in the
namespace:openshift-etcd$ oc get pods -n openshift-etcdReview the status of pods in the
namespace:openshift-etcd-operator$ oc get pods -n openshift-etcd-operator
If any of the pods listed by the previous commands are not showing a
or aRunningstatus, gather diagnostic information for the pod.CompletedReview events for the pod:
$ oc describe pod/<pod_name> -n <namespace>Inspect the pod’s logs:
$ oc logs pod/<pod_name> -n <namespace>If the pod has more than one container, the preceding command will create an error, and the container names will be provided in the error message. Inspect logs for each container:
$ oc logs pod/<pod_name> -c <container_name> -n <namespace>
If the API is not functional, review etcd pod and container logs on each control plane node by using SSH instead. Replace
with appropriate values.<master-node>.<cluster_name>.<base_domain>List etcd pods on each control plane node:
$ ssh core@<master-node>.<cluster_name>.<base_domain> sudo crictl pods --name=etcd-For any pods not showing
status, inspect pod status in detail. ReplaceReadywith the pod’s ID listed in the output of the preceding command:<pod_id>$ ssh core@<master-node>.<cluster_name>.<base_domain> sudo crictl inspectp <pod_id>List containers related to a pod:
$ ssh core@<master-node>.<cluster_name>.<base_domain> sudo crictl ps | grep '<pod_id>'For any containers not showing
status, inspect container status in detail. ReplaceReadywith container IDs listed in the output of the preceding command:<container_id>$ ssh core@<master-node>.<cluster_name>.<base_domain> sudo crictl inspect <container_id>Review the logs for any containers not showing a
status. ReplaceReadywith the container IDs listed in the output of the preceding command:<container_id>$ ssh core@<master-node>.<cluster_name>.<base_domain> sudo crictl logs -f <container_id>NoteOpenShift Container Platform 4.11 cluster nodes running Red Hat Enterprise Linux CoreOS (RHCOS) are immutable and rely on Operators to apply cluster changes. Accessing cluster nodes by using SSH is not recommended. Before attempting to collect diagnostic data over SSH, review whether the data collected by running
and otheroc adm must gathercommands is sufficient instead. However, if the OpenShift Container Platform API is not available, or the kubelet is not properly functioning on the target node,ocoperations will be impacted. In such situations, it is possible to access nodes usingoc.ssh core@<node>.<cluster_name>.<base_domain>
- Validate primary and secondary DNS server connectivity from control plane nodes.
7.1.10. Investigating control plane node kubelet and API server issues Copia collegamentoCollegamento copiato negli appunti!
To investigate control plane node kubelet and API server issues during installation, check DNS, DHCP, and load balancer functionality. Also, verify that certificates have not expired.
Prerequisites
-
You have access to the cluster as a user with the role.
cluster-admin -
You have installed the OpenShift CLI ().
oc - You have SSH access to your hosts.
- You have the fully qualified domain names of the control plane nodes.
Procedure
-
Verify that the API server’s DNS record directs the kubelet on control plane nodes to . Ensure that the record references the load balancer.
https://api-int.<cluster_name>.<base_domain>:6443 - Ensure that the load balancer’s port 6443 definition references each control plane node.
- Check that unique control plane node hostnames have been provided by DHCP.
Inspect the
journald unit logs on each control plane node.kubelet.serviceRetrieve the logs using
:oc$ oc adm node-logs --role=master -u kubeletIf the API is not functional, review the logs using SSH instead. Replace
with appropriate values:<master-node>.<cluster_name>.<base_domain>$ ssh core@<master-node>.<cluster_name>.<base_domain> journalctl -b -f -u kubelet.serviceNoteOpenShift Container Platform 4.11 cluster nodes running Red Hat Enterprise Linux CoreOS (RHCOS) are immutable and rely on Operators to apply cluster changes. Accessing cluster nodes by using SSH is not recommended. Before attempting to collect diagnostic data over SSH, review whether the data collected by running
and otheroc adm must gathercommands is sufficient instead. However, if the OpenShift Container Platform API is not available, or the kubelet is not properly functioning on the target node,ocoperations will be impacted. In such situations, it is possible to access nodes usingoc.ssh core@<node>.<cluster_name>.<base_domain>
Check for certificate expiration messages in the control plane node kubelet logs.
Retrieve the log using
:oc$ oc adm node-logs --role=master -u kubelet | grep -is 'x509: certificate has expired'If the API is not functional, review the logs using SSH instead. Replace
with appropriate values:<master-node>.<cluster_name>.<base_domain>$ ssh core@<master-node>.<cluster_name>.<base_domain> journalctl -b -f -u kubelet.service | grep -is 'x509: certificate has expired'
7.1.11. Investigating worker node installation issues Copia collegamentoCollegamento copiato negli appunti!
If you experience worker node installation issues, you can review the worker node status. Collect
kubelet.service
crio.service
Prerequisites
-
You have access to the cluster as a user with the role.
cluster-admin -
You have installed the OpenShift CLI ().
oc - You have SSH access to your hosts.
- You have the fully qualified domain names of the bootstrap and worker nodes.
If you are hosting Ignition configuration files by using an HTTP server, you must have the HTTP server’s fully qualified domain name and the port number. You must also have SSH access to the HTTP host.
NoteThe initial
password can be found inkubeadminon the installation host.<install_directory>/auth/kubeadmin-password
Procedure
- If you have access to the worker node’s console, monitor the console until the node reaches the login prompt. During the installation, Ignition log messages are output to the console.
Verify Ignition file configuration.
If you are hosting Ignition configuration files by using an HTTP server.
Verify the worker node Ignition file URL. Replace
with HTTP server’s fully qualified domain name:<http_server_fqdn>$ curl -I http://<http_server_fqdn>:<port>/worker.ign1 - 1
- The
-Ioption returns the header only. If the Ignition file is available on the specified URL, the command returns200 OKstatus. If it is not available, the command returns404 file not found.
To verify that the Ignition file was received by the worker node, query the HTTP server logs on the HTTP host. For example, if you are using an Apache web server to serve Ignition files:
$ grep -is 'worker.ign' /var/log/httpd/access_logIf the worker Ignition file is received, the associated
log message will include aHTTP GETsuccess status, indicating that the request succeeded.200 OK- If the Ignition file was not received, check that it exists on the serving host directly. Ensure that the appropriate file and web server permissions are in place.
If you are using a cloud provider mechanism to inject Ignition configuration files into hosts as part of their initial deployment.
- Review the worker node’s console to determine if the mechanism is injecting the worker node Ignition file correctly.
- Check the availability of the worker node’s assigned storage device.
- Verify that the worker node has been assigned an IP address from the DHCP server.
Determine worker node status.
Query node status:
$ oc get nodesRetrieve a detailed node description for any worker nodes not showing a
status:Ready$ oc describe node <worker_node>NoteIt is not possible to run
commands if an installation issue prevents the OpenShift Container Platform API from running or if the kubelet is not running yet on each node.oc
Unlike control plane nodes, worker nodes are deployed and scaled using the Machine API Operator. Check the status of the Machine API Operator.
Review Machine API Operator pod status:
$ oc get pods -n openshift-machine-apiIf the Machine API Operator pod does not have a
status, detail the pod’s events:Ready$ oc describe pod/<machine_api_operator_pod_name> -n openshift-machine-apiInspect
container logs. The container runs within themachine-api-operatorpod:machine-api-operator$ oc logs pod/<machine_api_operator_pod_name> -n openshift-machine-api -c machine-api-operatorAlso inspect
container logs. The container also runs within thekube-rbac-proxypod:machine-api-operator$ oc logs pod/<machine_api_operator_pod_name> -n openshift-machine-api -c kube-rbac-proxy
Monitor
journald unit logs on worker nodes, after they have booted. This provides visibility into worker node agent activity.kubelet.serviceRetrieve the logs using
:oc$ oc adm node-logs --role=worker -u kubeletIf the API is not functional, review the logs using SSH instead. Replace
with appropriate values:<worker-node>.<cluster_name>.<base_domain>$ ssh core@<worker-node>.<cluster_name>.<base_domain> journalctl -b -f -u kubelet.serviceNoteOpenShift Container Platform 4.11 cluster nodes running Red Hat Enterprise Linux CoreOS (RHCOS) are immutable and rely on Operators to apply cluster changes. Accessing cluster nodes by using SSH is not recommended. Before attempting to collect diagnostic data over SSH, review whether the data collected by running
and otheroc adm must gathercommands is sufficient instead. However, if the OpenShift Container Platform API is not available, or the kubelet is not properly functioning on the target node,ocoperations will be impacted. In such situations, it is possible to access nodes usingoc.ssh core@<node>.<cluster_name>.<base_domain>
Retrieve
journald unit logs on worker nodes, after they have booted. This provides visibility into worker node CRI-O container runtime activity.crio.serviceRetrieve the logs using
:oc$ oc adm node-logs --role=worker -u crioIf the API is not functional, review the logs using SSH instead:
$ ssh core@<worker-node>.<cluster_name>.<base_domain> journalctl -b -f -u crio.service
Collect logs from specific subdirectories under
on worker nodes./var/log/Retrieve a list of logs contained within a
subdirectory. The following example lists files in/var/log/on all worker nodes:/var/log/sssd/$ oc adm node-logs --role=worker --path=sssdInspect a specific log within a
subdirectory. The following example outputs/var/log/contents from all worker nodes:/var/log/sssd/audit.log$ oc adm node-logs --role=worker --path=sssd/sssd.logIf the API is not functional, review the logs on each node using SSH instead. The following example tails
:/var/log/sssd/sssd.log$ ssh core@<worker-node>.<cluster_name>.<base_domain> sudo tail -f /var/log/sssd/sssd.log
Review worker node container logs using SSH.
List the containers:
$ ssh core@<worker-node>.<cluster_name>.<base_domain> sudo crictl ps -aRetrieve a container’s logs using
:crictl$ ssh core@<worker-node>.<cluster_name>.<base_domain> sudo crictl logs -f <container_id>
If you experience worker node configuration issues, verify that the MCO, MCO endpoint, and DNS record are functioning. The Machine Config Operator (MCO) manages operating system configuration during the installation procedure. Also verify system clock accuracy and certificate validity.
Test whether the MCO endpoint is available. Replace
with appropriate values:<cluster_name>$ curl https://api-int.<cluster_name>:22623/config/worker- If the endpoint is unresponsive, verify load balancer configuration. Ensure that the endpoint is configured to run on port 22623.
Verify that the MCO endpoint’s DNS record is configured and resolves to the load balancer.
Run a DNS lookup for the defined MCO endpoint name:
$ dig api-int.<cluster_name> @<dns_server>Run a reverse lookup to the assigned MCO IP address on the load balancer:
$ dig -x <load_balancer_mco_ip_address> @<dns_server>
Verify that the MCO is functioning from the bootstrap node directly. Replace
with the bootstrap node’s fully qualified domain name:<bootstrap_fqdn>$ ssh core@<bootstrap_fqdn> curl https://api-int.<cluster_name>:22623/config/workerSystem clock time must be synchronized between bootstrap, master, and worker nodes. Check each node’s system clock reference time and time synchronization statistics:
$ ssh core@<node>.<cluster_name>.<base_domain> chronyc trackingReview certificate validity:
$ openssl s_client -connect api-int.<cluster_name>:22623 | openssl x509 -noout -text
7.1.12. Querying Operator status after installation Copia collegamentoCollegamento copiato negli appunti!
You can check Operator status at the end of an installation. Retrieve diagnostic data for Operators that do not become available. Review logs for any Operator pods that are listed as
Pending
Prerequisites
-
You have access to the cluster as a user with the role.
cluster-admin -
You have installed the OpenShift CLI ().
oc
Procedure
Check that cluster Operators are all available at the end of an installation.
$ oc get clusteroperatorsVerify that all of the required certificate signing requests (CSRs) are approved. Some nodes might not move to a
status and some cluster Operators might not become available if there are pending CSRs.ReadyCheck the status of the CSRs and ensure that you see a client and server request with the
orPendingstatus for each machine that you added to the cluster:Approved$ oc get csrExample output
NAME AGE REQUESTOR CONDITION csr-8b2br 15m system:serviceaccount:openshift-machine-config-operator:node-bootstrapper Pending1 csr-8vnps 15m system:serviceaccount:openshift-machine-config-operator:node-bootstrapper Pending csr-bfd72 5m26s system:node:ip-10-0-50-126.us-east-2.compute.internal Pending2 csr-c57lv 5m26s system:node:ip-10-0-95-157.us-east-2.compute.internal Pending ...In this example, two machines are joining the cluster. You might see more approved CSRs in the list.
If the CSRs were not approved, after all of the pending CSRs for the machines you added are in
status, approve the CSRs for your cluster machines:PendingNoteBecause the CSRs rotate automatically, approve your CSRs within an hour of adding the machines to the cluster. If you do not approve them within an hour, the certificates will rotate, and more than two certificates will be present for each node. You must approve all of these certificates. After you approve the initial CSRs, the subsequent node client CSRs are automatically approved by the cluster
.kube-controller-managerNoteFor clusters running on platforms that are not machine API enabled, such as bare metal and other user-provisioned infrastructure, you must implement a method of automatically approving the kubelet serving certificate requests (CSRs). If a request is not approved, then the
,oc exec, andoc rshcommands cannot succeed, because a serving certificate is required when the API server connects to the kubelet. Any operation that contacts the Kubelet endpoint requires this certificate approval to be in place. The method must watch for new CSRs, confirm that the CSR was submitted by theoc logsservice account in thenode-bootstrapperorsystem:nodegroups, and confirm the identity of the node.system:adminTo approve them individually, run the following command for each valid CSR:
$ oc adm certificate approve <csr_name>1 - 1
<csr_name>is the name of a CSR from the list of current CSRs.
To approve all pending CSRs, run the following command:
$ oc get csr -o go-template='{{range .items}}{{if not .status}}{{.metadata.name}}{{"\n"}}{{end}}{{end}}' | xargs oc adm certificate approve
View Operator events:
$ oc describe clusteroperator <operator_name>Review Operator pod status within the Operator’s namespace:
$ oc get pods -n <operator_namespace>Obtain a detailed description for pods that do not have
status:Running$ oc describe pod/<operator_pod_name> -n <operator_namespace>Inspect pod logs:
$ oc logs pod/<operator_pod_name> -n <operator_namespace>When experiencing pod base image related issues, review base image status.
Obtain details of the base image used by a problematic pod:
$ oc get pod -o "jsonpath={range .status.containerStatuses[*]}{.name}{'\t'}{.state}{'\t'}{.image}{'\n'}{end}" <operator_pod_name> -n <operator_namespace>List base image release information:
$ oc adm release info <image_path>:<tag> --commits
7.1.13. Gathering logs from a failed installation Copia collegamentoCollegamento copiato negli appunti!
If you gave an SSH key to your installation program, you can gather data about your failed installation.
You use a different command to gather logs about an unsuccessful installation than to gather logs from a running cluster. If you must gather logs from a running cluster, use the
oc adm must-gather
Prerequisites
- Your OpenShift Container Platform installation failed before the bootstrap process finished. The bootstrap node is running and accessible through SSH.
-
The process is active on your computer, and you provided the same SSH key to both the
ssh-agentprocess and the installation program.ssh-agent - If you tried to install a cluster on infrastructure that you provisioned, you must have the fully qualified domain names of the bootstrap and control plane nodes.
Procedure
Generate the commands that are required to obtain the installation logs from the bootstrap and control plane machines:
If you used installer-provisioned infrastructure, change to the directory that contains the installation program and run the following command:
$ ./openshift-install gather bootstrap --dir <installation_directory>1 - 1
installation_directoryis the directory you specified when you ran./openshift-install create cluster. This directory contains the OpenShift Container Platform definition files that the installation program creates.
For installer-provisioned infrastructure, the installation program stores information about the cluster, so you do not specify the hostnames or IP addresses.
If you used infrastructure that you provisioned yourself, change to the directory that contains the installation program and run the following command:
$ ./openshift-install gather bootstrap --dir <installation_directory> \1 --bootstrap <bootstrap_address> \2 --master <master_1_address> \3 --master <master_2_address> \4 --master <master_3_address>"5 - 1
- For
installation_directory, specify the same directory you specified when you ran./openshift-install create cluster. This directory contains the OpenShift Container Platform definition files that the installation program creates. - 2
<bootstrap_address>is the fully qualified domain name or IP address of the cluster’s bootstrap machine.- 3 4 5
- For each control plane, or master, machine in your cluster, replace
<master_*_address>with its fully qualified domain name or IP address.
NoteA default cluster contains three control plane machines. List all of your control plane machines as shown, no matter how many your cluster uses.
Example output
INFO Pulling debug logs from the bootstrap machine INFO Bootstrap gather logs captured here "<installation_directory>/log-bundle-<timestamp>.tar.gz"If you open a Red Hat support case about your installation failure, include the compressed logs in the case.
7.2. Verifying node health Copia collegamentoCollegamento copiato negli appunti!
7.2.1. Reviewing node status, resource usage, and configuration Copia collegamentoCollegamento copiato negli appunti!
Review cluster node health status, resource consumption statistics, and node logs. Additionally, query
kubelet
Prerequisites
-
You have access to the cluster as a user with the role.
cluster-admin -
You have installed the OpenShift CLI ().
oc
Procedure
List the name, status, and role for all nodes in the cluster:
$ oc get nodesSummarize CPU and memory usage for each node within the cluster:
$ oc adm top nodesSummarize CPU and memory usage for a specific node:
$ oc adm top node my-node
7.2.2. Querying the kubelet’s status on a node Copia collegamentoCollegamento copiato negli appunti!
You can review cluster node health status, resource consumption statistics, and node logs. Additionally, you can query
kubelet
Prerequisites
-
You have access to the cluster as a user with the role.
cluster-admin - Your API service is still functional.
-
You have installed the OpenShift CLI ().
oc
Procedure
The kubelet is managed using a systemd service on each node. Review the kubelet’s status by querying the
systemd service within a debug pod.kubeletStart a debug pod for a node:
$ oc debug node/my-nodeNoteIf you are running
on a control plane node, you can find administrativeoc debugfiles in thekubeconfigdirectory./etc/kubernetes/static-pod-resources/kube-apiserver-certs/secrets/node-kubeconfigsSet
as the root directory within the debug shell. The debug pod mounts the host’s root file system in/hostwithin the pod. By changing the root directory to/host, you can run binaries contained in the host’s executable paths:/host# chroot /hostNoteOpenShift Container Platform 4.11 cluster nodes running Red Hat Enterprise Linux CoreOS (RHCOS) are immutable and rely on Operators to apply cluster changes. Accessing cluster nodes by using SSH is not recommended. However, if the OpenShift Container Platform API is not available, or
is not properly functioning on the target node,kubeletoperations will be impacted. In such situations, it is possible to access nodes usingocinstead.ssh core@<node>.<cluster_name>.<base_domain>Check whether the
systemd service is active on the node:kubelet# systemctl is-active kubeletOutput a more detailed
status summary:kubelet.service# systemctl status kubelet
7.2.3. Querying cluster node journal logs Copia collegamentoCollegamento copiato negli appunti!
You can gather
journald
/var/log
Prerequisites
-
You have access to the cluster as a user with the role.
cluster-admin - Your API service is still functional.
-
You have installed the OpenShift CLI ().
oc - You have SSH access to your hosts.
Procedure
Query
kubeletunit logs from OpenShift Container Platform cluster nodes. The following example queries control plane nodes only:journald$ oc adm node-logs --role=master -u kubelet1 - 1
- Replace
kubeletas appropriate to query other unit logs.
Collect logs from specific subdirectories under
on cluster nodes./var/log/Retrieve a list of logs contained within a
subdirectory. The following example lists files in/var/log/on all control plane nodes:/var/log/openshift-apiserver/$ oc adm node-logs --role=master --path=openshift-apiserverInspect a specific log within a
subdirectory. The following example outputs/var/log/contents from all control plane nodes:/var/log/openshift-apiserver/audit.log$ oc adm node-logs --role=master --path=openshift-apiserver/audit.logIf the API is not functional, review the logs on each node using SSH instead. The following example tails
:/var/log/openshift-apiserver/audit.log$ ssh core@<master-node>.<cluster_name>.<base_domain> sudo tail -f /var/log/openshift-apiserver/audit.logNoteOpenShift Container Platform 4.11 cluster nodes running Red Hat Enterprise Linux CoreOS (RHCOS) are immutable and rely on Operators to apply cluster changes. Accessing cluster nodes by using SSH is not recommended. Before attempting to collect diagnostic data over SSH, review whether the data collected by running
and otheroc adm must gathercommands is sufficient instead. However, if the OpenShift Container Platform API is not available, or the kubelet is not properly functioning on the target node,ocoperations will be impacted. In such situations, it is possible to access nodes usingoc.ssh core@<node>.<cluster_name>.<base_domain>
7.3. Troubleshooting CRI-O container runtime issues Copia collegamentoCollegamento copiato negli appunti!
7.3.1. About CRI-O container runtime engine Copia collegamentoCollegamento copiato negli appunti!
CRI-O is a Kubernetes-native container runtime implementation that integrates closely with the operating system to deliver an efficient and optimized Kubernetes experience. CRI-O provides facilities for running, stopping, and restarting containers.
The CRI-O container runtime engine is managed using a systemd service on each OpenShift Container Platform cluster node. When container runtime issues occur, verify the status of the
crio
7.3.2. Verifying CRI-O runtime engine status Copia collegamentoCollegamento copiato negli appunti!
You can verify CRI-O container runtime engine status on each cluster node.
Prerequisites
-
You have access to the cluster as a user with the role.
cluster-admin -
You have installed the OpenShift CLI ().
oc
Procedure
Review CRI-O status by querying the
systemd service on a node, within a debug pod.crioStart a debug pod for a node:
$ oc debug node/my-nodeSet
as the root directory within the debug shell. The debug pod mounts the host’s root file system in/hostwithin the pod. By changing the root directory to/host, you can run binaries contained in the host’s executable paths:/host# chroot /hostNoteOpenShift Container Platform 4.11 cluster nodes running Red Hat Enterprise Linux CoreOS (RHCOS) are immutable and rely on Operators to apply cluster changes. Accessing cluster nodes by using SSH is not recommended. However, if the OpenShift Container Platform API is not available, or the kubelet is not properly functioning on the target node,
operations will be impacted. In such situations, it is possible to access nodes usingocinstead.ssh core@<node>.<cluster_name>.<base_domain>Check whether the
systemd service is active on the node:crio# systemctl is-active crioOutput a more detailed
status summary:crio.service# systemctl status crio.service
7.3.3. Gathering CRI-O journald unit logs Copia collegamentoCollegamento copiato negli appunti!
If you experience CRI-O issues, you can obtain CRI-O journald unit logs from a node.
Prerequisites
-
You have access to the cluster as a user with the role.
cluster-admin - Your API service is still functional.
-
You have installed the OpenShift CLI ().
oc - You have the fully qualified domain names of the control plane or control plane machines.
Procedure
Gather CRI-O journald unit logs. The following example collects logs from all control plane nodes (within the cluster:
$ oc adm node-logs --role=master -u crioGather CRI-O journald unit logs from a specific node:
$ oc adm node-logs <node_name> -u crioIf the API is not functional, review the logs using SSH instead. Replace
with appropriate values:<node>.<cluster_name>.<base_domain>$ ssh core@<node>.<cluster_name>.<base_domain> journalctl -b -f -u crio.serviceNoteOpenShift Container Platform 4.11 cluster nodes running Red Hat Enterprise Linux CoreOS (RHCOS) are immutable and rely on Operators to apply cluster changes. Accessing cluster nodes by using SSH is not recommended. Before attempting to collect diagnostic data over SSH, review whether the data collected by running
and otheroc adm must gathercommands is sufficient instead. However, if the OpenShift Container Platform API is not available, or the kubelet is not properly functioning on the target node,ocoperations will be impacted. In such situations, it is possible to access nodes usingoc.ssh core@<node>.<cluster_name>.<base_domain>
7.3.4. Cleaning CRI-O storage Copia collegamentoCollegamento copiato negli appunti!
You can manually clear the CRI-O ephemeral storage if you experience the following issues:
A node cannot run on any pods and this error appears:
Failed to create pod sandbox: rpc error: code = Unknown desc = failed to mount container XXX: error recreating the missing symlinks: error reading name of symlink for XXX: open /var/lib/containers/storage/overlay/XXX/link: no such file or directoryYou cannot create a new container on a working node and the “can’t stat lower layer” error appears:
can't stat lower layer ... because it does not exist. Going through storage to recreate the missing symlinks.-
Your node is in the state after a cluster upgrade or if you attempt to reboot it.
NotReady -
The container runtime implementation () is not working properly.
crio -
You are unable to start a debug shell on the node using because the container runtime instance (
oc debug node/<node_name>) is not working.crio
Follow this process to completely wipe the CRI-O storage and resolve the errors.
Prerequisites:
-
You have access to the cluster as a user with the role.
cluster-admin -
You have installed the OpenShift CLI ().
oc
Procedure
Use
on the node. This is to avoid any workload getting scheduled if the node gets into thecordonstatus. You will know that scheduling is disabled whenReadyis in your Status section:SchedulingDisabled$ oc adm cordon <node_name>Drain the node as the cluster-admin user:
$ oc adm drain <node_name> --ignore-daemonsets --delete-emptydir-dataNoteThe
attribute of a pod or pod template controls the graceful termination period. This attribute defaults at 30 seconds, but can be customized for each application as necessary. If set to more than 90 seconds, the pod might be marked asterminationGracePeriodSecondsand fail to terminate successfully.SIGKILLedWhen the node returns, connect back to the node via SSH or Console. Then connect to the root user:
$ ssh core@node1.example.com $ sudo -iManually stop the kubelet:
# systemctl stop kubeletStop the containers and pods:
Use the following command to stop the pods that are not in the
. They must be removed first because their removal relies on the networking plugin pods, which are in theHostNetwork.HostNetwork.. for pod in $(crictl pods -q); do if [[ "$(crictl inspectp $pod | jq -r .status.linux.namespaces.options.network)" != "NODE" ]]; then crictl rmp -f $pod; fi; doneStop all other pods:
# crictl rmp -fa
Manually stop the crio services:
# systemctl stop crioAfter you run those commands, you can completely wipe the ephemeral storage:
# crio wipe -fStart the crio and kubelet service:
# systemctl start crio # systemctl start kubeletYou will know if the clean up worked if the crio and kubelet services are started, and the node is in the
status:Ready$ oc get nodesExample output
NAME STATUS ROLES AGE VERSION ci-ln-tkbxyft-f76d1-nvwhr-master-1 Ready, SchedulingDisabled master 133m v1.24.0Mark the node schedulable. You will know that the scheduling is enabled when
is no longer in status:SchedulingDisabled$ oc adm uncordon <node_name>Example output
NAME STATUS ROLES AGE VERSION ci-ln-tkbxyft-f76d1-nvwhr-master-1 Ready master 133m v1.24.0
7.4. Troubleshooting operating system issues Copia collegamentoCollegamento copiato negli appunti!
OpenShift Container Platform runs on RHCOS. You can follow these procedures to troubleshoot problems related to the operating system.
7.4.1. Investigating kernel crashes Copia collegamentoCollegamento copiato negli appunti!
The
kdump
kexec-tools
The
x86_64
The following table provides details about the support level of kdump for different architectures.
| Architecture | Support level |
|---|---|
|
|
GA |
|
|
TP |
|
|
TP |
|
|
TP |
Kdump support, for the preceding three architectures in the table, is a Technology Preview feature only. Technology Preview features are not supported with Red Hat production service level agreements (SLAs) and might not be functionally complete. Red Hat does not recommend using them in production. These features provide early access to upcoming product features, enabling customers to test functionality and provide feedback during the development process.
For more information about the support scope of Red Hat Technology Preview features, see Technology Preview Features Support Scope.
7.4.1.1. Enabling kdump Copia collegamentoCollegamento copiato negli appunti!
RHCOS ships with the
kexec-tools
kdump
Procedure
Perform the following steps to enable kdump on RHCOS.
To reserve memory for the crash kernel during the first kernel booting, provide kernel arguments by entering the following command:
# rpm-ostree kargs --append='crashkernel=256M'NoteFor the
platform, the recommended value forppc64leiscrashkernel.crashkernel=2G-4G:384M,4G-16G:512M,16G-64G:1G,64G-128G:2G,128G-:4GOptional: To write the crash dump over the network or to some other location, rather than to the default local
location, edit the/var/crashconfiguration file./etc/kdump.confNoteIf your node uses LUKS-encrypted devices, you must use network dumps as kdump does not support saving crash dumps to LUKS-encrypted devices.
For details on configuring the
service, see the comments inkdump,/etc/sysconfig/kdump, and the/etc/kdump.confmanual page. Also refer to the RHEL kdump documentation for further information on configuring the dump target.kdump.confEnable the
systemd service.kdump# systemctl enable kdump.serviceReboot your system.
# systemctl reboot-
Ensure that kdump has loaded a crash kernel by checking that the systemd service has started and exited successfully and that the command,
kdump.service, prints the valuecat /sys/kernel/kexec_crash_loaded.1
7.4.1.2. Enabling kdump on day-1 Copia collegamentoCollegamento copiato negli appunti!
The
kdump
kdump
kdump
- Less available RAM due to memory being reserved for the crash kernel.
- Node unavailability while the kernel is dumping the core.
- Additional storage space being used to store the crash dumps.
If you are aware of the downsides and trade-offs of having the
kdump
systemd
MachineConfig
MachineConfig
See "Customizing nodes" in the Installing
Procedure
Create a
MachineConfig
Create a Butane config file,
, that configures and enables kdump:99-worker-kdump.buvariant: openshift version: 4.11.0 metadata: name: 99-worker-kdump1 labels: machineconfiguration.openshift.io/role: worker2 openshift: kernel_arguments:3 - crashkernel=256M storage: files: - path: /etc/kdump.conf4 mode: 0644 overwrite: true contents: inline: | path /var/crash core_collector makedumpfile -l --message-level 7 -d 31 - path: /etc/sysconfig/kdump5 mode: 0644 overwrite: true contents: inline: | KDUMP_COMMANDLINE_REMOVE="hugepages hugepagesz slub_debug quiet log_buf_len swiotlb" KDUMP_COMMANDLINE_APPEND="irqpoll nr_cpus=1 reset_devices cgroup_disable=memory mce=off numa=off udev.children-max=2 panic=10 rootflags=nofail acpi_no_memhotplug transparent_hugepage=never nokaslr novmcoredd hest_disable"6 KEXEC_ARGS="-s" KDUMP_IMG="vmlinuz" systemd: units: - name: kdump.service enabled: true- 1 2
- Replace
workerwithmasterin both locations when creating aMachineConfigobject for control plane nodes. - 3
- Provide kernel arguments to reserve memory for the crash kernel. You can add other kernel arguments if necessary. For the
ppc64leplatform, the recommended value forcrashkerneliscrashkernel=2G-4G:384M,4G-16G:512M,16G-64G:1G,64G-128G:2G,128G-:4G. - 4
- If you want to change the contents of
/etc/kdump.conffrom the default, include this section and modify theinlinesubsection accordingly. - 5
- If you want to change the contents of
/etc/sysconfig/kdumpfrom the default, include this section and modify theinlinesubsection accordingly. - 6
- For the
ppc64leplatform, replacenr_cpus=1withmaxcpus=1, which is not supported on this platform.
Use Butane to generate a machine config YAML file,
, containing the configuration to be delivered to the nodes:99-worker-kdump.yaml$ butane 99-worker-kdump.bu -o 99-worker-kdump.yamlPut the YAML file into the
directory during cluster setup. You can also create this<installation_directory>/manifests/object after cluster setup with the YAML file:MachineConfig$ oc create -f 99-worker-kdump.yaml
7.4.1.3. Testing the kdump configuration Copia collegamentoCollegamento copiato negli appunti!
See the Testing the kdump configuration section in the RHEL documentation for kdump.
7.4.1.4. Analyzing a core dump Copia collegamentoCollegamento copiato negli appunti!
See the Analyzing a core dump section in the RHEL documentation for kdump.
It is recommended to perform vmcore analysis on a separate RHEL system.
Additional resources
- Setting up kdump in RHEL
- Linux kernel documentation for kdump
-
kdump.conf(5) — a manual page for the configuration file containing the full documentation of available options
/etc/kdump.conf -
kexec(8) — a manual page for the package
kexec - Red Hat Knowledgebase article regarding kexec and kdump
7.4.2. Debugging Ignition failures Copia collegamentoCollegamento copiato negli appunti!
If a machine cannot be provisioned, Ignition fails and RHCOS will boot into the emergency shell. Use the following procedure to get debugging information.
Procedure
Run the following command to show which service units failed:
$ systemctl --failedOptional: Run the following command on an individual service unit to find out more information:
$ journalctl -u <unit>.service
7.5. Troubleshooting network issues Copia collegamentoCollegamento copiato negli appunti!
7.5.1. How the network interface is selected Copia collegamentoCollegamento copiato negli appunti!
For installations on bare metal or with virtual machines that have more than one network interface controller (NIC), the NIC that OpenShift Container Platform uses for communication with the Kubernetes API server is determined by the
nodeip-configuration.service
After the
nodeip-configuration.service
/etc/systemd/system/kubelet.service.d/20-nodenet.conf
20-nodenet.conf
KUBELET_NODE_IP
When the kubelet service starts, it reads the value of the environment variable from the
20-nodenet.conf
--node-ip
If hardware or networking is reconfigured after installation, it is possible that the
nodeip-configuration.service
INTERNAL-IP
oc get nodes -o wide
If network communication is disrupted or misconfigured because a different NIC is selected, one strategy for overriding the selection process is to set the correct IP address explicitly. The following list identifies the high-level steps and considerations:
-
Create a shell script that determines the IP address to use for OpenShift Container Platform communication. Have the script create a custom unit file such as . Use the custom unit file,
/etc/systemd/system/kubelet.service.d/98-nodenet-override.conf, to set the98-nodenet-override.confenvironment variable to the IP address.KUBELET_NODE_IP -
Do not overwrite the file. Specify a file name with a numerically higher value such as
/etc/systemd/system/kubelet.service.d/20-nodenet.confin the same directory path. The goal is to have the custom unit file run after98-nodenet-override.confand override the value of the environment variable.20-nodenet.conf -
Create a machine config object with the shell script as a base64-encoded string and use the Machine Config Operator to deploy the script to the nodes at a file system path such as .
/usr/local/bin/override-node-ip.sh -
Ensure that runs after the shell script runs. The simplest method is to specify
systemctl daemon-reloadin the machine config, as shown in the following sample.ExecStart=systemctl daemon-reload
Sample machine config to override the network interface for kubelet
apiVersion: machineconfiguration.openshift.io/v1
kind: MachineConfig
metadata:
labels:
machineconfiguration.openshift.io/role: worker
name: 98-nodenet-override
spec:
config:
ignition:
version: 3.2.0
storage:
files:
- contents:
source: data:text/plain;charset=utf-8;base64,<encoded_script>
mode: 0755
overwrite: true
path: /usr/local/bin/override-node-ip.sh
systemd:
units:
- contents: |
[Unit]
Description=Override node IP detection
Wants=network-online.target
Before=kubelet.service
After=network-online.target
[Service]
Type=oneshot
ExecStart=/usr/local/bin/override-node-ip.sh
ExecStart=systemctl daemon-reload
[Install]
WantedBy=multi-user.target
enabled: true
name: nodenet-override.service
7.5.2. Troubleshooting Open vSwitch issues Copia collegamentoCollegamento copiato negli appunti!
To troubleshoot some Open vSwitch (OVS) issues, you might need to configure the log level to include more information.
If you modify the log level on a node temporarily, be aware that you can receive log messages from the machine config daemon on the node like the following example:
E0514 12:47:17.998892 2281 daemon.go:1350] content mismatch for file /etc/systemd/system/ovs-vswitchd.service: [Unit]
To avoid the log messages related to the mismatch, revert the log level change after you complete your troubleshooting.
7.5.2.1. Configuring the Open vSwitch log level temporarily Copia collegamentoCollegamento copiato negli appunti!
For short-term troubleshooting, you can configure the Open vSwitch (OVS) log level temporarily. The following procedure does not require rebooting the node. In addition, the configuration change does not persist whenever you reboot the node.
After you perform this procedure to change the log level, you can receive log messages from the machine config daemon that indicate a content mismatch for the
ovs-vswitchd.service
Prerequisites
-
You have access to the cluster as a user with the role.
cluster-admin -
You have installed the OpenShift CLI ().
oc
Procedure
Start a debug pod for a node:
$ oc debug node/<node_name>Set
as the root directory within the debug shell. The debug pod mounts the root file system from the host in/hostwithin the pod. By changing the root directory to/host, you can run binaries from the host file system:/host# chroot /hostView the current syslog level for OVS modules:
# ovs-appctl vlog/listThe following example output shows the log level for syslog set to
.infoExample output
console syslog file ------- ------ ------ backtrace OFF INFO INFO bfd OFF INFO INFO bond OFF INFO INFO bridge OFF INFO INFO bundle OFF INFO INFO bundles OFF INFO INFO cfm OFF INFO INFO collectors OFF INFO INFO command_line OFF INFO INFO connmgr OFF INFO INFO conntrack OFF INFO INFO conntrack_tp OFF INFO INFO coverage OFF INFO INFO ct_dpif OFF INFO INFO daemon OFF INFO INFO daemon_unix OFF INFO INFO dns_resolve OFF INFO INFO dpdk OFF INFO INFO ...Specify the log level in the
file:/etc/systemd/system/ovs-vswitchd.service.d/10-ovs-vswitchd-restart.confRestart=always ExecStartPre=-/bin/sh -c '/usr/bin/chown -R :$${OVS_USER_ID##*:} /var/lib/openvswitch' ExecStartPre=-/bin/sh -c '/usr/bin/chown -R :$${OVS_USER_ID##*:} /etc/openvswitch' ExecStartPre=-/bin/sh -c '/usr/bin/chown -R :$${OVS_USER_ID##*:} /run/openvswitch' ExecStartPost=-/usr/bin/ovs-appctl vlog/set syslog:dbg ExecReload=-/usr/bin/ovs-appctl vlog/set syslog:dbgIn the preceding example, the log level is set to
. Change the last two lines by settingdbgtosyslog:<log_level>,off,emer,err,warn, orinfo. Thedbglog level filters out all log messages.offRestart the service:
# systemctl daemon-reload# systemctl restart ovs-vswitchd
7.5.2.2. Configuring the Open vSwitch log level permanently Copia collegamentoCollegamento copiato negli appunti!
For long-term changes to the Open vSwitch (OVS) log level, you can change the log level permanently.
Prerequisites
-
You have access to the cluster as a user with the role.
cluster-admin -
You have installed the OpenShift CLI ().
oc
Procedure
Create a file, such as
, with a99-change-ovs-loglevel.yamlobject like the following example:MachineConfigapiVersion: machineconfiguration.openshift.io/v1 kind: MachineConfig metadata: labels: machineconfiguration.openshift.io/role: master1 name: 99-change-ovs-loglevel spec: config: ignition: version: 3.2.0 systemd: units: - dropins: - contents: | [Service] ExecStartPost=-/usr/bin/ovs-appctl vlog/set syslog:dbg2 ExecReload=-/usr/bin/ovs-appctl vlog/set syslog:dbg name: 20-ovs-vswitchd-restart.conf name: ovs-vswitchd.serviceApply the machine config:
$ oc apply -f 99-change-ovs-loglevel.yaml
7.5.2.3. Displaying Open vSwitch logs Copia collegamentoCollegamento copiato negli appunti!
Use the following procedure to display Open vSwitch (OVS) logs.
Prerequisites
-
You have access to the cluster as a user with the role.
cluster-admin -
You have installed the OpenShift CLI ().
oc
Procedure
Run one of the following commands:
Display the logs by using the
command from outside the cluster:oc$ oc adm node-logs <node_name> -u ovs-vswitchdDisplay the logs after logging on to a node in the cluster:
# journalctl -b -f -u ovs-vswitchd.serviceOne way to log on to a node is by using the
command.oc debug node/<node_name>
7.6. Troubleshooting Operator issues Copia collegamentoCollegamento copiato negli appunti!
Operators are a method of packaging, deploying, and managing an OpenShift Container Platform application. They act like an extension of the software vendor’s engineering team, watching over an OpenShift Container Platform environment and using its current state to make decisions in real time. Operators are designed to handle upgrades seamlessly, react to failures automatically, and not take shortcuts, such as skipping a software backup process to save time.
OpenShift Container Platform 4.11 includes a default set of Operators that are required for proper functioning of the cluster. These default Operators are managed by the Cluster Version Operator (CVO).
As a cluster administrator, you can install application Operators from the OperatorHub using the OpenShift Container Platform web console or the CLI. You can then subscribe the Operator to one or more namespaces to make it available for developers on your cluster. Application Operators are managed by Operator Lifecycle Manager (OLM).
If you experience Operator issues, verify Operator subscription status. Check Operator pod health across the cluster and gather Operator logs for diagnosis.
7.6.1. Operator subscription condition types Copia collegamentoCollegamento copiato negli appunti!
Subscriptions can report the following condition types:
| Condition | Description |
|---|---|
|
| Some or all of the catalog sources to be used in resolution are unhealthy. |
|
| An install plan for a subscription is missing. |
|
| An install plan for a subscription is pending installation. |
|
| An install plan for a subscription has failed. |
|
| The dependency resolution for a subscription has failed. |
Default OpenShift Container Platform cluster Operators are managed by the Cluster Version Operator (CVO) and they do not have a
Subscription
Subscription
7.6.2. Viewing Operator subscription status by using the CLI Copia collegamentoCollegamento copiato negli appunti!
You can view Operator subscription status by using the CLI.
Prerequisites
-
You have access to the cluster as a user with the role.
cluster-admin -
You have installed the OpenShift CLI ().
oc
Procedure
List Operator subscriptions:
$ oc get subs -n <operator_namespace>Use the
command to inspect aoc describeresource:Subscription$ oc describe sub <subscription_name> -n <operator_namespace>In the command output, find the
section for the status of Operator subscription condition types. In the following example, theConditionscondition type has a status ofCatalogSourcesUnhealthybecause all available catalog sources are healthy:falseExample output
Name: cluster-logging Namespace: openshift-logging Labels: operators.coreos.com/cluster-logging.openshift-logging= Annotations: <none> API Version: operators.coreos.com/v1alpha1 Kind: Subscription # ... Conditions: Last Transition Time: 2019-07-29T13:42:57Z Message: all available catalogsources are healthy Reason: AllCatalogSourcesHealthy Status: False Type: CatalogSourcesUnhealthy # ...
Default OpenShift Container Platform cluster Operators are managed by the Cluster Version Operator (CVO) and they do not have a
Subscription
Subscription
7.6.3. Viewing Operator catalog source status by using the CLI Copia collegamentoCollegamento copiato negli appunti!
You can view the status of an Operator catalog source by using the CLI.
Prerequisites
-
You have access to the cluster as a user with the role.
cluster-admin -
You have installed the OpenShift CLI ().
oc
Procedure
List the catalog sources in a namespace. For example, you can check the
namespace, which is used for cluster-wide catalog sources:openshift-marketplace$ oc get catalogsources -n openshift-marketplaceExample output
NAME DISPLAY TYPE PUBLISHER AGE certified-operators Certified Operators grpc Red Hat 55m community-operators Community Operators grpc Red Hat 55m example-catalog Example Catalog grpc Example Org 2m25s redhat-marketplace Red Hat Marketplace grpc Red Hat 55m redhat-operators Red Hat Operators grpc Red Hat 55mUse the
command to get more details and status about a catalog source:oc describe$ oc describe catalogsource example-catalog -n openshift-marketplaceExample output
Name: example-catalog Namespace: openshift-marketplace Labels: <none> Annotations: operatorframework.io/managed-by: marketplace-operator target.workload.openshift.io/management: {"effect": "PreferredDuringScheduling"} API Version: operators.coreos.com/v1alpha1 Kind: CatalogSource # ... Status: Connection State: Address: example-catalog.openshift-marketplace.svc:50051 Last Connect: 2021-09-09T17:07:35Z Last Observed State: TRANSIENT_FAILURE Registry Service: Created At: 2021-09-09T17:05:45Z Port: 50051 Protocol: grpc Service Name: example-catalog Service Namespace: openshift-marketplace # ...In the preceding example output, the last observed state is
. This state indicates that there is a problem establishing a connection for the catalog source.TRANSIENT_FAILUREList the pods in the namespace where your catalog source was created:
$ oc get pods -n openshift-marketplaceExample output
NAME READY STATUS RESTARTS AGE certified-operators-cv9nn 1/1 Running 0 36m community-operators-6v8lp 1/1 Running 0 36m marketplace-operator-86bfc75f9b-jkgbc 1/1 Running 0 42m example-catalog-bwt8z 0/1 ImagePullBackOff 0 3m55s redhat-marketplace-57p8c 1/1 Running 0 36m redhat-operators-smxx8 1/1 Running 0 36mWhen a catalog source is created in a namespace, a pod for the catalog source is created in that namespace. In the preceding example output, the status for the
pod isexample-catalog-bwt8z. This status indicates that there is an issue pulling the catalog source’s index image.ImagePullBackOffUse the
command to inspect a pod for more detailed information:oc describe$ oc describe pod example-catalog-bwt8z -n openshift-marketplaceExample output
Name: example-catalog-bwt8z Namespace: openshift-marketplace Priority: 0 Node: ci-ln-jyryyg2-f76d1-ggdbq-worker-b-vsxjd/10.0.128.2 ... Events: Type Reason Age From Message ---- ------ ---- ---- ------- Normal Scheduled 48s default-scheduler Successfully assigned openshift-marketplace/example-catalog-bwt8z to ci-ln-jyryyf2-f76d1-fgdbq-worker-b-vsxjd Normal AddedInterface 47s multus Add eth0 [10.131.0.40/23] from openshift-sdn Normal BackOff 20s (x2 over 46s) kubelet Back-off pulling image "quay.io/example-org/example-catalog:v1" Warning Failed 20s (x2 over 46s) kubelet Error: ImagePullBackOff Normal Pulling 8s (x3 over 47s) kubelet Pulling image "quay.io/example-org/example-catalog:v1" Warning Failed 8s (x3 over 47s) kubelet Failed to pull image "quay.io/example-org/example-catalog:v1": rpc error: code = Unknown desc = reading manifest v1 in quay.io/example-org/example-catalog: unauthorized: access to the requested resource is not authorized Warning Failed 8s (x3 over 47s) kubelet Error: ErrImagePullIn the preceding example output, the error messages indicate that the catalog source’s index image is failing to pull successfully because of an authorization issue. For example, the index image might be stored in a registry that requires login credentials.
7.6.4. Querying Operator pod status Copia collegamentoCollegamento copiato negli appunti!
You can list Operator pods within a cluster and their status. You can also collect a detailed Operator pod summary.
Prerequisites
-
You have access to the cluster as a user with the role.
cluster-admin - Your API service is still functional.
-
You have installed the OpenShift CLI ().
oc
Procedure
List Operators running in the cluster. The output includes Operator version, availability, and up-time information:
$ oc get clusteroperatorsList Operator pods running in the Operator’s namespace, plus pod status, restarts, and age:
$ oc get pod -n <operator_namespace>Output a detailed Operator pod summary:
$ oc describe pod <operator_pod_name> -n <operator_namespace>If an Operator issue is node-specific, query Operator container status on that node.
Start a debug pod for the node:
$ oc debug node/my-nodeSet
as the root directory within the debug shell. The debug pod mounts the host’s root file system in/hostwithin the pod. By changing the root directory to/host, you can run binaries contained in the host’s executable paths:/host# chroot /hostNoteOpenShift Container Platform 4.11 cluster nodes running Red Hat Enterprise Linux CoreOS (RHCOS) are immutable and rely on Operators to apply cluster changes. Accessing cluster nodes by using SSH is not recommended. However, if the OpenShift Container Platform API is not available, or the kubelet is not properly functioning on the target node,
operations will be impacted. In such situations, it is possible to access nodes usingocinstead.ssh core@<node>.<cluster_name>.<base_domain>List details about the node’s containers, including state and associated pod IDs:
# crictl psList information about a specific Operator container on the node. The following example lists information about the
container:network-operator# crictl ps --name network-operator- Exit from the debug shell.
7.6.5. Gathering Operator logs Copia collegamentoCollegamento copiato negli appunti!
If you experience Operator issues, you can gather detailed diagnostic information from Operator pod logs.
Prerequisites
-
You have access to the cluster as a user with the role.
cluster-admin - Your API service is still functional.
-
You have installed the OpenShift CLI ().
oc - You have the fully qualified domain names of the control plane or control plane machines.
Procedure
List the Operator pods that are running in the Operator’s namespace, plus the pod status, restarts, and age:
$ oc get pods -n <operator_namespace>Review logs for an Operator pod:
$ oc logs pod/<pod_name> -n <operator_namespace>If an Operator pod has multiple containers, the preceding command will produce an error that includes the name of each container. Query logs from an individual container:
$ oc logs pod/<operator_pod_name> -c <container_name> -n <operator_namespace>If the API is not functional, review Operator pod and container logs on each control plane node by using SSH instead. Replace
with appropriate values.<master-node>.<cluster_name>.<base_domain>List pods on each control plane node:
$ ssh core@<master-node>.<cluster_name>.<base_domain> sudo crictl podsFor any Operator pods not showing a
status, inspect the pod’s status in detail. ReplaceReadywith the Operator pod’s ID listed in the output of the preceding command:<operator_pod_id>$ ssh core@<master-node>.<cluster_name>.<base_domain> sudo crictl inspectp <operator_pod_id>List containers related to an Operator pod:
$ ssh core@<master-node>.<cluster_name>.<base_domain> sudo crictl ps --pod=<operator_pod_id>For any Operator container not showing a
status, inspect the container’s status in detail. ReplaceReadywith a container ID listed in the output of the preceding command:<container_id>$ ssh core@<master-node>.<cluster_name>.<base_domain> sudo crictl inspect <container_id>Review the logs for any Operator containers not showing a
status. ReplaceReadywith a container ID listed in the output of the preceding command:<container_id>$ ssh core@<master-node>.<cluster_name>.<base_domain> sudo crictl logs -f <container_id>NoteOpenShift Container Platform 4.11 cluster nodes running Red Hat Enterprise Linux CoreOS (RHCOS) are immutable and rely on Operators to apply cluster changes. Accessing cluster nodes by using SSH is not recommended. Before attempting to collect diagnostic data over SSH, review whether the data collected by running
and otheroc adm must gathercommands is sufficient instead. However, if the OpenShift Container Platform API is not available, or the kubelet is not properly functioning on the target node,ocoperations will be impacted. In such situations, it is possible to access nodes usingoc.ssh core@<node>.<cluster_name>.<base_domain>
7.6.6. Disabling the Machine Config Operator from automatically rebooting Copia collegamentoCollegamento copiato negli appunti!
When configuration changes are made by the Machine Config Operator (MCO), Red Hat Enterprise Linux CoreOS (RHCOS) must reboot for the changes to take effect. Whether the configuration change is automatic or manual, an RHCOS node reboots automatically unless it is paused.
The following modifications do not trigger a node reboot:
When the MCO detects any of the following changes, it applies the update without draining or rebooting the node:
-
Changes to the SSH key in the parameter of a machine config.
spec.config.passwd.users.sshAuthorizedKeys -
Changes to the global pull secret or pull secret in the namespace.
openshift-config -
Automatic rotation of the certificate authority (CA) by the Kubernetes API Server Operator.
/etc/kubernetes/kubelet-ca.crt
-
Changes to the SSH key in the
When the MCO detects changes to the
file, such as adding or editing an/etc/containers/registries.conf(ICSP) object, it drains the corresponding nodes, applies the changes, and uncordons the nodes.The node drain does not happen for the following changes:ImageContentSourcePolicy-
The addition of a registry with the parameter set for each mirror.
pull-from-mirror = "digest-only" -
The addition of a mirror with the parameter set in a registry.
pull-from-mirror = "digest-only" -
The addition of items to the list.
unqualified-search-registries
-
The addition of a registry with the
To avoid unwanted disruptions, you can modify the machine config pool (MCP) to prevent automatic rebooting after the Operator makes changes to the machine config.
Pausing an MCP prevents the MCO from applying any configuration changes on the associated nodes. Pausing an MCP also prevents any automatically rotated certificates from being pushed to the associated nodes, including the automatic rotation of the
kube-apiserver-to-kubelet-signer
If the MCP is paused when the
kube-apiserver-to-kubelet-signer
oc
oc debug
oc logs
oc exec
oc attach
Pausing an MCP should be done with careful consideration about the
kube-apiserver-to-kubelet-signer
New CA certificates are generated at 292 days from the installation date and removed at 365 days from that date. To determine the next automatic CA certificate rotation, see the Understand CA cert auto renewal in Red Hat OpenShift 4.
7.6.6.1. Disabling the Machine Config Operator from automatically rebooting by using the console Copia collegamentoCollegamento copiato negli appunti!
To avoid unwanted disruptions from changes made by the Machine Config Operator (MCO), you can use the OpenShift Container Platform web console to modify the machine config pool (MCP) to prevent the MCO from making any changes to nodes in that pool. This prevents any reboots that would normally be part of the MCO update process.
Prerequisites
-
You have access to the cluster as a user with the role.
cluster-admin
Procedure
To pause or unpause automatic MCO update rebooting:
Pause the autoreboot process:
-
Log in to the OpenShift Container Platform web console as a user with the role.
cluster-admin -
Click Compute
MachineConfigPools. - On the MachineConfigPools page, click either master or worker, depending upon which nodes you want to pause rebooting for.
- On the master or worker page, click YAML.
In the YAML, update the
field tospec.paused.trueSample MachineConfigPool object
apiVersion: machineconfiguration.openshift.io/v1 kind: MachineConfigPool # ... spec: # ... paused: true1 # ...- 1
- Update the
spec.pausedfield totrueto pause rebooting.
To verify that the MCP is paused, return to the MachineConfigPools page.
On the MachineConfigPools page, the Paused column reports True for the MCP you modified.
If the MCP has pending changes while paused, the Updated column is False and Updating is False. When Updated is True and Updating is False, there are no pending changes.
ImportantIf there are pending changes (where both the Updated and Updating columns are False), it is recommended to schedule a maintenance window for a reboot as early as possible. Use the following steps for unpausing the autoreboot process to apply the changes that were queued since the last reboot.
-
Log in to the OpenShift Container Platform web console as a user with the
Unpause the autoreboot process:
-
Log in to the OpenShift Container Platform web console as a user with the role.
cluster-admin -
Click Compute
MachineConfigPools. - On the MachineConfigPools page, click either master or worker, depending upon which nodes you want to pause rebooting for.
- On the master or worker page, click YAML.
In the YAML, update the
field tospec.paused.falseSample MachineConfigPool object
apiVersion: machineconfiguration.openshift.io/v1 kind: MachineConfigPool # ... spec: # ... paused: false1 # ...- 1
- Update the
spec.pausedfield tofalseto allow rebooting.
NoteBy unpausing an MCP, the MCO applies all paused changes reboots Red Hat Enterprise Linux CoreOS (RHCOS) as needed.
To verify that the MCP is paused, return to the MachineConfigPools page.
On the MachineConfigPools page, the Paused column reports False for the MCP you modified.
If the MCP is applying any pending changes, the Updated column is False and the Updating column is True. When Updated is True and Updating is False, there are no further changes being made.
-
Log in to the OpenShift Container Platform web console as a user with the
7.6.6.2. Disabling the Machine Config Operator from automatically rebooting by using the CLI Copia collegamentoCollegamento copiato negli appunti!
To avoid unwanted disruptions from changes made by the Machine Config Operator (MCO), you can modify the machine config pool (MCP) using the OpenShift CLI (oc) to prevent the MCO from making any changes to nodes in that pool. This prevents any reboots that would normally be part of the MCO update process.
Prerequisites
-
You have access to the cluster as a user with the role.
cluster-admin -
You have installed the OpenShift CLI ().
oc
Procedure
To pause or unpause automatic MCO update rebooting:
Pause the autoreboot process:
Update the
custom resource to set theMachineConfigPoolfield tospec.paused.trueControl plane (master) nodes
$ oc patch --type=merge --patch='{"spec":{"paused":true}}' machineconfigpool/masterWorker nodes
$ oc patch --type=merge --patch='{"spec":{"paused":true}}' machineconfigpool/workerVerify that the MCP is paused:
Control plane (master) nodes
$ oc get machineconfigpool/master --template='{{.spec.paused}}'Worker nodes
$ oc get machineconfigpool/worker --template='{{.spec.paused}}'Example output
trueThe
field isspec.pausedand the MCP is paused.trueDetermine if the MCP has pending changes:
# oc get machineconfigpoolExample output
NAME CONFIG UPDATED UPDATING master rendered-master-33cf0a1254318755d7b48002c597bf91 True False worker rendered-worker-e405a5bdb0db1295acea08bcca33fa60 False FalseIf the UPDATED column is False and UPDATING is False, there are pending changes. When UPDATED is True and UPDATING is False, there are no pending changes. In the previous example, the worker node has pending changes. The control plane node does not have any pending changes.
ImportantIf there are pending changes (where both the Updated and Updating columns are False), it is recommended to schedule a maintenance window for a reboot as early as possible. Use the following steps for unpausing the autoreboot process to apply the changes that were queued since the last reboot.
Unpause the autoreboot process:
Update the
custom resource to set theMachineConfigPoolfield tospec.paused.falseControl plane (master) nodes
$ oc patch --type=merge --patch='{"spec":{"paused":false}}' machineconfigpool/masterWorker nodes
$ oc patch --type=merge --patch='{"spec":{"paused":false}}' machineconfigpool/workerNoteBy unpausing an MCP, the MCO applies all paused changes and reboots Red Hat Enterprise Linux CoreOS (RHCOS) as needed.
Verify that the MCP is unpaused:
Control plane (master) nodes
$ oc get machineconfigpool/master --template='{{.spec.paused}}'Worker nodes
$ oc get machineconfigpool/worker --template='{{.spec.paused}}'Example output
falseThe
field isspec.pausedand the MCP is unpaused.falseDetermine if the MCP has pending changes:
$ oc get machineconfigpoolExample output
NAME CONFIG UPDATED UPDATING master rendered-master-546383f80705bd5aeaba93 True False worker rendered-worker-b4c51bb33ccaae6fc4a6a5 False TrueIf the MCP is applying any pending changes, the UPDATED column is False and the UPDATING column is True. When UPDATED is True and UPDATING is False, there are no further changes being made. In the previous example, the MCO is updating the worker node.
7.6.7. Refreshing failing subscriptions Copia collegamentoCollegamento copiato negli appunti!
In Operator Lifecycle Manager (OLM), if you subscribe to an Operator that references images that are not accessible on your network, you can find jobs in the
openshift-marketplace
Example output
ImagePullBackOff for
Back-off pulling image "example.com/openshift4/ose-elasticsearch-operator-bundle@sha256:6d2587129c846ec28d384540322b40b05833e7e00b25cca584e004af9a1d292e"
Example output
rpc error: code = Unknown desc = error pinging docker registry example.com: Get "https://example.com/v2/": dial tcp: lookup example.com on 10.0.0.1:53: no such host
As a result, the subscription is stuck in this failing state and the Operator is unable to install or upgrade.
You can refresh a failing subscription by deleting the subscription, cluster service version (CSV), and other related objects. After recreating the subscription, OLM then reinstalls the correct version of the Operator.
Prerequisites
- You have a failing subscription that is unable to pull an inaccessible bundle image.
- You have confirmed that the correct bundle image is accessible.
Procedure
Get the names of the
andSubscriptionobjects from the namespace where the Operator is installed:ClusterServiceVersion$ oc get sub,csv -n <namespace>Example output
NAME PACKAGE SOURCE CHANNEL subscription.operators.coreos.com/elasticsearch-operator elasticsearch-operator redhat-operators 5.0 NAME DISPLAY VERSION REPLACES PHASE clusterserviceversion.operators.coreos.com/elasticsearch-operator.5.0.0-65 OpenShift Elasticsearch Operator 5.0.0-65 SucceededDelete the subscription:
$ oc delete subscription <subscription_name> -n <namespace>Delete the cluster service version:
$ oc delete csv <csv_name> -n <namespace>Get the names of any failing jobs and related config maps in the
namespace:openshift-marketplace$ oc get job,configmap -n openshift-marketplaceExample output
NAME COMPLETIONS DURATION AGE job.batch/1de9443b6324e629ddf31fed0a853a121275806170e34c926d69e53a7fcbccb 1/1 26s 9m30s NAME DATA AGE configmap/1de9443b6324e629ddf31fed0a853a121275806170e34c926d69e53a7fcbccb 3 9m30sDelete the job:
$ oc delete job <job_name> -n openshift-marketplaceThis ensures pods that try to pull the inaccessible image are not recreated.
Delete the config map:
$ oc delete configmap <configmap_name> -n openshift-marketplace- Reinstall the Operator using OperatorHub in the web console.
Verification
Check that the Operator has been reinstalled successfully:
$ oc get sub,csv,installplan -n <namespace>
7.6.8. Reinstalling Operators after failed uninstallation Copia collegamentoCollegamento copiato negli appunti!
You must successfully and completely uninstall an Operator prior to attempting to reinstall the same Operator. Failure to fully uninstall the Operator properly can leave resources, such as a project or namespace, stuck in a "Terminating" state and cause "error resolving resource" messages. For example:
Example Project resource description
...
message: 'Failed to delete all resource types, 1 remaining: Internal error occurred:
error resolving resource'
...
These types of issues can prevent an Operator from being reinstalled successfully.
Forced deletion of a namespace is not likely to resolve "Terminating" state issues and can lead to unstable or unpredictable cluster behavior, so it is better to try to find related resources that might be preventing the namespace from being deleted. For more information, see the Red Hat Knowledgebase Solution #4165791, paying careful attention to the cautions and warnings.
The following procedure shows how to troubleshoot when an Operator cannot be reinstalled because an existing custom resource definition (CRD) from a previous installation of the Operator is preventing a related namespace from deleting successfully.
Procedure
Check if there are any namespaces related to the Operator that are stuck in "Terminating" state:
$ oc get namespacesExample output
operator-ns-1 TerminatingCheck if there are any CRDs related to the Operator that are still present after the failed uninstallation:
$ oc get crdsNoteCRDs are global cluster definitions; the actual custom resource (CR) instances related to the CRDs could be in other namespaces or be global cluster instances.
If there are any CRDs that you know were provided or managed by the Operator and that should have been deleted after uninstallation, delete the CRD:
$ oc delete crd <crd_name>Check if there are any remaining CR instances related to the Operator that are still present after uninstallation, and if so, delete the CRs:
The type of CRs to search for can be difficult to determine after uninstallation and can require knowing what CRDs the Operator manages. For example, if you are troubleshooting an uninstallation of the etcd Operator, which provides the
CRD, you can search for remainingEtcdClusterCRs in a namespace:EtcdCluster$ oc get EtcdCluster -n <namespace_name>Alternatively, you can search across all namespaces:
$ oc get EtcdCluster --all-namespacesIf there are any remaining CRs that should be removed, delete the instances:
$ oc delete <cr_name> <cr_instance_name> -n <namespace_name>
Check that the namespace deletion has successfully resolved:
$ oc get namespace <namespace_name>ImportantIf the namespace or other Operator resources are still not uninstalled cleanly, contact Red Hat Support.
- Reinstall the Operator using OperatorHub in the web console.
Verification
Check that the Operator has been reinstalled successfully:
$ oc get sub,csv,installplan -n <namespace>
7.7. Investigating pod issues Copia collegamentoCollegamento copiato negli appunti!
OpenShift Container Platform leverages the Kubernetes concept of a pod, which is one or more containers deployed together on one host. A pod is the smallest compute unit that can be defined, deployed, and managed on OpenShift Container Platform 4.11.
After a pod is defined, it is assigned to run on a node until its containers exit, or until it is removed. Depending on policy and exit code, Pods are either removed after exiting or retained so that their logs can be accessed.
The first thing to check when pod issues arise is the pod’s status. If an explicit pod failure has occurred, observe the pod’s error state to identify specific image, container, or pod network issues. Focus diagnostic data collection according to the error state. Review pod event messages, as well as pod and container log information. Diagnose issues dynamically by accessing running Pods on the command line, or start a debug pod with root access based on a problematic pod’s deployment configuration.
7.7.1. Understanding pod error states Copia collegamentoCollegamento copiato negli appunti!
Pod failures return explicit error states that can be observed in the
status
oc get pods
The following table provides a list of pod error states along with their descriptions.
| Pod error state | Description |
|---|---|
|
| Generic image retrieval error. |
|
| Image retrieval failed and is backed off. |
|
| The specified image name was invalid. |
|
| Image inspection did not succeed. |
|
|
|
|
| When attempting to retrieve an image from a registry, an HTTP error was encountered. |
|
| The specified container is either not present or not managed by the kubelet, within the declared pod. |
|
| Container initialization failed. |
|
| None of the pod’s containers started successfully. |
|
| None of the pod’s containers were killed successfully. |
|
| A container has terminated. The kubelet will not attempt to restart it. |
|
| A container or image attempted to run with root privileges. |
|
| Pod sandbox creation did not succeed. |
|
| Pod sandbox configuration was not obtained. |
|
| A pod sandbox did not stop successfully. |
|
| Network initialization failed. |
|
| Network termination failed. |
7.7.2. Reviewing pod status Copia collegamentoCollegamento copiato negli appunti!
You can query pod status and error states. You can also query a pod’s associated deployment configuration and review base image availability.
Prerequisites
-
You have access to the cluster as a user with the role.
cluster-admin -
You have installed the OpenShift CLI ().
oc -
is installed.
skopeo
Procedure
Switch into a project:
$ oc project <project_name>List pods running within the namespace, as well as pod status, error states, restarts, and age:
$ oc get podsDetermine whether the namespace is managed by a deployment configuration:
$ oc statusIf the namespace is managed by a deployment configuration, the output includes the deployment configuration name and a base image reference.
Inspect the base image referenced in the preceding command’s output:
$ skopeo inspect docker://<image_reference>If the base image reference is not correct, update the reference in the deployment configuration:
$ oc edit deployment/my-deploymentWhen deployment configuration changes on exit, the configuration will automatically redeploy. Watch pod status as the deployment progresses, to determine whether the issue has been resolved:
$ oc get pods -wReview events within the namespace for diagnostic information relating to pod failures:
$ oc get events
7.7.3. Inspecting pod and container logs Copia collegamentoCollegamento copiato negli appunti!
You can inspect pod and container logs for warnings and error messages related to explicit pod failures. Depending on policy and exit code, pod and container logs remain available after pods have been terminated.
Prerequisites
-
You have access to the cluster as a user with the role.
cluster-admin - Your API service is still functional.
-
You have installed the OpenShift CLI ().
oc
Procedure
Query logs for a specific pod:
$ oc logs <pod_name>Query logs for a specific container within a pod:
$ oc logs <pod_name> -c <container_name>Logs retrieved using the preceding
commands are composed of messages sent to stdout within pods or containers.oc logsInspect logs contained in
within a pod./var/log/List log files and subdirectories contained in
within a pod:/var/log$ oc exec <pod_name> -- ls -alh /var/logExample output
total 124K drwxr-xr-x. 1 root root 33 Aug 11 11:23 . drwxr-xr-x. 1 root root 28 Sep 6 2022 .. -rw-rw----. 1 root utmp 0 Jul 10 10:31 btmp -rw-r--r--. 1 root root 33K Jul 17 10:07 dnf.librepo.log -rw-r--r--. 1 root root 69K Jul 17 10:07 dnf.log -rw-r--r--. 1 root root 8.8K Jul 17 10:07 dnf.rpm.log -rw-r--r--. 1 root root 480 Jul 17 10:07 hawkey.log -rw-rw-r--. 1 root utmp 0 Jul 10 10:31 lastlog drwx------. 2 root root 23 Aug 11 11:14 openshift-apiserver drwx------. 2 root root 6 Jul 10 10:31 private drwxr-xr-x. 1 root root 22 Mar 9 08:05 rhsm -rw-rw-r--. 1 root utmp 0 Jul 10 10:31 wtmpQuery a specific log file contained in
within a pod:/var/log$ oc exec <pod_name> cat /var/log/<path_to_log>Example output
2023-07-10T10:29:38+0000 INFO --- logging initialized --- 2023-07-10T10:29:38+0000 DDEBUG timer: config: 13 ms 2023-07-10T10:29:38+0000 DEBUG Loaded plugins: builddep, changelog, config-manager, copr, debug, debuginfo-install, download, generate_completion_cache, groups-manager, needs-restarting, playground, product-id, repoclosure, repodiff, repograph, repomanage, reposync, subscription-manager, uploadprofile 2023-07-10T10:29:38+0000 INFO Updating Subscription Management repositories. 2023-07-10T10:29:38+0000 INFO Unable to read consumer identity 2023-07-10T10:29:38+0000 INFO Subscription Manager is operating in container mode. 2023-07-10T10:29:38+0000 INFOList log files and subdirectories contained in
within a specific container:/var/log$ oc exec <pod_name> -c <container_name> ls /var/logQuery a specific log file contained in
within a specific container:/var/log$ oc exec <pod_name> -c <container_name> cat /var/log/<path_to_log>
7.7.4. Accessing running pods Copia collegamentoCollegamento copiato negli appunti!
You can review running pods dynamically by opening a shell inside a pod or by gaining network access through port forwarding.
Prerequisites
-
You have access to the cluster as a user with the role.
cluster-admin - Your API service is still functional.
-
You have installed the OpenShift CLI ().
oc
Procedure
Switch into the project that contains the pod you would like to access. This is necessary because the
command does not accept theoc rshnamespace option:-n$ oc project <namespace>Start a remote shell into a pod:
$ oc rsh <pod_name>1 - 1
- If a pod has multiple containers,
oc rshdefaults to the first container unless-c <container_name>is specified.
Start a remote shell into a specific container within a pod:
$ oc rsh -c <container_name> pod/<pod_name>Create a port forwarding session to a port on a pod:
$ oc port-forward <pod_name> <host_port>:<pod_port>1 - 1
- Enter
Ctrl+Cto cancel the port forwarding session.
7.7.5. Starting debug pods with root access Copia collegamentoCollegamento copiato negli appunti!
You can start a debug pod with root access, based on a problematic pod’s deployment or deployment configuration. Pod users typically run with non-root privileges, but running troubleshooting pods with temporary root privileges can be useful during issue investigation.
Prerequisites
-
You have access to the cluster as a user with the role.
cluster-admin - Your API service is still functional.
-
You have installed the OpenShift CLI ().
oc
Procedure
Start a debug pod with root access, based on a deployment.
Obtain a project’s deployment name:
$ oc get deployment -n <project_name>Start a debug pod with root privileges, based on the deployment:
$ oc debug deployment/my-deployment --as-root -n <project_name>
Start a debug pod with root access, based on a deployment configuration.
Obtain a project’s deployment configuration name:
$ oc get deploymentconfigs -n <project_name>Start a debug pod with root privileges, based on the deployment configuration:
$ oc debug deploymentconfig/my-deployment-configuration --as-root -n <project_name>
You can append
-- <command>
oc debug
7.7.6. Copying files to and from pods and containers Copia collegamentoCollegamento copiato negli appunti!
You can copy files to and from a pod to test configuration changes or gather diagnostic information.
Prerequisites
-
You have access to the cluster as a user with the role.
cluster-admin - Your API service is still functional.
-
You have installed the OpenShift CLI ().
oc
Procedure
Copy a file to a pod:
$ oc cp <local_path> <pod_name>:/<path> -c <container_name>1 - 1
- The first container in a pod is selected if the
-coption is not specified.
Copy a file from a pod:
$ oc cp <pod_name>:/<path> -c <container_name> <local_path>1 - 1
- The first container in a pod is selected if the
-coption is not specified.
NoteFor
to function, theoc cpbinary must be available within the container.tar
7.8. Troubleshooting the Source-to-Image process Copia collegamentoCollegamento copiato negli appunti!
7.8.1. Strategies for Source-to-Image troubleshooting Copia collegamentoCollegamento copiato negli appunti!
Use Source-to-Image (S2I) to build reproducible, Docker-formatted container images. You can create ready-to-run images by injecting application source code into a container image and assembling a new image. The new image incorporates the base image (the builder) and built source.
To determine where in the S2I process a failure occurs, you can observe the state of the pods relating to each of the following S2I stages:
- During the build configuration stage, a build pod is used to create an application container image from a base image and application source code.
- During the deployment configuration stage, a deployment pod is used to deploy application pods from the application container image that was built in the build configuration stage. The deployment pod also deploys other resources such as services and routes. The deployment configuration begins after the build configuration succeeds.
-
After the deployment pod has started the application pods, application failures can occur within the running application pods. For instance, an application might not behave as expected even though the application pods are in a state. In this scenario, you can access running application pods to investigate application failures within a pod.
Running
When troubleshooting S2I issues, follow this strategy:
- Monitor build, deployment, and application pod status
- Determine the stage of the S2I process where the problem occurred
- Review logs corresponding to the failed stage
7.8.2. Gathering Source-to-Image diagnostic data Copia collegamentoCollegamento copiato negli appunti!
The S2I tool runs a build pod and a deployment pod in sequence. The deployment pod is responsible for deploying the application pods based on the application container image created in the build stage. Watch build, deployment and application pod status to determine where in the S2I process a failure occurs. Then, focus diagnostic data collection accordingly.
Prerequisites
-
You have access to the cluster as a user with the role.
cluster-admin - Your API service is still functional.
-
You have installed the OpenShift CLI ().
oc
Procedure
Watch the pod status throughout the S2I process to determine at which stage a failure occurs:
$ oc get pods -w1 - 1
- Use
-wto monitor pods for changes until you quit the command usingCtrl+C.
Review a failed pod’s logs for errors.
If the build pod fails, review the build pod’s logs:
$ oc logs -f pod/<application_name>-<build_number>-buildNoteAlternatively, you can review the build configuration’s logs using
. The build configuration’s logs include the logs from the build pod.oc logs -f bc/<application_name>If the deployment pod fails, review the deployment pod’s logs:
$ oc logs -f pod/<application_name>-<build_number>-deployNoteAlternatively, you can review the deployment configuration’s logs using
. This outputs logs from the deployment pod until the deployment pod completes successfully. The command outputs logs from the application pods if you run it after the deployment pod has completed. After a deployment pod completes, its logs can still be accessed by runningoc logs -f dc/<application_name>.oc logs -f pod/<application_name>-<build_number>-deployIf an application pod fails, or if an application is not behaving as expected within a running application pod, review the application pod’s logs:
$ oc logs -f pod/<application_name>-<build_number>-<random_string>
7.8.3. Gathering application diagnostic data to investigate application failures Copia collegamentoCollegamento copiato negli appunti!
Application failures can occur within running application pods. In these situations, you can retrieve diagnostic information with these strategies:
- Review events relating to the application pods.
- Review the logs from the application pods, including application-specific log files that are not collected by the OpenShift Logging framework.
- Test application functionality interactively and run diagnostic tools in an application container.
Prerequisites
-
You have access to the cluster as a user with the role.
cluster-admin -
You have installed the OpenShift CLI ().
oc
Procedure
List events relating to a specific application pod. The following example retrieves events for an application pod named
:my-app-1-akdlg$ oc describe pod/my-app-1-akdlgReview logs from an application pod:
$ oc logs -f pod/my-app-1-akdlgQuery specific logs within a running application pod. Logs that are sent to stdout are collected by the OpenShift Logging framework and are included in the output of the preceding command. The following query is only required for logs that are not sent to stdout.
If an application log can be accessed without root privileges within a pod, concatenate the log file as follows:
$ oc exec my-app-1-akdlg -- cat /var/log/my-application.logIf root access is required to view an application log, you can start a debug container with root privileges and then view the log file from within the container. Start the debug container from the project’s
object. Pod users typically run with non-root privileges, but running troubleshooting pods with temporary root privileges can be useful during issue investigation:DeploymentConfig$ oc debug dc/my-deployment-configuration --as-root -- cat /var/log/my-application.logNoteYou can access an interactive shell with root access within the debug pod if you run
without appendingoc debug dc/<deployment_configuration> --as-root.-- <command>
Test application functionality interactively and run diagnostic tools, in an application container with an interactive shell.
Start an interactive shell on the application container:
$ oc exec -it my-app-1-akdlg /bin/bash- Test application functionality interactively from within the shell. For example, you can run the container’s entry point command and observe the results. Then, test changes from the command line directly, before updating the source code and rebuilding the application container through the S2I process.
Run diagnostic binaries available within the container.
NoteRoot privileges are required to run some diagnostic binaries. In these situations you can start a debug pod with root access, based on a problematic pod’s
object, by runningDeploymentConfig. Then, you can run diagnostic binaries as root from within the debug pod.oc debug dc/<deployment_configuration> --as-root
If diagnostic binaries are not available within a container, you can run a host’s diagnostic binaries within a container’s namespace by using
. The following example runsnsenterwithin a container’s namespace, using the host`sip adbinary.ipEnter into a debug session on the target node. This step instantiates a debug pod called
:<node_name>-debug$ oc debug node/my-cluster-nodeSet
as the root directory within the debug shell. The debug pod mounts the host’s root file system in/hostwithin the pod. By changing the root directory to/host, you can run binaries contained in the host’s executable paths:/host# chroot /hostNoteOpenShift Container Platform 4.11 cluster nodes running Red Hat Enterprise Linux CoreOS (RHCOS) are immutable and rely on Operators to apply cluster changes. Accessing cluster nodes by using SSH is not recommended. However, if the OpenShift Container Platform API is not available, or the kubelet is not properly functioning on the target node,
operations will be impacted. In such situations, it is possible to access nodes usingocinstead.ssh core@<node>.<cluster_name>.<base_domain>Determine the target container ID:
# crictl psDetermine the container’s process ID. In this example, the target container ID is
:a7fe32346b120# crictl inspect a7fe32346b120 --output yaml | grep 'pid:' | awk '{print $2}'Run
within the container’s namespace, using the host’sip adbinary. This example usesipas the container’s process ID. The31150command enters the namespace of a target process and runs a command in its namespace. Because the target process in this example is a container’s process ID, thensentercommand is run in the container’s namespace from the host:ip ad# nsenter -n -t 31150 -- ip adNoteRunning a host’s diagnostic binaries within a container’s namespace is only possible if you are using a privileged container such as a debug node.
7.9. Troubleshooting storage issues Copia collegamentoCollegamento copiato negli appunti!
7.9.1. Resolving multi-attach errors Copia collegamentoCollegamento copiato negli appunti!
When a node crashes or shuts down abruptly, the attached ReadWriteOnce (RWO) volume is expected to be unmounted from the node so that it can be used by a pod scheduled on another node.
However, mounting on a new node is not possible because the failed node is unable to unmount the attached volume.
A multi-attach error is reported:
Example output
Unable to attach or mount volumes: unmounted volumes=[sso-mysql-pvol], unattached volumes=[sso-mysql-pvol default-token-x4rzc]: timed out waiting for the condition
Multi-Attach error for volume "pvc-8837384d-69d7-40b2-b2e6-5df86943eef9" Volume is already used by pod(s) sso-mysql-1-ns6b4
Procedure
To resolve the multi-attach issue, use one of the following solutions:
Enable multiple attachments by using RWX volumes.
For most storage solutions, you can use ReadWriteMany (RWX) volumes to prevent multi-attach errors.
Recover or delete the failed node when using an RWO volume.
For storage that does not support RWX, such as VMware vSphere, RWO volumes must be used instead. However, RWO volumes cannot be mounted on multiple nodes.
If you encounter a multi-attach error message with an RWO volume, force delete the pod on a shutdown or crashed node to avoid data loss in critical workloads, such as when dynamic persistent volumes are attached.
$ oc delete pod <old_pod> --force=true --grace-period=0This command deletes the volumes stuck on shutdown or crashed nodes after six minutes.
7.10. Troubleshooting Windows container workload issues Copia collegamentoCollegamento copiato negli appunti!
7.10.1. Windows Machine Config Operator does not install Copia collegamentoCollegamento copiato negli appunti!
If you have completed the process of installing the Windows Machine Config Operator (WMCO), but the Operator is stuck in the
InstallWaiting
The WMCO requires your OpenShift Container Platform cluster to be configured with hybrid networking using OVN-Kubernetes; the WMCO cannot complete the installation process without hybrid networking available. This is necessary to manage nodes on multiple operating systems (OS) and OS variants. This must be completed during the installation of your cluster.
For more information, see Configuring hybrid networking.
7.10.2. Investigating why Windows Machine does not become compute node Copia collegamentoCollegamento copiato negli appunti!
There are various reasons why a Windows Machine does not become a compute node. The best way to investigate this problem is to collect the Windows Machine Config Operator (WMCO) logs.
Prerequisites
- You installed the Windows Machine Config Operator (WMCO) using Operator Lifecycle Manager (OLM).
- You have created a Windows machine set.
Procedure
Run the following command to collect the WMCO logs:
$ oc logs -f deployment/windows-machine-config-operator -n openshift-windows-machine-config-operator
7.10.3. Accessing a Windows node Copia collegamentoCollegamento copiato negli appunti!
Windows nodes cannot be accessed using the
oc debug node
7.10.3.1. Accessing a Windows node using SSH Copia collegamentoCollegamento copiato negli appunti!
You can access a Windows node by using a secure shell (SSH).
Prerequisites
- You have installed the Windows Machine Config Operator (WMCO) using Operator Lifecycle Manager (OLM).
- You have created a Windows machine set.
-
You have added the key used in the secret and the key used when creating the cluster to the ssh-agent. For security reasons, remember to remove the keys from the ssh-agent after use.
cloud-private-key -
You have connected to the Windows node using an
ssh-bastionpod.
Procedure
Access the Windows node by running the following command:
$ ssh -t -o StrictHostKeyChecking=no -o ProxyCommand='ssh -A -o StrictHostKeyChecking=no \ -o ServerAliveInterval=30 -W %h:%p core@$(oc get service --all-namespaces -l run=ssh-bastion \ -o go-template="{{ with (index (index .items 0).status.loadBalancer.ingress 0) }}{{ or .hostname .ip }}{{end}}")' <username>@<windows_node_internal_ip>1 2 $ oc get nodes <node_name> -o jsonpath={.status.addresses[?\(@.type==\"InternalIP\"\)].address}
7.10.3.2. Accessing a Windows node using RDP Copia collegamentoCollegamento copiato negli appunti!
You can access a Windows node by using a Remote Desktop Protocol (RDP).
Prerequisites
- You installed the Windows Machine Config Operator (WMCO) using Operator Lifecycle Manager (OLM).
- You have created a Windows machine set.
-
You have added the key used in the secret and the key used when creating the cluster to the ssh-agent. For security reasons, remember to remove the keys from the ssh-agent after use.
cloud-private-key -
You have connected to the Windows node using an
ssh-bastionpod.
Procedure
Run the following command to set up an SSH tunnel:
$ ssh -L 2020:<windows_node_internal_ip>:3389 \1 core@$(oc get service --all-namespaces -l run=ssh-bastion -o go-template="{{ with (index (index .items 0).status.loadBalancer.ingress 0) }}{{ or .hostname .ip }}{{end}}")- 1
- Specify the internal IP address of the node, which can be discovered by running the following command:
$ oc get nodes <node_name> -o jsonpath={.status.addresses[?\(@.type==\"InternalIP\"\)].address}From within the resulting shell, SSH into the Windows node and run the following command to create a password for the user:
C:\> net user <username> *1 - 1
- Specify the cloud provider user name, such as
Administratorfor AWS orcapifor Azure.
You can now remotely access the Windows node at
localhost:2020
7.10.4. Collecting Kubernetes node logs for Windows containers Copia collegamentoCollegamento copiato negli appunti!
Windows container logging works differently from Linux container logging; the Kubernetes node logs for Windows workloads are streamed to the
C:\var\logs
Prerequisites
- You installed the Windows Machine Config Operator (WMCO) using Operator Lifecycle Manager (OLM).
- You have created a Windows machine set.
Procedure
To view the logs under all directories in
, run the following command:C:\var\logs$ oc adm node-logs -l kubernetes.io/os=windows --path= \ /ip-10-0-138-252.us-east-2.compute.internal containers \ /ip-10-0-138-252.us-east-2.compute.internal hybrid-overlay \ /ip-10-0-138-252.us-east-2.compute.internal kube-proxy \ /ip-10-0-138-252.us-east-2.compute.internal kubelet \ /ip-10-0-138-252.us-east-2.compute.internal podsYou can now list files in the directories using the same command and view the individual log files. For example, to view the kubelet logs, run the following command:
$ oc adm node-logs -l kubernetes.io/os=windows --path=/kubelet/kubelet.log
7.10.5. Collecting Windows application event logs Copia collegamentoCollegamento copiato negli appunti!
The
Get-WinEvent
logs
Prerequisites
- You installed the Windows Machine Config Operator (WMCO) using Operator Lifecycle Manager (OLM).
- You have created a Windows machine set.
Procedure
To view logs from all applications logging to the event logs on the Windows machine, run:
$ oc adm node-logs -l kubernetes.io/os=windows --path=journalThe same command is executed when collecting logs with
.oc adm must-gatherOther Windows application logs from the event log can also be collected by specifying the respective service with a
flag. For example, you can run the following command to collect logs for the docker runtime service:-u$ oc adm node-logs -l kubernetes.io/os=windows --path=journal -u docker
7.10.6. Collecting Docker logs for Windows containers Copia collegamentoCollegamento copiato negli appunti!
The Windows Docker service does not stream its logs to stdout, but instead, logs to the event log for Windows. You can view the Docker event logs to investigate issues you think might be caused by the Windows Docker service.
Prerequisites
- You installed the Windows Machine Config Operator (WMCO) using Operator Lifecycle Manager (OLM).
- You have created a Windows machine set.
Procedure
SSH into the Windows node and enter PowerShell:
C:\> powershellView the Docker logs by running the following command:
C:\> Get-EventLog -LogName Application -Source Docker
7.11. Investigating monitoring issues Copia collegamentoCollegamento copiato negli appunti!
OpenShift Container Platform includes a preconfigured, preinstalled, and self-updating monitoring stack that provides monitoring for core platform components. In OpenShift Container Platform 4.11, cluster administrators can optionally enable monitoring for user-defined projects.
You can follow these procedures if your own metrics are unavailable or if Prometheus is consuming a lot of disk space.
7.11.2. Determining why Prometheus is consuming a lot of disk space Copia collegamentoCollegamento copiato negli appunti!
Developers can create labels to define attributes for metrics in the form of key-value pairs. The number of potential key-value pairs corresponds to the number of possible values for an attribute. An attribute that has an unlimited number of potential values is called an unbound attribute. For example, a
customer_id
Every assigned key-value pair has a unique time series. The use of many unbound attributes in labels can result in an exponential increase in the number of time series created. This can impact Prometheus performance and can consume a lot of disk space.
You can use the following measures when Prometheus consumes a lot of disk:
- Check the number of scrape samples that are being collected.
- Check the time series database (TSDB) status using the Prometheus HTTP API for more information about which labels are creating the most time series. Doing so requires cluster administrator privileges.
Reduce the number of unique time series that are created by reducing the number of unbound attributes that are assigned to user-defined metrics.
NoteUsing attributes that are bound to a limited set of possible values reduces the number of potential key-value pair combinations.
- Enforce limits on the number of samples that can be scraped across user-defined projects. This requires cluster administrator privileges.
Prerequisites
-
You have access to the cluster as a user with the cluster role.
cluster-admin -
You have installed the OpenShift CLI ().
oc
Procedure
-
In the Administrator perspective, navigate to Observe
Metrics. Run the following Prometheus Query Language (PromQL) query in the Expression field. This returns the ten metrics that have the highest number of scrape samples:
topk(10,count by (job)({__name__=~".+"}))Investigate the number of unbound label values assigned to metrics with higher than expected scrape sample counts.
- If the metrics relate to a user-defined project, review the metrics key-value pairs assigned to your workload. These are implemented through Prometheus client libraries at the application level. Try to limit the number of unbound attributes referenced in your labels.
- If the metrics relate to a core OpenShift Container Platform project, create a Red Hat support case on the Red Hat Customer Portal.
Review the TSDB status using the Prometheus HTTP API by running the following commands as a cluster administrator:
$ oc login -u <username> -p <password>$ host=$(oc -n openshift-monitoring get route prometheus-k8s -ojsonpath={.spec.host})$ token=$(oc whoami -t)$ curl -H "Authorization: Bearer $token" -k "https://$host/api/v1/status/tsdb"Example output
"status": "success",
7.12. Diagnosing OpenShift CLI (oc) issues Copia collegamentoCollegamento copiato negli appunti!
7.12.1. Understanding OpenShift CLI (oc) log levels Copia collegamentoCollegamento copiato negli appunti!
With the OpenShift CLI (
oc
If
oc
oc
curl
oc
oc
oc
| Log level | Description |
|---|---|
| 1 to 5 | No additional logging to stderr. |
| 6 | Log API requests to stderr. |
| 7 | Log API requests and headers to stderr. |
| 8 | Log API requests, headers, and body, plus API response headers and body to stderr. |
| 9 | Log API requests, headers, and body, API response headers and body, plus
|
| 10 | Log API requests, headers, and body, API response headers and body, plus
|
7.12.2. Specifying OpenShift CLI (oc) log levels Copia collegamentoCollegamento copiato negli appunti!
You can investigate OpenShift CLI (
oc
The OpenShift Container Platform user’s current session token is typically included in logged
curl
oc
Prerequisites
-
Install the OpenShift CLI ().
oc
Procedure
Specify the
log level when running anoccommand:oc$ oc <command> --loglevel <log_level>where:
- <command>
- Specifies the command you are running.
- <log_level>
- Specifies the log level to apply to the command.
To obtain the current user’s session token, run the following command:
$ oc whoami -tExample output
sha256~RCV3Qcn7H-OEfqCGVI0CvnZ6...