This documentation is for a release that is no longer maintained
See documentation for the latest supported version 3 or the latest supported version 4.Chapter 19. Configuring for GCE
19.1. Overview Copy linkLink copied to clipboard!
OpenShift Container Platform can be configured to access a Google Compute Engine (GCE) infrastructure, including using GCE volumes as persistent storage for application data. After GCE is configured properly, some additional configurations will need to be completed on the OpenShift Container Platform hosts.
19.2. Configuring Masters Copy linkLink copied to clipboard!
You can set the GCE configuration on your OpenShift Container Platform master hosts in two ways:
19.2.1. Configuring OpenShift Container Platform Masters for GCE with Ansible Copy linkLink copied to clipboard!
During advanced installations, GCE can be configured using the openshift_cloudprovider_kind
parameter, which is configurable in the inventory file.
Example GCE Configuration with Ansible
Cloud Provider Configuration
# Cloud Provider Configuration
#
openshift_cloudprovider_kind=gce
When Ansible configures GCE, the following files are created for you:
- /etc/origin/cloudprovider/gce.conf
- /etc/origin/master/master-config.yaml
- /etc/origin/node/node-config.yaml
The advanced installation configures multizone support by default. If you want single-zone support, edit the /etc/origin/cloudprovider/gce.conf as shown in Configuring Multizone Support in a GCE Deployment.
19.2.2. Manually Configuring OpenShift Container Platform Masters for GCE Copy linkLink copied to clipboard!
To configure the OpenShift Container Platform masters for GCE:
Edit or create the master configuration file (/etc/origin/master/master-config.yaml by default) on all masters and update the contents of the
apiServerArguments
andcontrollerArguments
sections:Copy to Clipboard Copied! Toggle word wrap Toggle overflow ImportantWhen triggering a containerized installation, only the directories of /etc/origin and /var/lib/origin are mounted to the master and node container. Therefore, master-config.yaml should be in /etc/origin/master instead of /etc/.
Start or restart the OpenShift Container Platform services:
systemctl restart atomic-openshift-master
# systemctl restart atomic-openshift-master
Copy to Clipboard Copied! Toggle word wrap Toggle overflow
19.3. Configuring Nodes Copy linkLink copied to clipboard!
To configure the OpenShift Container Platform nodes for GCE:
Edit or create the node configuration file (/etc/origin/node/node-config.yaml by default) on all nodes and update the contents of the
kubeletArguments
section:kubeletArguments: cloud-provider: - "gce" cloud-config: - "/etc/origin/cloudprovider/gce.conf"
kubeletArguments: cloud-provider: - "gce" cloud-config: - "/etc/origin/cloudprovider/gce.conf"
Copy to Clipboard Copied! Toggle word wrap Toggle overflow
Currently, the nodeName
must match the instance name in GCE in order for the cloud provider integration to work properly. The name must also be RFC1123 compliant.
When triggering a containerized installation, only the directories of /etc/origin and /var/lib/origin are mounted to the master and node container. Therefore, node-config.yaml should be in /etc/origin/node instead of /etc/.
Start or restart the OpenShift Container Platform services all nodes.
systemctl restart atomic-openshift-node
# systemctl restart atomic-openshift-node
Copy to Clipboard Copied! Toggle word wrap Toggle overflow
19.4. Configuring Multizone Support in a GCE Deployment Copy linkLink copied to clipboard!
If manually congifuring GCE, multizone support is not configured by default.
The advanced installation configures multizone support by default.
If you want multizone support:
- Edit or create a /etc/origin/cloudprovider/gce.conf file on all of your OpenShift Container Platform hosts, both masters and nodes.
Add the following contents:
[Global] multizone = true
[Global] multizone = true
Copy to Clipboard Copied! Toggle word wrap Toggle overflow
To return to single-zone support, set the multizone
value to false
.
19.5. Applying Configuration Changes Copy linkLink copied to clipboard!
Start or restart OpenShift Container Platform services on all master and node hosts to apply your configuration changes:
systemctl restart atomic-openshift-master systemctl restart atomic-openshift-node
# systemctl restart atomic-openshift-master
# systemctl restart atomic-openshift-node
Switching from not using a cloud provider to using a cloud provider produces an error message. Adding the cloud provider tries to delete the node because the node switches from using the hostname as the externalID
(which would have been the case when no cloud provider was being used) to using the GCE instance-id
(which is what the GCE cloud provider specifies). To resolve this issue:
- Log in to the CLI as a cluster administrator.
Check and backup existing node labels:
oc describe node <node_name> | grep -Poz '(?s)Labels.*\n.*(?=Taints)'
$ oc describe node <node_name> | grep -Poz '(?s)Labels.*\n.*(?=Taints)'
Copy to Clipboard Copied! Toggle word wrap Toggle overflow Delete the nodes:
oc delete node <node_name>
$ oc delete node <node_name>
Copy to Clipboard Copied! Toggle word wrap Toggle overflow On each node host, restart the OpenShift Container Platform service.
systemctl restart atomic-openshift-node
$ systemctl restart atomic-openshift-node
Copy to Clipboard Copied! Toggle word wrap Toggle overflow - Add back any labels on each node that you previously had.