CLI Reference
OpenShift Container Platform 3.10 CLI Reference
Abstract
Chapter 1. Overview
With the OpenShift Container Platform command line interface (CLI), you can create applications and manage OpenShift Container Platform projects from a terminal. The CLI is ideal in situations where you are:
- Working directly with project source code.
- Scripting OpenShift Container Platform operations.
- Restricted by bandwidth resources and cannot use the web console.
The CLI is available using the oc
command:
$ oc <command>
See Get Started with the CLI for installation and setup instructions.
Chapter 2. Get Started with the CLI
2.1. Overview
The OpenShift Container Platform CLI exposes commands for managing your applications, as well as lower level tools to interact with each component of your system. This topic guides you through getting started with the CLI, including installation and logging in to create your first project.
2.2. Prerequisites
Certain operations require Git to be locally installed on a client. For example, the command to create an application using a remote Git repository:
$ oc new-app https://github.com/<your_user>/<your_git_repo>
Before proceeding, install Git on your workstation. See the official Git documentation for instructions per your workstation’s operating system.
2.3. Installing the CLI
The easiest way to download the CLI is by accessing the About page on the web console if your cluster administrator has enabled the download links:
Installation options for the CLI vary depending on your operating system.
To log in using the CLI, collect your token from the web console’s Command Line page, which is accessed from Command Line Tools in the Help menu. The token is hidden, so you must click the copy to clipboard button at the end of the oc login
line on the Command Line Tools page, then paste the copied contents to show the token.
2.3.1. For Windows
The CLI for Windows is provided as a zip archive; you can download it from the Red Hat Customer Portal. After logging in with your Red Hat account, you must have an active OpenShift Enterprise subscription to access the downloads page:
Download the CLI from the Red Hat Customer Portal
Alternatively, if the cluster administrator has enabled it, you can download and unpack the CLI from the About page on the web console.
Tutorial Video:
The following video walks you through this process: Click here to watch
Then, unzip the archive with a ZIP program and move the oc
binary to a directory on your PATH. To check your PATH, open the Command Prompt and run:
C:\> path
2.3.2. For Mac OS X
The CLI for Mac OS X is provided as a tar.gz archive; you can download it from the Red Hat Customer Portal. After logging in with your Red Hat account, you must have an active OpenShift Enterprise subscription to access the downloads page:
Download the CLI from the Red Hat Customer Portal
Alternatively, if the cluster administrator has enabled it, you can download and unpack the CLI from the About page on the web console.
Tutorial Video:
The following video walks you through this process: Click here to watch
Then, unpack the archive and move the oc
binary to a directory on your PATH. To check your PATH, open a Terminal window and run:
$ echo $PATH
2.3.3. For Linux
For Red Hat Enterprise Linux (RHEL) 7, you can install the CLI as an RPM using Red Hat Subscription Management (RHSM) if you have an active OpenShift Enterprise subscription on your Red Hat account:
# subscription-manager register
# subscription-manager refresh
# subscription-manager attach --pool=<pool_id> 1
# subscription-manager repos --enable="rhel-7-server-ose-3.10-rpms"
# yum install atomic-openshift-clients
- 1
- Pool ID for an active OpenShift Enterprise subscription
For RHEL, Fedora, and other Linux distributions, you can also download the CLI directly from the Red Hat Customer Portal as a tar.gz archive. After logging in with your Red Hat account, you must have an active OpenShift Enterprise subscription to access the downloads page.
Download the CLI from the Red Hat Customer Portal
Tutorial Video:
The following video walks you through this process: Click here to watch
Alternatively, if the cluster administrator has enabled it, you can download and unpack the CLI from the About page on the web console.
Then, unpack the archive and move the oc
binary to a directory on your PATH. To check your path, run:
$ echo $PATH
To unpack the archive:
$ tar -xf <file>
If you do not use RHEL or Fedora, ensure that libc is installed in a directory on your library path. If libc is not available, you might see the following error when you run CLI commands:
oc: No such file or directory
2.4. Basic Setup and Login
The oc login
command is the best way to initially set up the CLI, and it serves as the entry point for most users. The interactive flow helps you establish a session to an OpenShift Container Platform server with the provided credentials. The information is automatically saved in a CLI configuration file that is then used for subsequent commands.
The following example shows the interactive setup and login using the oc login
command:
Example 2.1. Initial CLI Setup
$ oc login OpenShift server [https://localhost:8443]: https://openshift.example.com 1 Username: alice 2 Authentication required for https://openshift.example.com (openshift) Password: ****** Login successful. 3 You don't have any projects. You can try to create a new project, by running $ oc new-project <projectname> 4 Welcome to OpenShift! See 'oc help' to get started.
When you have completed the CLI configuration, subsequent commands use the configuration file for the server, session token, and project information.
You can log out of CLI using the oc logout
command:
$ oc logout User, alice, logged out of https://openshift.example.com
If you log in after creating or being granted access to a project, a project you have access to is automatically set as the current default, until switching to another one:
$ oc login Username: alice Authentication required for https://openshift.example.com (openshift) Password: Login successful. Using project "aliceproject".
Additional options are also available for the oc login
command.
If you have access to administrator credentials but are no longer logged in as the default system user system:admin, you can log back in as this user at any time as long as the credentials are still present in your CLI configuration file. The following command logs in and switches to the default project:
$ oc login -u system:admin -n default
2.5. CLI Configuration Files
A CLI configuration file permanently stores oc
options and contains a series of authentication mechanisms and OpenShift Container Platform server connection information associated with nicknames.
As described in the previous section, the oc login
command automatically creates and manages CLI configuration files. All information gathered by the command is stored in a configuration file located in ~/.kube/config. The current CLI configuration can be viewed using the following command:
Example 2.2. Viewing the CLI Configuration
$ oc config view apiVersion: v1 clusters: - cluster: server: https://openshift.example.com name: openshift contexts: - context: cluster: openshift namespace: aliceproject user: alice name: alice current-context: alice kind: Config preferences: {} users: - name: alice user: token: NDM2N2MwODgtNjI1Yy10N3VhLTg1YmItYzI4NDEzZDUyYzVi
CLI configuration files can be used to setup multiple CLI profiles using various OpenShift Container Platform servers, namespaces, and users so that you can switch easily between them. The CLI can support multiple configuration files; they are loaded at runtime and merged together along with any override options specified from the command line.
2.6. Projects
A project in OpenShift Container Platform contains multiple objects to make up a logical application.
Most oc
commands run in the context of a project. The oc login
selects a default project during initial setup to be used with subsequent commands. Use the following command to display the project currently in use:
$ oc project
If you have access to multiple projects, use the following syntax to switch to a particular project by specifying the project name:
$ oc project <project_name>
For example:
$ oc project project02 Now using project 'project02'. $ oc project project03 Now using project 'project03'. $ oc project Using project 'project03'.
The oc status
command shows a high level overview of the project currently in use, with its components and their relationships, as shown in the following example:
$ oc status In project OpenShift 3 Sample (test) service database-test (172.30.17.113:6434 -> 3306) database-test deploys docker.io/library/mysql:latest #1 deployed 47 hours ago service frontend-test (172.30.17.236:5432 -> 8080) frontend-test deploys origin-ruby-sample:test <- builds https://github.com/openshift/ruby-hello-world with docker.io/openshift/ruby-20-centos7:latest not built yet #1 deployment waiting on image To see more information about a service or deployment config, use 'oc describe service <name>' or 'oc describe dc <name>'. You can use 'oc get pods,svc,dc,bc,builds' to see lists of each of the types described above.
2.7. What’s Next?
After you have logged in, you can create a new application and explore some common CLI operations.
Chapter 3. Managing CLI Profiles
3.1. Overview
A CLI configuration file allows you to configure different profiles, or contexts, for use with the OpenShift CLI. A context consists of user authentication and OpenShift Container Platform server information associated with a nickname.
3.2. Switching Between CLI Profiles
Contexts allow you to easily switch between multiple users across multiple OpenShift Container Platform servers, or clusters, when using issuing CLI operations. Nicknames make managing CLI configuration easier by providing short-hand references to contexts, user credentials, and cluster details.
After logging in with the CLI for the first time, OpenShift Container Platform creates a ~/.kube/config file if one does not already exist. As more authentication and connection details are provided to the CLI, either automatically during an oc login
operation or by setting them explicitly, the updated information is stored in the configuration file:
Example 3.1. CLI Configuration File
apiVersion: v1 clusters: 1 - cluster: insecure-skip-tls-verify: true server: https://openshift1.example.com:8443 name: openshift1.example.com:8443 - cluster: insecure-skip-tls-verify: true server: https://openshift2.example.com:8443 name: openshift2.example.com:8443 contexts: 2 - context: cluster: openshift1.example.com:8443 namespace: alice-project user: alice/openshift1.example.com:8443 name: alice-project/openshift1.example.com:8443/alice - context: cluster: openshift1.example.com:8443 namespace: joe-project user: alice/openshift1.example.com:8443 name: joe-project/openshift1/alice current-context: joe-project/openshift1.example.com:8443/alice 3 kind: Config preferences: {} users: 4 - name: alice/openshift1.example.com:8443 user: token: xZHd2piv5_9vQrg-SKXRJ2Dsl9SceNJdhNTljEKTb8k
- 1
- The
clusters
section defines connection details for OpenShift Container Platform clusters, including the address for their master server. In this example, one cluster is nicknamed openshift1.example.com:8443 and another is nicknamed openshift2.example.com:8443. - 2
- This
contexts
section defines two contexts: one nicknamed alice-project/openshift1.example.com:8443/alice, using the alice-project project, openshift1.example.com:8443 cluster, and alice user, and another nicknamed joe-project/openshift1.example.com:8443/alice, using the joe-project project, openshift1.example.com:8443 cluster and alice user. - 3
- The
current-context
parameter shows that the joe-project/openshift1.example.com:8443/alice context is currently in use, allowing the alice user to work in the joe-project project on the openshift1.example.com:8443 cluster. - 4
- The
users
section defines user credentials. In this example, the user nickname alice/openshift1.example.com:8443 uses an access token.
The CLI can support multiple configuration files; they are loaded at runtime and merged together along with any override options specified from the command line.
After you are logged in, you can use the oc status
command or the oc project
command to verify your current working environment:
Example 3.2. Verifying the Current Working Environment
$ oc status oc status In project Joe's Project (joe-project) service database (172.30.43.12:5434 -> 3306) database deploys docker.io/openshift/mysql-55-centos7:latest #1 deployed 25 minutes ago - 1 pod service frontend (172.30.159.137:5432 -> 8080) frontend deploys origin-ruby-sample:latest <- builds https://github.com/openshift/ruby-hello-world with joe-project/ruby-20-centos7:latest #1 deployed 22 minutes ago - 2 pods To see more information about a service or deployment, use 'oc describe service <name>' or 'oc describe dc <name>'. You can use 'oc get all' to see lists of each of the types described above.
$ oc project Using project "joe-project" from context named "joe-project/openshift1.example.com:8443/alice" on server "https://openshift1.example.com:8443".
To log in using any other combination of user credentials and cluster details, run the oc login
command again and supply the relevant information during the interactive process. A context is constructed based on the supplied information if one does not already exist.
If you are already logged in and want to switch to another project the current user already has access to, use the oc project
command and supply the name of the project:
$ oc project alice-project Now using project "alice-project" on server "https://openshift1.example.com:8443".
At any time, you can use the oc config view
command to view your current, full CLI configuration, as seen in the output.
Additional CLI configuration commands are also available for more advanced usage.
If you have access to administrator credentials but are no longer logged in as the default system user system:admin, you can log back in as this user at any time as long as the credentials are still present in your CLI configuration file. The following command logs in and switches to the default project:
$ oc login -u system:admin -n default
3.3. Manually Configuring CLI Profiles
This section covers more advanced usage of CLI configurations. In most situations, you can simply use the oc login
and oc project
commands to log in and switch between contexts and projects.
If you want to manually configure your CLI configuration files, you can use the oc config
command instead of modifying the files themselves. The oc config
command includes a number of helpful subcommands for this purpose:
Subcommand | Usage |
---|---|
| Sets a cluster entry in the CLI configuration file. If the referenced cluster nickname already exists, the specified information is merged in. $ oc config set-cluster <cluster_nickname> [--server=<master_ip_or_fqdn>] [--certificate-authority=<path/to/certificate/authority>] [--api-version=<apiversion>] [--insecure-skip-tls-verify=true] |
| Sets a context entry in the CLI configuration file. If the referenced context nickname already exists, the specified information is merged in. $ oc config set-context <context_nickname> [--cluster=<cluster_nickname>] [--user=<user_nickname>] [--namespace=<namespace>] |
| Sets the current context using the specified context nickname. $ oc config use-context <context_nickname> |
| Sets an individual value in the CLI configuration file. $ oc config set <property_name> <property_value>
The |
| Unsets individual values in the CLI configuration file. $ oc config unset <property_name>
The |
| Displays the merged CLI configuration currently in use. $ oc config view Displays the result of the specified CLI configuration file. $ oc config view --config=<specific_filename> |
Example Usage
Consider the following configuration workflow. First, login as a user that uses an access token. This token is used by the alice user:
$ oc login https://openshift1.example.com --token=ns7yVhuRNpDM9cgzfhhxQ7bM5s7N2ZVrkZepSRf4LC0
View the cluster entry automatically created:
$ oc config view apiVersion: v1 clusters: - cluster: insecure-skip-tls-verify: true server: https://openshift1.example.com name: openshift1-example-com contexts: - context: cluster: openshift1-example-com namespace: default user: alice/openshift1-example-com name: default/openshift1-example-com/alice current-context: default/openshift1-example-com/alice kind: Config preferences: {} users: - name: alice/openshift1.example.com user: token: ns7yVhuRNpDM9cgzfhhxQ7bM5s7N2ZVrkZepSRf4LC0
Update the current context to have users login to the desired namespace:
$ oc config set-context `oc config current-context` --namespace=<project_name>
To confirm that the changes have taken effect, examine the current context:
$ oc whoami -c
All subsequent CLI operations will use the new context, unless otherwise specified by overriding CLI options or until the context is switched.
3.4. Loading and Merging Rules
When issuing CLI operations, the loading and merging order for the CLI configuration follows these rules:
CLI configuration files are retrieved from your workstation, using the following hierarchy and merge rules:
-
If the
--config
option is set, then only that file is loaded. The flag may only be set once and no merging takes place. -
If
$KUBECONFIG
environment variable is set, then it is used. The variable can be a list of paths, and if so the paths are merged together. When a value is modified, it is modified in the file that defines the stanza. When a value is created, it is created in the first file that exists. If no files in the chain exist, then it creates the last file in the list. -
Otherwise, the ~/.kube/config file is used and no merging takes place.
-
If the
The context to use is determined based on the first hit in the following chain:
-
The value of the
--context
option. -
The
current-context
value from the CLI configuration file. -
An empty value is allowed at this stage.
-
The value of the
The user and cluster to use is determined. At this point, you may or may not have a context; they are built based on the first hit in the following chain, which is run once for the user and once for the cluster:
-
The value of the
--user
option for user name and the--cluster
option for cluster name. -
If the
--context
option is present, then use the context’s value. -
An empty value is allowed at this stage.
-
The value of the
The actual cluster information to use is determined. At this point, you may or may not have cluster information. Each piece of the cluster information is built based on the first hit in the following chain:
The values of any of the following command line options:
-
--server
, -
--api-version
-
--certificate-authority
-
--insecure-skip-tls-verify
-
- If cluster information and a value for the attribute is present, then use it.
-
If you do not have a server location, then there is an error.
The actual user information to use is determined. Users are built using the same rules as clusters, except that you can only have one authentication technique per user; conflicting techniques cause the operation to fail. Command line options take precedence over configuration file values. Valid command line options are:
-
--auth-path
-
--client-certificate
-
--client-key
-
--token
-
- For any information that is still missing, default values are used and prompts are given for additional information.
Chapter 4. Developer CLI Operations
4.1. Overview
This topic provides information on the developer CLI operations and their syntax. You must setup and login with the CLI before you can perform these operations.
The developer CLI uses the oc
command, and is used for project-level operations. This differs from the administrator CLI, which uses the oc adm
command for more advanced, administrator operations.
4.2. Common Operations
The developer CLI allows interaction with the various objects that are managed by OpenShift Container Platform. Many common oc
operations are invoked using the following syntax:
$ oc <action> <object_type> <object_name>
This specifies:
-
An
<action>
to perform, such asget
ordescribe
. -
The
<object_type>
to perform the action on, such asservice
or the abbreviatedsvc
. -
The
<object_name>
of the specified<object_type>
.
For example, the oc get
operation returns a complete list of services that are currently defined:
$ oc get svc NAME LABELS SELECTOR IP PORT(S) docker-registry docker-registry=default docker-registry=default 172.30.78.158 5000/TCP kubernetes component=apiserver,provider=kubernetes <none> 172.30.0.2 443/TCP kubernetes-ro component=apiserver,provider=kubernetes <none> 172.30.0.1 80/TCP
The oc describe
operation can then be used to return detailed information about a specific object:
$ oc describe svc docker-registry Name: docker-registry Labels: docker-registry=default Selector: docker-registry=default IP: 172.30.78.158 Port: <unnamed> 5000/TCP Endpoints: 10.128.0.2:5000 Session Affinity: None No events.
Versions of oc
prior to 3.0.2.0 did not have the ability to negotiate API versions against a server. So if you are using oc
up to 3.0.1.0 with a server that only supports v1 or higher versions of the API, make sure to pass --api-version
in order to point the oc
client to the correct API endpoint. For example: oc get svc --api-version=v1
.
4.3. Object Types
The CLI supports the following object types, some of which have abbreviated syntax:
Object Type | Abbreviated Version |
---|---|
| |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| |
|
|
| |
|
|
|
|
|
|
|
|
| |
|
|
|
|
|
|
|
|
4.4. Basic CLI Operations
The following table describes basic oc
operations and their general syntax:
4.4.1. types
Display an introduction to some core OpenShift Container Platform concepts:
$ oc types
4.4.2. login
Log in to the OpenShift Container Platform server:
$ oc login
4.4.3. logout
End the current session:
$ oc logout
4.4.4. new-project
Create a new project:
$ oc new-project <project_name>
4.4.5. new-app
Creates a new application based on the source code in the current directory:
$ oc new-app .
Creates a new application based on the source code in a remote repository:
$ oc new-app https://github.com/sclorg/cakephp-ex
Creates a new application based on the source code in a private remote repository:
$ oc new-app https://github.com/youruser/yourprivaterepo --source-secret=yoursecret
4.4.6. status
Show an overview of the current project:
$ oc status
4.4.7. project
Switch to another project. Run without options to display the current project. To view all projects you have access to run oc projects
.
$ oc project <project_name>
4.5. Application Modification CLI Operations
4.5.1. get
Return a list of objects for the specified object type. If the optional <object_name>
is included in the request, then the list of results is filtered by that value.
$ oc get <object_type> [<object_name>]
You can use the -o
or --output
option to modify the output format.
$ oc get <object_type> [<object_name>]-o|--output=json|yaml|wide|custom-columns=...|custom-columns-file=...|go-template=...|go-template-file=...|jsonpath=...|jsonpath-file=...]
The output format can be a JSON or YAML, or an extensible format like custom columns, golang template, and jsonpath.
For example, the following command lists the name of the pods running in a specific project:
$ oc get pods -n default -o jsonpath='{range .items[*].metadata}{"Pod Name: "}{.name}{"\n"}{end}' Pod Name: docker-registry-1-wvhrx Pod Name: registry-console-1-ntq65 Pod Name: router-1-xzw69
4.5.2. describe
Returns information about the specific object returned by the query. A specific <object_name>
must be provided. The actual information that is available varies as described in object type.
$ oc describe <object_type> <object_name>
4.5.3. edit
Edit the desired object type:
$ oc edit <object_type>/<object_name>
Edit the desired object type with a specified text editor:
$ OC_EDITOR="<text_editor>" oc edit <object_type>/<object_name>
Edit the desired object in a specified format (eg: JSON):
$ oc edit <object_type>/<object_name> \ --output-version=<object_type_version> \ -o <object_type_format>
4.5.4. volume
Modify a volume:
$ oc volume <object_type>/<object_name> [--option]
4.5.5. label
Update the labels on a object:
$ oc label <object_type> <object_name> <label>
4.5.6. expose
Look up a service and expose it as a route. There is also the ability to expose a deployment configuration, replication controller, service, or pod as a new service on a specified port. If no labels are specified, the new object will re-use the labels from the object it exposes.
If you are exposing a service, the default generator is --generator=route/v1
. For all other cases the default is --generator=service/v2
, which leaves the port unnamed. Generally, there is no need to set a generator with the oc expose
command. A third generator, --generator=service/v1
, is available with the port name default.
$ oc expose <object_type> <object_name>
4.5.7. delete
Delete the specified object. An object configuration can also be passed in through STDIN. The oc delete all -l <label>
operation deletes all objects matching the specified <label>
, including the replication controller so that pods are not re-created.
$ oc delete -f <file_path>
$ oc delete <object_type> <object_name>
$ oc delete <object_type> -l <label>
$ oc delete all -l <label>
4.5.8. set
Modify a specific property of the specified object.
4.5.8.1. set env
Sets an environment variable on a deployment configuration or a build configuration:
$ oc set env dc/mydc VAR1=value1
4.5.8.2. set build-secret
Sets the name of a secret on a build configuration. The secret may be an image pull or push secret or a source repository secret:
$ oc set build-secret --source bc/mybc mysecret
4.6. Build and Deployment CLI Operations
One of the fundamental capabilities of OpenShift Container Platform is the ability to build applications into a container from source.
OpenShift Container Platform provides CLI access to inspect and manipulate deployment configurations using standard oc
resource operations, such as get
, create
, and describe
.
4.6.1. start-build
Manually start the build process with the specified build configuration file:
$ oc start-build <buildconfig_name>
Manually start the build process by specifying the name of a previous build as a starting point:
$ oc start-build --from-build=<build_name>
Manually start the build process by specifying either a configuration file or the name of a previous build and retrieve its build logs:
$ oc start-build --from-build=<build_name> --follow
$ oc start-build <buildconfig_name> --follow
Wait for a build to complete and exit with a non-zero return code if the build fails:
$ oc start-build --from-build=<build_name> --wait
Set or override environment variables for the current build without changing the build configuration. Alternatively, use -e
.
$ oc start-build --env <var_name>=<value>
Set or override the default build log level output during the build:
$ oc start-build --build-loglevel [0-5]
Specify the source code commit identifier the build should use; requires a build based on a Git repository:
$ oc start-build --commit=<hash>
Re-run build with name <build_name>
:
$ oc start-build --from-build=<build_name>
Archive <dir_name>
and build with it as the binary input:
$ oc start-build --from-dir=<dir_name>
Use existing archive as the binary input; unlike --from-file
the archive will be extracted by the builder prior to the build process:
$ oc start-build --from-archive=<archive_name>
Use <file_name>
as the binary input for the build. This file must be the only one in the build source. For example, pom.xml or Dockerfile.
$ oc start-build --from-file=<file_name>
Download the binary input using HTTP or HTTPS instead of reading it from the file system:
$ oc start-build --from-file=<file_URL>
Download an archive and use its contents as the build source:
$ oc start-build --from-archive=<archive_URL>
The path to a local source code repository to use as the binary input for a build:
$ oc start-build --from-repo=<path_to_repo>
Specify a webhook URL for an existing build configuration to trigger:
$ oc start-build --from-webhook=<webhook_URL>
The contents of the post-receive hook to trigger a build:
$ oc start-build --git-post-receive=<contents>
The path to the Git repository for post-receive; defaults to the current directory:
$ oc start-build --git-repository=<path_to_repo>
List the webhooks for the specified build configuration or build; accepts all
, generic
, or github
:
$ oc start-build --list-webhooks
Override the Spec.Strategy.SourceStrategy.Incremental option of a source-strategy build:
$ oc start-build --incremental
Override the Spec.Strategy.DockerStrategy.NoCache option of a docker-strategy build:
$oc start-build --no-cache
4.6.2. rollback
Perform a rollback:
$ oc rollback <deployment_name>
4.6.3. new-build
Create a build configuration based on the source code in the current Git repository (with a public remote) and a container image:
$ oc new-build .
Create a build configuration based on a remote git repository:
$ oc new-build https://github.com/sclorg/cakephp-ex
Create a build configuration based on a private remote git repository:
$ oc new-build https://github.com/youruser/yourprivaterepo --source-secret=yoursecret
4.6.4. cancel-build
Stop a build that is in progress:
$ oc cancel-build <build_name>
Cancel multiple builds at the same time:
$ oc cancel-build <build1_name> <build2_name> <build3_name>
Cancel all builds created from the build configuration:
$ oc cancel-build bc/<buildconfig_name>
Specify the builds to be canceled:
$ oc cancel-build bc/<buildconfig_name> --state=<state>
Example values for state
are new or pending.
4.6.5. import-image
Import tag and image information from an external image repository:
$ oc import-image <image_stream>
4.6.6. scale
Set the number of desired replicas for a replication controller or a deployment configuration to the number of specified replicas:
$ oc scale <object_type> <object_name> --replicas=<#_of_replicas>
4.6.7. tag
Take an existing tag or image from an image stream, or a container image "pull spec", and set it as the most recent image for a tag in one or more other image streams:
$ oc tag <current_image> <image_stream>
4.7. Advanced Commands
4.7.1. create
Parse a configuration file and create one or more OpenShift Container Platform objects based on the file contents. The -f
flag can be passed multiple times with different file or directory paths. When the flag is passed multiple times, oc create
iterates through each one, creating the objects described in all of the indicated files. Any existing resources are ignored.
$ oc create -f <file_or_dir_path>
4.7.2. replace
Attempt to modify an existing object based on the contents of the specified configuration file. The -f
flag can be passed multiple times with different file or directory paths. When the flag is passed multiple times, oc replace
iterates through each one, updating the objects described in all of the indicated files.
$ oc replace -f <file_or_dir_path>
4.7.3. process
Transform a project template into a project configuration file:
$ oc process -f <template_file_path>
4.7.4. run
Create and run a particular image, possibly replicated. By default, create a deployment configuration to manage the created container(s). You can choose to create a different resource using the --generator
flag:
API Resource | --generator Option |
---|---|
Deployment configuration |
|
Pod |
|
Replication controller |
|
Deployment using |
|
Deployment using |
|
Job |
|
Cron job |
|
You can choose to run in the foreground for an interactive container execution.
$ oc run NAME --image=<image> \ [--generator=<resource>] \ [--port=<port>] \ [--replicas=<replicas>] \ [--dry-run=<bool>] \ [--overrides=<inline_json>] \ [options]
4.7.5. patch
Updates one or more fields of an object using strategic merge patch:
$ oc patch <object_type> <object_name> -p <changes>
The <changes> is a JSON or YAML expression containing the new fields and the values. For example, to update the spec.unschedulable
field of the node node1
to the value true
, the json expression is:
$ oc patch node node1 -p '{"spec":{"unschedulable":true}}'
4.7.6. export
Export resources to be used elsewhere:
$ oc export <object_type> [--options]
See Creating a Template from Existing Objects for more information on exporting existing objects from your project in template form.
4.7.7. policy
Manage authorization policies:
$ oc policy [--options]
4.7.8. secrets
Configure secrets:
$ oc secrets [--options] path/to/ssh_key
4.7.9. autoscale
Setup an autoscaler for your application. Requires metrics to be enabled in the cluster. See Enabling Cluster Metrics for cluster administrator instructions, if needed.
$ oc autoscale dc/<dc_name> [--options]
4.8. Troubleshooting and Debugging CLI Operations
4.8.1. debug
Launch a command shell to debug a running application.
$ oc debug -h
When debugging images and setup problems, you can get an exact copy of a running pod configuration and troubleshoot with a shell. Since a failing pod may not be started and not accessible to rsh
or exec
, running the debug
command creates a carbon copy of that setup.
The default mode is to start a shell inside of the first container of the referenced pod, replication controller, or deployment configuration. The started pod will be a copy of your source pod, with labels stripped, the command changed to /bin/sh
, and readiness and liveness checks disabled. If you just want to run a command, add --
and a command to run. Passing a command will not create a TTY or send STDIN by default. Other flags are supported for altering the container or pod in common ways.
A common problem running containers is a security policy that prohibits you from running as a root user on the cluster. You can use this command to test running a pod as non-root (with --as-user
) or to run a non-root pod as root (with --as-root
).
The debug pod is deleted when the remote command completes or you interrupt the shell.
4.8.1.1. Usage
$ oc debug RESOURCE/NAME [ENV1=VAL1 ...] [-c CONTAINER] [options] [-- COMMAND]
4.8.1.2. Examples
To debug a currently running deployment:
$ oc debug dc/test
To test running a deployment as a non-root user:
$ oc debug dc/test --as-user=1000000
To debug a specific failing container by running the env
command in the second
container:
$ oc debug dc/test -c second -- /bin/env
To view the pod that would be created to debug:
$ oc debug dc/test -o yaml
4.8.2. logs
Retrieve the log output for a specific build, deployment, or pod. This command works for builds, build configurations, deployment configurations, and pods.
$ oc logs -f <pod>
4.8.3. exec
Execute a command in an already-running container. You can optionally specify a container ID, otherwise it defaults to the first container.
$ oc exec <pod> [-c <container>] <command>
4.8.4. rsh
Open a remote shell session to a container:
$ oc rsh <pod>
4.8.5. rsync
Copy the contents to or from a directory in an already-running pod container. If you do not specify a container, it defaults to the first container in the pod.
To copy contents from a local directory to a directory in a pod:
$ oc rsync <local_dir> <pod>:<pod_dir> -c <container>
To copy contents from a directory in a pod to a local directory:
$ oc rsync <pod>:<pod_dir> <local_dir> -c <container>
4.8.6. port-forward
Forward one or more local ports to a pod:
$ oc port-forward <pod> <local_port>:<remote_port>
4.8.7. proxy
Run a proxy to the Kubernetes API server:
$ oc proxy --port=<port> --www=<static_directory>
For security purposes, the oc exec
command does not work when accessing privileged containers except when the command is executed by a cluster-admin
user. Administrators can SSH into a node host, then use the docker exec
command on the desired container.
Chapter 5. Administrator CLI Operations
5.1. Overview
This topic provides information on the administrator CLI operations and their syntax. You must setup and login with the CLI before you can perform these operations.
The openshift
command is used for starting services that make up the OpenShift Container Platform cluster. For example, openshift start [master|node]
. However, it is also an all-in-one command that can perform all the same actions as the oc
and oc adm
commands via openshift cli
and openshift admin
respectively.
The administrator CLI differs from the normal set of commands under the developer CLI, which uses the oc
command, and is used more for project-level operations.
5.2. Common Operations
The administrator CLI allows interaction with the various objects that are managed by OpenShift Container Platform. Many common oc adm
operations are invoked using the following syntax:
$ oc adm <action> <option>
This specifies:
-
An
<action>
to perform, such asnew-project
orgroups
. -
An available
<option>
to perform the action on as well as a value for the option. Options include--output
.
When running oc adm
commands, you should run them only from the first master listed in the Ansible host inventory file, by default /etc/ansible/hosts.
5.3. Basic CLI Operations
5.3.1. new-project
Creates a new project:
$ oc adm new-project <project_name>
5.3.2. policy
Manages authorization policies:
$ oc adm policy
5.3.3. groups
Manages groups:
$ oc adm groups
5.4. Install CLI Operations
5.4.1. router
Installs a router:
$ ocadm router <router_name>
5.4.2. ipfailover
Installs an IP failover group for a set of nodes:
$ oc adm ipfailover <ipfailover_config>
5.4.3. registry
Installs an integrated container registry:
$ oc adm registry
5.5. Maintenance CLI Operations
5.5.1. build-chain
Outputs the inputs and dependencies of any builds:
$ oc adm build-chain <image_stream>[:<tag>]
5.5.2. manage-node
Manages nodes. For example, list or evacuate pods, or mark them ready:
$ oc adm manage-node
5.5.3. prune
Removes older versions of resources from the server:
$ oc adm prune
5.6. Settings CLI Operations
5.6.1. config
Changes kubelet configuration files:
$ oc adm config <subcommand>
5.6.2. create-kubeconfig
Creates a basic .kubeconfig file from client certificates:
$ oc adm create-kubeconfig
5.6.3. create-api-client-config
Creates a configuration file for connecting to the server as a user:
$ oc adm create-api-client-config
5.7. Advanced CLI Operations
5.7.1. create-bootstrap-project-template
Creates a bootstrap project template:
$ oc adm create-bootstrap-project-template
5.7.2. create-bootstrap-policy-file
Creates the default bootstrap policy:
$ oc adm create-bootstrap-policy-file
5.7.3. create-login-template
Creates a login template:
$ oc adm create-login-template
5.7.4. overwrite-policy
Resets the policy to the default values:
$ oc adm overwrite-policy
5.7.5. create-node-config
Creates a configuration bundle for a node:
$ oc adm create-node-config
5.7.6. ca
Manages certificates and keys:
$ oc adm ca
5.8. Other CLI Operations
5.8.1. version
Displays the version of the indicated object:
$ oc adm version
5.8.2. help
Displays help about any command:
$ oc adm help <command>
Chapter 6. Differences Between oc and kubectl
6.1. Why Use oc Over kubectl?
Kubernetes' command line interface (CLI), kubectl
, is used to run commands against any Kubernetes cluster. Because OpenShift Container Platform runs on top of a Kubernetes cluster, a copy of kubectl
is also included with oc
, OpenShift Container Platform’s command line interface (CLI).
Although there are several similarities between these two clients, this guide’s aim is to clarify the main reasons and scenarios for using one over the other.
6.2. Using oc
The oc
binary offers the same capabilities as the kubectl
binary, but it is further extended to natively support OpenShift Container Platform features, such as:
- Full support for OpenShift resources
-
Resources such as
DeploymentConfigs
,BuildConfigs
,Routes
,ImageStreams
, andImageStreamTags
are specific to OpenShift distributions, and not available in standard Kubernetes. - Authentication
-
The
oc
binary offers a built-inlogin
command which allows authentication. See developer authentication and configuring authentication for more information. - Additional commands
-
For example, the additional command
new-app
makes it easier to get new applications started using existing source code or pre-built images.
6.3. Using kubectl
The kubectl
binary is provided as a means to support existing workflows and scripts for new OpenShift Container Platform users coming from a standard Kubernetes environment. Existing users of kubectl
can continue to use the binary with no changes to the API, but should consider upgrading to oc
in order to gain the added functionality mentioned in the previous section.
Because oc
is built on top of kubectl
, converting a kubectl
binary to oc
is as simple as changing the binary’s name from kubectl
to oc
.
See Get Started with the CLI for installation and setup instructions.
Chapter 7. Extending the CLI
7.1. Overview
This topic reviews how to install and write extensions for the CLI. Usually called plug-ins or binary extensions, this feature allows you to extend the default set of oc
commands available and, therefore, allows you to perform new tasks.
A plug-in is a set of files: typically at least one plugin.yaml descriptor and one or more binary, script, or assets files.
CLI plug-ins are currently only available under the oc plugin
subcommand.
CLI plug-ins are currently a Technology Preview feature. Technology Preview features are not supported with Red Hat production service level agreements (SLAs), might not be functionally complete, and Red Hat does not recommend to use them for production. These features provide early access to upcoming product features, enabling customers to test functionality and provide feedback during the development process.
See the Red Hat Technology Preview features support scope for more information.
7.2. Prerequisites
You must have:
-
A working
oc
binary installed. -
An
oc
version of 3.7 or later (recommended).
7.3. Installing Plug-ins
Copy the plug-in’s plugin.yaml descriptor, binaries, scripts, and assets files to one of the locations in the file system where oc
searches for plug-ins.
Currently, OpenShift Container Platform does not provide a package manager for plug-ins. Therefore, it is your responsibility to place the plug-in files in the correct location. It is recommended that each plug-in is located on its own directory.
To install a plug-in that is distributed as a compressed file, extract it to one of the locations specified in The Plug-in Loader section.
7.3.1. The Plug-in Loader
The plug-in loader is responsible for searching plug-in files, and checking if the plug-in provides the minimum amount of information required for it to run. Files placed in the correct location that do not provide the minimum amount of information (for example, an incomplete plugin.yaml descriptor) are ignored.
7.3.1.1. Search Order
The plug-in loader uses the following search order:
${KUBECTL_PLUGINS_PATH}
If specified, the search stops here.
If the
KUBECTL_PLUGINS_PATH
environment variable is present, the loader uses it as the only location to look for plug-ins. TheKUBECTL_PLUGINS_PATH
environment variable is a list of directories. In Linux and Mac, the list is colon-delimited. In Windows, the list is semicolon-delimited.If
KUBECTL_PLUGINS_PATH
is not present, the loader begins to search the additional locations.${XDG_DATA_DIRS}/kubectl/plugins
The plug-in loader searches one or more directories specified according to the XDG System Directory Structure specification.
Specifically, the loader locates the directories specified by the
XDG_DATA_DIRS
environment variable. The plug-in loader searches the kubectl/plugins directory inside of directories specified by theXDG_DATA_DIRS
environment variable. IfXDG_DATA_DIRS
is not specified, it defaults to /usr/local/share:/usr/share.~/.kube/plugins
The
plugins
directory under the user’s kubeconfig directory. In most cases, this is ~/.kube/plugins:# Loads plugins from both /path/to/dir1 and /path/to/dir2 $ KUBECTL_PLUGINS_PATH=/path/to/dir1:/path/to/dir2 kubectl plugin -h
7.4. Writing Plug-ins
You can write a plug-in in any programming language or script that allows you to write CLI commands. A plug-in does not necessarily need to have a binary component. It could rely entirely on operating system utilities like echo
, sed
, or grep
. Alternatively, it could rely on the oc
binary.
The only strong requirement for an oc
plug-in is the plugin.yaml descriptor file. This file is responsible for declaring at least the minimum attributes required to register a plug-in and must be located under one of the locations specified in the Search Order section.
7.4.1. The plugin.yaml Descriptor
The descriptor file supports the following attributes:
name: "great-plugin" # REQUIRED: the plug-in command name, to be invoked under 'kubectl' shortDesc: "great-plugin plug-in" # REQUIRED: the command short description, for help longDesc: "" # the command long description, for help example: "" # command example(s), for help command: "./example" # REQUIRED: the command, binary, or script to invoke when running the plug-in flags: # flags supported by the plug-in - name: "flag-name" # REQUIRED for each flag: flag name shorthand: "f" # short version of the flag name desc: "example flag" # REQUIRED for each flag: flag description defValue: "extreme" # default value of the flag tree: # allows the declaration of subcommands - ... # subcommands support the same set of attributes
The preceding descriptor declares the great-plugin
plug-in, which has one flag named -f | --flag-name
. It could be invoked as:
$ oc plugin great-plugin -f value
When the plug-in is invoked, it calls the example
binary or script, which is located in the same directory as the descriptor file, passing a number of arguments and environment variables. The Accessing Runtime Attributes section describes how the example
command accesses the flag value and other runtime context.
7.4.2. Recommended Directory Structure
It is recommended that each plug-in has its own subdirectory in the file system, preferably with the same name as the plug-in command. The directory must contain the plugin.yaml descriptor and any binary, script, asset, or other dependency it might require.
For example, the directory structure for the great-plugin
plug-in could look like this:
~/.kube/plugins/ └── great-plugin ├── plugin.yaml └── example
7.4.3. Accessing Runtime Attributes
In most use cases, the binary or script file you write to support the plug-in must have access to some contextual information provided by the plug-in framework. For example, if you declared flags in the descriptor file, your plug-in must have access to the user-provided flag values at runtime.
The same is true for global flags. The plug-in framework is responsible for doing that, so plug-in writers do not need to worry about parsing arguments. This also ensures the best level of consistency between plug-ins and regular oc
commands.
Plug-ins have access to runtime context attributes through environment variables. To access the value provided through a flag, for example, look for the value of the proper environment variable using the appropriate function call for your binary or script.
The supported environment variables are:
-
KUBECTL_PLUGINS_CALLER
: The full path to theoc
binary that was used in the current command invocation. As a plug-in writer, you do not have to implement logic to authenticate and access the Kubernetes API. Instead, you can use the value provided by this environment variable to invokeoc
and obtain the information you need, using for exampleoc get --raw=/apis
. -
KUBECTL_PLUGINS_CURRENT_NAMESPACE
: The current namespace that is the context for this call. This is the actual namespace to be considered in namespaced operations, meaning it was already processed in terms of the precedence between what was provided through the kubeconfig, the--namespace
global flag, environment variables, and so on. -
KUBECTL_PLUGINS_DESCRIPTOR_*
: One environment variable for every attribute declared in the plugin.yaml descriptor. For example,KUBECTL_PLUGINS_DESCRIPTOR_NAME
,KUBECTL_PLUGINS_DESCRIPTOR_COMMAND
. -
KUBECTL_PLUGINS_GLOBAL_FLAG_*
: One environment variable for every global flag supported byoc
. For example,KUBECTL_PLUGINS_GLOBAL_FLAG_NAMESPACE
,KUBECTL_PLUGINS_GLOBAL_FLAG_LOGLEVEL
. -
KUBECTL_PLUGINS_LOCAL_FLAG_*
: One environment variable for every local flag declared in the plugin.yaml descriptor. For example,KUBECTL_PLUGINS_LOCAL_FLAG_HEAT
in the precedinggreat-plugin
example.
Legal Notice
Copyright © 2024 Red Hat, Inc.
OpenShift documentation is licensed under the Apache License 2.0 (https://www.apache.org/licenses/LICENSE-2.0).
Modified versions must remove all Red Hat trademarks.
Portions adapted from https://github.com/kubernetes-incubator/service-catalog/ with modifications by Red Hat.
Red Hat, Red Hat Enterprise Linux, the Red Hat logo, the Shadowman logo, JBoss, OpenShift, Fedora, the Infinity logo, and RHCE are trademarks of Red Hat, Inc., registered in the United States and other countries.
Linux® is the registered trademark of Linus Torvalds in the United States and other countries.
Java® is a registered trademark of Oracle and/or its affiliates.
XFS® is a trademark of Silicon Graphics International Corp. or its subsidiaries in the United States and/or other countries.
MySQL® is a registered trademark of MySQL AB in the United States, the European Union and other countries.
Node.js® is an official trademark of Joyent. Red Hat Software Collections is not formally related to or endorsed by the official Joyent Node.js open source or commercial project.
The OpenStack® Word Mark and OpenStack logo are either registered trademarks/service marks or trademarks/service marks of the OpenStack Foundation, in the United States and other countries and are used with the OpenStack Foundation’s permission. We are not affiliated with, endorsed or sponsored by the OpenStack Foundation, or the OpenStack community.
All other trademarks are the property of their respective owners.