API overview
Overview content for the OpenShift Container Platform API
Abstract
Chapter 1. Understanding API tiers
This guidance does not cover layered OpenShift Container Platform offerings.
API tiers for bare-metal configurations also apply to virtualized configurations except for any feature that directly interacts with hardware. Those features directly related to hardware have no application operating environment (AOE) compatibility level beyond that which is provided by the hardware vendor. For example, applications that rely on Graphics Processing Units (GPU) features are subject to the AOE compatibility provided by the GPU vendor driver.
API tiers in a cloud environment for cloud specific integration points have no API or AOE compatibility level beyond that which is provided by the hosting cloud vendor. For example, APIs that exercise dynamic management of compute, ingress, or storage are dependent upon the underlying API capabilities exposed by the cloud platform. Where a cloud vendor modifies a prerequisite API, Red Hat will provide commercially reasonable efforts to maintain support for the API with the capability presently offered by the cloud infrastructure vendor.
Red Hat requests that application developers validate that any behavior they depend on is explicitly defined in the formal API documentation to prevent introducing dependencies on unspecified implementation-specific behavior or dependencies on bugs in a particular implementation of an API. For example, new releases of an ingress router may not be compatible with older releases if an application uses an undocumented API or relies on undefined behavior.
1.1. API tiers
All commercially supported APIs, components, and features are associated under one of the following support levels:
API tier 1
APIs and application operating environments (AOEs) are stable within a major release. They may be deprecated within a major release, but they will not be removed until a subsequent major release.
API tier 2
APIs and AOEs are stable within a major release for a minimum of 9 months or 3 minor releases from the announcement of deprecation, whichever is longer.
API tier 3
This level applies to languages, tools, applications, and optional Operators included with OpenShift Container Platform through Operator Hub. Each component will specify a lifetime during which the API and AOE will be supported. Newer versions of language runtime specific components will attempt to be as API and AOE compatible from minor version to minor version as possible. Minor version to minor version compatibility is not guaranteed, however.
Components and developer tools that receive continuous updates through the Operator Hub, referred to as Operators and operands, should be considered API tier 3. Developers should use caution and understand how these components may change with each minor release. Users are encouraged to consult the compatibility guidelines documented by the component.
API tier 4
No compatibility is provided. API and AOE can change at any point. These capabilities should not be used by applications needing long-term support.
It is common practice for Operators to use custom resource definitions (CRDs) internally to accomplish a task. These objects are not meant for use by actors external to the Operator and are intended to be hidden. If any CRD is not meant for use by actors external to the Operator, the operators.operatorframework.io/internal-objects
annotation in the Operators ClusterServiceVersion
(CSV) should be specified to signal that the corresponding resource is internal use only and the CRD may be explicitly labeled as tier 4.
1.2. Mapping API tiers to API groups
For each API tier defined by Red Hat, we provide a mapping table for specific API groups where the upstream communities are committed to maintain forward compatibility. Any API group that does not specify an explicit compatibility level and is not specifically discussed below is assigned API tier 3 by default except for v1alpha1
APIs which are assigned tier 4 by default.
1.2.1. Support for Kubernetes API groups
API groups that end with the suffix *.k8s.io
or have the form version.<name>
with no suffix are governed by the Kubernetes deprecation policy and follow a general mapping between API version exposed and corresponding support tier unless otherwise specified.
API version example | API tier |
---|---|
| Tier 1 |
| Tier 2 |
| Tier 4 |
1.2.2. Support for OpenShift API groups
API groups that end with the suffix *.openshift.io
are governed by the OpenShift Container Platform deprecation policy and follow a general mapping between API version exposed and corresponding compatibility level unless otherwise specified.
API version example | API tier |
---|---|
| Tier 1 |
| Tier 1, some tier 1 deprecated |
| Tier 1, some tier 1 deprecated |
| Tier 1 |
| Tier 1 |
| Tier 1 |
| Tier 1 |
| Tier 1 |
| Tier 1 |
| Tier 1 |
| Tier 1 |
| Tier 1 |
| Tier 1 |
|
Tier 1 except for |
| Tier 1 |
| Tier 2 |
1.2.3. Support for Monitoring API groups
API groups that end with the suffix monitoring.coreos.com
have the following mapping:
API version example | API tier |
---|---|
| Tier 1 |
| Tier 1 |
| Tier 1 |
1.2.4. Support for Operator Lifecycle Manager API groups
Operator Lifecycle Manager (OLM) provides APIs that include API groups with the suffix operators.coreos.com
. These APIs have the following mapping:
API version example | API tier |
---|---|
| Tier 1 |
| Tier 1 |
| Tier 1 |
1.3. API deprecation policy
OpenShift Container Platform is composed of many components sourced from many upstream communities. It is anticipated that the set of components, the associated API interfaces, and correlated features will evolve over time and might require formal deprecation in order to remove the capability.
1.3.1. Deprecating parts of the API
OpenShift Container Platform is a distributed system where multiple components interact with a shared state managed by the cluster control plane through a set of structured APIs. Per Kubernetes conventions, each API presented by OpenShift Container Platform is associated with a group identifier and each API group is independently versioned. Each API group is managed in a distinct upstream community including Kubernetes, Metal3, Multus, Operator Framework, Open Cluster Management, OpenShift itself, and more.
While each upstream community might define their own unique deprecation policy for a given API group and version, Red Hat normalizes the community specific policy to one of the compatibility levels defined prior based on our integration in and awareness of each upstream community to simplify end-user consumption and support.
The deprecation policy and schedule for APIs vary by compatibility level.
The deprecation policy covers all elements of the API including:
- REST resources, also known as API objects
- Fields of REST resources
- Annotations on REST resources, excluding version-specific qualifiers
- Enumerated or constant values
Other than the most recent API version in each group, older API versions must be supported after their announced deprecation for a duration of no less than:
API tier | Duration |
---|---|
Tier 1 | Stable within a major release. They may be deprecated within a major release, but they will not be removed until a subsequent major release. |
Tier 2 | 9 months or 3 releases from the announcement of deprecation, whichever is longer. |
Tier 3 | See the component-specific schedule. |
Tier 4 | None. No compatibility is guaranteed. |
The following rules apply to all tier 1 APIs:
- API elements can only be removed by incrementing the version of the group.
- API objects must be able to round-trip between API versions without information loss, with the exception of whole REST resources that do not exist in some versions. In cases where equivalent fields do not exist between versions, data will be preserved in the form of annotations during conversion.
- API versions in a given group can not deprecate until a new API version at least as stable is released, except in cases where the entire API object is being removed.
1.3.2. Deprecating CLI elements
Client-facing CLI commands are not versioned in the same way as the API, but are user-facing component systems. The two major ways a user interacts with a CLI are through a command or flag, which is referred to in this context as CLI elements.
All CLI elements default to API tier 1 unless otherwise noted or the CLI depends on a lower tier API.
Element | API tier | |
---|---|---|
Generally available (GA) | Flags and commands | Tier 1 |
Technology Preview | Flags and commands | Tier 3 |
Developer Preview | Flags and commands | Tier 4 |
1.3.3. Deprecating an entire component
The duration and schedule for deprecating an entire component maps directly to the duration associated with the highest API tier of an API exposed by that component. For example, a component that surfaced APIs with tier 1 and 2 could not be removed until the tier 1 deprecation schedule was met.
API tier | Duration |
---|---|
Tier 1 | Stable within a major release. They may be deprecated within a major release, but they will not be removed until a subsequent major release. |
Tier 2 | 9 months or 3 releases from the announcement of deprecation, whichever is longer. |
Tier 3 | See the component-specific schedule. |
Tier 4 | None. No compatibility is guaranteed. |
Chapter 2. Understanding API compatibility guidelines
This guidance does not cover layered OpenShift Container Platform offerings.
2.1. API compatibility guidelines
Red Hat recommends that application developers adopt the following principles in order to improve compatibility with OpenShift Container Platform:
- Use APIs and components with support tiers that match the application’s need.
- Build applications using the published client libraries where possible.
- Applications are only guaranteed to run correctly if they execute in an environment that is as new as the environment it was built to execute against. An application that was built for OpenShift Container Platform 4.14 is not guaranteed to function properly on OpenShift Container Platform 4.13.
- Do not design applications that rely on configuration files provided by system packages or other components. These files can change between versions unless the upstream community is explicitly committed to preserving them. Where appropriate, depend on any Red Hat provided interface abstraction over those configuration files in order to maintain forward compatibility. Direct file system modification of configuration files is discouraged, and users are strongly encouraged to integrate with an Operator provided API where available to avoid dual-writer conflicts.
-
Do not depend on API fields prefixed with
unsupported<FieldName>
or annotations that are not explicitly mentioned in product documentation. - Do not depend on components with shorter compatibility guarantees than your application.
- Do not perform direct storage operations on the etcd server. All etcd access must be performed via the api-server or through documented backup and restore procedures.
Red Hat recommends that application developers follow the compatibility guidelines defined by Red Hat Enterprise Linux (RHEL). OpenShift Container Platform strongly recommends the following guidelines when building an application or hosting an application on the platform:
- Do not depend on a specific Linux kernel or OpenShift Container Platform version.
-
Avoid reading from
proc
,sys
, anddebug
file systems, or any other pseudo file system. -
Avoid using
ioctls
to directly interact with hardware. -
Avoid direct interaction with
cgroups
in order to not conflict with OpenShift Container Platform host-agents that provide the container execution environment.
During the lifecycle of a release, Red Hat makes commercially reasonable efforts to maintain API and application operating environment (AOE) compatibility across all minor releases and z-stream releases. If necessary, Red Hat might make exceptions to this compatibility goal for critical impact security or other significant issues.
2.2. API compatibility exceptions
The following are exceptions to compatibility in OpenShift Container Platform:
RHEL CoreOS file system modifications not made with a supported Operator
No assurances are made at this time that a modification made to the host operating file system is preserved across minor releases except for where that modification is made through the public interface exposed via a supported Operator, such as the Machine Config Operator or Node Tuning Operator.
Modifications to cluster infrastructure in cloud or virtualized environments
No assurances are made at this time that a modification to the cloud hosting environment that supports the cluster is preserved except for where that modification is made through a public interface exposed in the product or is documented as a supported configuration. Cluster infrastructure providers are responsible for preserving their cloud or virtualized infrastructure except for where they delegate that authority to the product through an API.
Functional defaults between an upgraded cluster and a new installation
No assurances are made at this time that a new installation of a product minor release will have the same functional defaults as a version of the product that was installed with a prior minor release and upgraded to the equivalent version. For example, future versions of the product may provision cloud infrastructure with different defaults than prior minor versions. In addition, different default security choices may be made in future versions of the product than those made in past versions of the product. Past versions of the product will forward upgrade, but preserve legacy choices where appropriate specifically to maintain backwards compatibility.
Usage of API fields that have the prefix "unsupported” or undocumented annotations
Select APIs in the product expose fields with the prefix unsupported<FieldName>
. No assurances are made at this time that usage of this field is supported across releases or within a release. Product support can request a customer to specify a value in this field when debugging specific problems, but its usage is not supported outside of that interaction. Usage of annotations on objects that are not explicitly documented are not assured support across minor releases.
API availability per product installation topology
The OpenShift distribution will continue to evolve its supported installation topology, and not all APIs in one install topology will necessarily be included in another. For example, certain topologies may restrict read/write access to particular APIs if they are in conflict with the product installation topology or not include a particular API at all if not pertinent to that topology. APIs that exist in a given topology will be supported in accordance with the compatibility tiers defined above.
2.3. API compatibility common terminology
2.3.1. Application Programming Interface (API)
An API is a public interface implemented by a software program that enables it to interact with other software. In OpenShift Container Platform, the API is served from a centralized API server and is used as the hub for all system interaction.
2.3.2. Application Operating Environment (AOE)
An AOE is the integrated environment that executes the end-user application program. The AOE is a containerized environment that provides isolation from the host operating system (OS). At a minimum, AOE allows the application to run in an isolated manner from the host OS libraries and binaries, but still share the same OS kernel as all other containers on the host. The AOE is enforced at runtime and it describes the interface between an application and its operating environment. It includes intersection points between the platform, operating system and environment, with the user application including projection of downward API, DNS, resource accounting, device access, platform workload identity, isolation among containers, isolation between containers and host OS.
The AOE does not include components that might vary by installation, such as Container Network Interface (CNI) plugin selection or extensions to the product such as admission hooks. Components that integrate with the cluster at a level below the container environment might be subjected to additional variation between versions.
2.3.3. Compatibility in a virtualized environment
Virtual environments emulate bare-metal environments such that unprivileged applications that run on bare-metal environments will run, unmodified, in corresponding virtual environments. Virtual environments present simplified abstracted views of physical resources, so some differences might exist.
2.3.4. Compatibility in a cloud environment
OpenShift Container Platform might choose to offer integration points with a hosting cloud environment via cloud provider specific integrations. The compatibility of these integration points are specific to the guarantee provided by the native cloud vendor and its intersection with the OpenShift Container Platform compatibility window. Where OpenShift Container Platform provides an integration with a cloud environment natively as part of the default installation, Red Hat develops against stable cloud API endpoints to provide commercially reasonable support with forward looking compatibility that includes stable deprecation policies. Example areas of integration between the cloud provider and OpenShift Container Platform include, but are not limited to, dynamic volume provisioning, service load balancer integration, pod workload identity, dynamic management of compute, and infrastructure provisioned as part of initial installation.
2.3.5. Major, minor, and z-stream releases
A Red Hat major release represents a significant step in the development of a product. Minor releases appear more frequently within the scope of a major release and represent deprecation boundaries that might impact future application compatibility. A z-stream release is an update to a minor release which provides a stream of continuous fixes to an associated minor release. API and AOE compatibility is never broken in a z-stream release except when this policy is explicitly overridden in order to respond to an unforeseen security impact.
For example, in the release 4.13.2:
- 4 is the major release version
- 13 is the minor release version
- 2 is the z-stream release version
2.3.6. Extended user support (EUS)
A minor release in an OpenShift Container Platform major release that has an extended support window for critical bug fixes. Users are able to migrate between EUS releases by incrementally adopting minor versions between EUS releases. It is important to note that the deprecation policy is defined across minor releases and not EUS releases. As a result, an EUS user might have to respond to a deprecation when migrating to a future EUS while sequentially upgrading through each minor release.
2.3.7. Developer Preview
An optional product capability that is not officially supported by Red Hat, but is intended to provide a mechanism to explore early phase technology. By default, Developer Preview functionality is opt-in, and subject to removal at any time. Enabling a Developer Preview feature might render a cluster unsupportable dependent upon the scope of the feature.
2.3.8. Technology Preview
An optional product capability that provides early access to upcoming product innovations to test functionality and provide feedback during the development process. The feature is not fully supported, might not be functionally complete, and is not intended for production use. Usage of a Technology Preview function requires explicit opt-in. Learn more about the Technology Preview Features Support Scope.
Chapter 3. Editing kubelet log level verbosity and gathering logs
To troubleshoot some issues with nodes, establish the kubelet’s log level verbosity depending on the issue to be tracked.
3.1. Modifying the kubelet as a one-time scenario
To modify the kubelet in a one-time scenario without rebooting the node due to the change of machine-config(spec":{"paused":false}})
, allowing you to modify the kubelet without affecting the service, follow this procedure.
Procedure
Connect to the node in debug mode:
$ oc debug node/<node>
$ chroot /host
Alternatively, it is possible to SSH to the node and become root.
After access is established, check the default log level:
$ systemctl cat kubelet
Example output
# /etc/systemd/system/kubelet.service.d/20-logging.conf [Service] Environment="KUBELET_LOG_LEVEL=2"
Define the new verbosity required in a new
/etc/systemd/system/kubelet.service.d/30-logging.conf
file, which overrides/etc/systemd/system/kubelet.service.d/20-logging.conf
. In this example, the verbosity is changed from2
to8
:$ echo -e "[Service]\nEnvironment=\"KUBELET_LOG_LEVEL=8\"" > /etc/systemd/system/kubelet.service.d/30-logging.conf
Reload systemd and restart the service:
$ systemctl daemon-reload
$ systemctl restart kubelet
Gather the logs, and then revert the log level increase:
$ rm -f /etc/systemd/system/kubelet.service.d/30-logging.conf
$ systemctl daemon-reload
$ systemctl restart kubelet
3.2. Persistent kubelet log level configuration
Procedure
Use the following
MachineConfig
object for persistent kubelet log level configuration:apiVersion: machineconfiguration.openshift.io/v1 kind: MachineConfig metadata: labels: machineconfiguration.openshift.io/role: master name: 99-master-kubelet-loglevel spec: config: ignition: version: 3.2.0 systemd: units: - name: kubelet.service enabled: true dropins: - name: 30-logging.conf contents: | [Service] Environment="KUBELET_LOG_LEVEL=2"
Generally, it is recommended to apply
0-4
as debug-level logs and5-8
as trace-level logs.
3.3. Log verbosity descriptions
Log verbosity | Description |
---|---|
| Always visible to an Operator. |
| A reasonable default log level if you do not want verbosity. |
| Useful steady state information about the service and important log messages that might correlate to significant changes in the system. This is the recommended default log level. |
| Extended information about changes. |
| Debug level verbosity. |
| Display requested resources. |
| Display HTTP request headers. |
| Display HTTP request contents. |
3.4. Gathering kubelet logs
Procedure
After the kubelet’s log level verbosity is configured properly, you can gather logs by running the following commands:
$ oc adm node-logs --role master -u kubelet
$ oc adm node-logs --role worker -u kubelet
Alternatively, inside the node, run the following command:
$ journalctl -b -f -u kubelet.service
To collect master container logs, run the following command:
$ sudo tail -f /var/log/containers/*
To directly gather the logs of all nodes, run the following command:
- for n in $(oc get node --no-headers | awk '{print $1}'); do oc adm node-logs $n | gzip > $n.log.gz; done
Chapter 4. API index
API | API group |
---|---|
policy.networking.k8s.io/v1alpha1 | |
k8s.ovn.org/v1 | |
monitoring.openshift.io/v1 | |
monitoring.coreos.com/v1 | |
monitoring.coreos.com/v1beta1 | |
monitoring.openshift.io/v1 | |
apiserver.openshift.io/v1 | |
config.openshift.io/v1 | |
apiregistration.k8s.io/v1 | |
quota.openshift.io/v1 | |
config.openshift.io/v1 | |
operator.openshift.io/v1 | |
metal3.io/v1alpha1 | |
policy.networking.k8s.io/v1alpha1 | |
v1 | |
metal3.io/v1alpha1 | |
template.openshift.io/v1 | |
build.openshift.io/v1 | |
config.openshift.io/v1 | |
build.openshift.io/v1 | |
build.openshift.io/v1 | |
build.openshift.io/v1 | |
operators.coreos.com/v1alpha1 | |
certificates.k8s.io/v1 | |
operator.openshift.io/v1 | |
cloud.network.openshift.io/v1 | |
autoscaling.openshift.io/v1 | |
operator.openshift.io/v1 | |
config.openshift.io/v1 | |
quota.openshift.io/v1 | |
authorization.openshift.io/v1 | |
rbac.authorization.k8s.io/v1 | |
authorization.openshift.io/v1 | |
rbac.authorization.k8s.io/v1 | |
operators.coreos.com/v1alpha1 | |
config.openshift.io/v1 | |
v1 | |
imageregistry.operator.openshift.io/v1 | |
operator.openshift.io/v1 | |
samples.operator.openshift.io/v1 | |
v1 | |
config.openshift.io/v1 | |
operator.openshift.io/v1 | |
console.openshift.io/v1 | |
console.openshift.io/v1 | |
console.openshift.io/v1 | |
console.openshift.io/v1 | |
console.openshift.io/v1 | |
console.openshift.io/v1 | |
console.openshift.io/v1 | |
console.openshift.io/v1 | |
machineconfiguration.openshift.io/v1 | |
machineconfiguration.openshift.io/v1 | |
apps/v1 | |
machine.openshift.io/v1 | |
cloudcredential.openshift.io/v1 | |
batch/v1 | |
storage.k8s.io/v1 | |
storage.k8s.io/v1 | |
operator.openshift.io/v1 | |
storage.k8s.io/v1 | |
apiextensions.k8s.io/v1 | |
apps/v1 | |
metal3.io/v1alpha1 | |
apps/v1 | |
apps.openshift.io/v1 | |
apps.openshift.io/v1 | |
apps.openshift.io/v1 | |
apps.openshift.io/v1 | |
config.openshift.io/v1 | |
operator.openshift.io/v1 | |
ingress.operator.openshift.io/v1 | |
k8s.ovn.org/v1 | |
k8s.ovn.org/v1 | |
k8s.ovn.org/v1 | |
network.operator.openshift.io/v1 | |
k8s.ovn.org/v1 | |
v1 | |
discovery.k8s.io/v1 | |
operator.openshift.io/v1 | |
v1 | |
events.k8s.io/v1 | |
policy/v1 | |
config.openshift.io/v1 | |
metal3.io/v1alpha1 | |
flowcontrol.apiserver.k8s.io/v1 | |
user.openshift.io/v1 | |
metal3.io/v1alpha1 | |
helm.openshift.io/v1beta1 | |
autoscaling/v2 | |
metal3.io/v1alpha1 | |
metal3.io/v1alpha1 | |
user.openshift.io/v1 | |
config.openshift.io/v1 | |
image.openshift.io/v1 | |
config.openshift.io/v1 | |
operator.openshift.io/v1alpha1 | |
config.openshift.io/v1 | |
imageregistry.operator.openshift.io/v1 | |
image.openshift.io/v1 | |
image.openshift.io/v1 | |
image.openshift.io/v1 | |
image.openshift.io/v1 | |
image.openshift.io/v1 | |
image.openshift.io/v1 | |
image.openshift.io/v1 | |
image.openshift.io/v1 | |
config.openshift.io/v1 | |
config.openshift.io/v1 | |
config.openshift.io/v1 | |
networking.k8s.io/v1 | |
networking.k8s.io/v1 | |
operator.openshift.io/v1 | |
operator.openshift.io/v1 | |
operators.coreos.com/v1alpha1 | |
ipam.cluster.x-k8s.io/v1beta1 | |
ipam.cluster.x-k8s.io/v1beta1 | |
whereabouts.cni.cncf.io/v1alpha1 | |
batch/v1 | |
operator.openshift.io/v1 | |
operator.openshift.io/v1 | |
machineconfiguration.openshift.io/v1 | |
operator.openshift.io/v1 | |
operator.openshift.io/v1 | |
coordination.k8s.io/v1 | |
v1 | |
authorization.openshift.io/v1 | |
authorization.k8s.io/v1 | |
authorization.openshift.io/v1 | |
machine.openshift.io/v1beta1 | |
autoscaling.openshift.io/v1beta1 | |
machineconfiguration.openshift.io/v1 | |
machineconfiguration.openshift.io/v1 | |
operator.openshift.io/v1 | |
machine.openshift.io/v1beta1 | |
machine.openshift.io/v1beta1 | |
infrastructure.cluster.x-k8s.io/v1beta1 | |
infrastructure.cluster.x-k8s.io/v1beta1 | |
admissionregistration.k8s.io/v1 | |
v1 | |
config.openshift.io/v1 | |
operator.openshift.io/v1 | |
k8s.cni.cncf.io/v1 | |
networking.k8s.io/v1 | |
v1 | |
config.openshift.io/v1 | |
config.openshift.io/v1 | |
oauth.openshift.io/v1 | |
oauth.openshift.io/v1 | |
oauth.openshift.io/v1 | |
oauth.openshift.io/v1 | |
operators.coreos.com/v1 | |
operator.openshift.io/v1 | |
operator.openshift.io/v1 | |
operators.coreos.com/v1 | |
operators.coreos.com/v2 | |
operators.coreos.com/v1 | |
config.openshift.io/v1 | |
network.operator.openshift.io/v1 | |
whereabouts.cni.cncf.io/v1alpha1 | |
packages.operators.coreos.com/v1 | |
performance.openshift.io/v2 | |
v1 | |
v1 | |
v1 | |
policy/v1 | |
monitoring.coreos.com/v1 | |
controlplane.operator.openshift.io/v1alpha1 | |
security.openshift.io/v1 | |
security.openshift.io/v1 | |
security.openshift.io/v1 | |
v1 | |
metal3.io/v1alpha1 | |
scheduling.k8s.io/v1 | |
flowcontrol.apiserver.k8s.io/v1 | |
monitoring.coreos.com/v1 | |
tuned.openshift.io/v1 | |
config.openshift.io/v1 | |
project.openshift.io/v1 | |
helm.openshift.io/v1beta1 | |
project.openshift.io/v1 | |
monitoring.coreos.com/v1 | |
monitoring.coreos.com/v1 | |
metal3.io/v1alpha1 | |
config.openshift.io/v1 | |
security.openshift.io/v1 | |
apps/v1 | |
v1 | |
authorization.openshift.io/v1 | |
v1 | |
authorization.openshift.io/v1 | |
rbac.authorization.k8s.io/v1 | |
authorization.openshift.io/v1 | |
rbac.authorization.k8s.io/v1 | |
authorization.openshift.io/v1 | |
route.openshift.io/v1 | |
node.k8s.io/v1 | |
autoscaling/v1 | |
config.openshift.io/v1 | |
v1 | |
image.openshift.io/v1 | |
security.openshift.io/v1 | |
authorization.k8s.io/v1 | |
authentication.k8s.io/v1 | |
authorization.k8s.io/v1 | |
authorization.openshift.io/v1 | |
v1 | |
v1 | |
operator.openshift.io/v1 | |
monitoring.coreos.com/v1 | |
apps/v1 | |
operator.openshift.io/v1 | |
storage.k8s.io/v1 | |
migration.k8s.io/v1alpha1 | |
migration.k8s.io/v1alpha1 | |
authorization.k8s.io/v1 | |
authorization.openshift.io/v1 | |
authorization.openshift.io/v1 | |
operators.coreos.com/v1alpha1 | |
template.openshift.io/v1 | |
template.openshift.io/v1 | |
monitoring.coreos.com/v1 | |
authentication.k8s.io/v1 | |
authentication.k8s.io/v1 | |
tuned.openshift.io/v1 | |
user.openshift.io/v1 | |
user.openshift.io/v1 | |
oauth.openshift.io/v1 | |
admissionregistration.k8s.io/v1 | |
storage.k8s.io/v1 | |
snapshot.storage.k8s.io/v1 | |
snapshot.storage.k8s.io/v1 | |
snapshot.storage.k8s.io/v1 |
Legal Notice
Copyright © 2024 Red Hat, Inc.
OpenShift documentation is licensed under the Apache License 2.0 (https://www.apache.org/licenses/LICENSE-2.0).
Modified versions must remove all Red Hat trademarks.
Portions adapted from https://github.com/kubernetes-incubator/service-catalog/ with modifications by Red Hat.
Red Hat, Red Hat Enterprise Linux, the Red Hat logo, the Shadowman logo, JBoss, OpenShift, Fedora, the Infinity logo, and RHCE are trademarks of Red Hat, Inc., registered in the United States and other countries.
Linux® is the registered trademark of Linus Torvalds in the United States and other countries.
Java® is a registered trademark of Oracle and/or its affiliates.
XFS® is a trademark of Silicon Graphics International Corp. or its subsidiaries in the United States and/or other countries.
MySQL® is a registered trademark of MySQL AB in the United States, the European Union and other countries.
Node.js® is an official trademark of Joyent. Red Hat Software Collections is not formally related to or endorsed by the official Joyent Node.js open source or commercial project.
The OpenStack® Word Mark and OpenStack logo are either registered trademarks/service marks or trademarks/service marks of the OpenStack Foundation, in the United States and other countries and are used with the OpenStack Foundation’s permission. We are not affiliated with, endorsed or sponsored by the OpenStack Foundation, or the OpenStack community.
All other trademarks are the property of their respective owners.