This documentation is for a release that is no longer maintained
See documentation for the latest supported version 3 or the latest supported version 4.이 콘텐츠는 선택한 언어로 제공되지 않습니다.
Container-native virtualization
Container-native virtualization installation, usage, and release notes
Abstract
Chapter 1. About container-native virtualization 링크 복사링크가 클립보드에 복사되었습니다!
Learn about container-native virtualization’s capabilities and support scope.
1.1. What you can do with container-native virtualization 링크 복사링크가 클립보드에 복사되었습니다!
Container-native virtualization is an add-on to OpenShift Container Platform that allows you to run and manage virtual machine workloads alongside container workloads.
Container-native virtualization adds new objects into your OpenShift Container Platform cluster via Kubernetes custom resources to enable virtualization tasks. These tasks include:
- Creating and managing Linux and Windows virtual machines
- Connecting to virtual machines through a variety of consoles and CLI tools
- Importing and cloning existing virtual machines
- Managing network interface controllers and storage disks attached to virtual machines
- Live migrating virtual machines between nodes
An enhanced web console provides a graphical portal to manage these virtualized resources alongside the OpenShift Container Platform cluster containers and infrastructure.
1.2. Container-native virtualization support 링크 복사링크가 클립보드에 복사되었습니다!
container-native virtualization is a Technology Preview feature only. Technology Preview features are not supported with Red Hat production service level agreements (SLAs) and might not be functionally complete. Red Hat does not recommend using them in production. These features provide early access to upcoming product features, enabling customers to test functionality and provide feedback during the development process.
For more information about the support scope of Red Hat Technology Preview features, see https://access.redhat.com/support/offerings/techpreview/.
Chapter 2. Container-native virtualization release notes 링크 복사링크가 클립보드에 복사되었습니다!
2.1. Container-native virtualization release notes 링크 복사링크가 클립보드에 복사되었습니다!
2.1.1. About container-native virtualization 2.2 링크 복사링크가 클립보드에 복사되었습니다!
2.1.1.1. What you can do with container-native virtualization 링크 복사링크가 클립보드에 복사되었습니다!
Container-native virtualization is an add-on to OpenShift Container Platform that allows you to run and manage virtual machine workloads alongside container workloads.
Container-native virtualization adds new objects into your OpenShift Container Platform cluster via Kubernetes custom resources to enable virtualization tasks. These tasks include:
- Creating and managing Linux and Windows virtual machines
- Connecting to virtual machines through a variety of consoles and CLI tools
- Importing and cloning existing virtual machines
- Managing network interface controllers and storage disks attached to virtual machines
- Live migrating virtual machines between nodes
An enhanced web console provides a graphical portal to manage these virtualized resources alongside the OpenShift Container Platform cluster containers and infrastructure.
2.1.1.2. Container-native virtualization support 링크 복사링크가 클립보드에 복사되었습니다!
container-native virtualization is a Technology Preview feature only. Technology Preview features are not supported with Red Hat production service level agreements (SLAs) and might not be functionally complete. Red Hat does not recommend using them in production. These features provide early access to upcoming product features, enabling customers to test functionality and provide feedback during the development process.
For more information about the support scope of Red Hat Technology Preview features, see https://access.redhat.com/support/offerings/techpreview/.
2.1.2. New and changed features 링크 복사링크가 클립보드에 복사되었습니다!
Managing virtual machines is simpler and more efficient due to improvements in design and workflow. You can now:
- Run the virtual machine wizard with less navigation. The wizard now uses a comprehensive in-page style and includes a review page for confirming configuration details before submission.
- Import a single VMware virtual machine with less navigation.
- Edit virtual machine templates as well as virtual machine configurations.
- Monitor health of virtual machine-backed services as you would for Pod-based services.
- Enable persistent local storage for virtual machine images.
- Add, edit, and view virtual CD-ROM devices attached to a virtual machine.
- Add and view network attachment definitions with a graphical editor.
2.1.3. Resolved issues 링크 복사링크가 클립보드에 복사되었습니다!
-
Previously, when you added a disk to a virtual machine via the Disks tab in the web console, the added disk had a
FilesystemvolumeMode regardless of the volumeMode set in thekubevirt-storage-class-defaultConfigMap. This issue has been fixed. (BZ#1753688) - Previously, when navigating to the Virtual Machines Console tab, sometimes no content was displayed. This issue has been fixed. (BZ#1753606)
- Previously, attempting to list all instances of the container-native virtualization operator from a browser resulted in a 404 (page not found) error. This issue has been fixed. (BZ#1757526)
-
Previously, if a virtual machine used guaranteed CPUs, it was not scheduled because the label
cpumanager=truewas not automatically set on nodes. This issue has been fixed. (BZ#1718944)
2.1.4. Known issues 링크 복사링크가 클립보드에 복사되었습니다!
- If you have container-native virtualization 2.1.0 deployed, you must first upgrade container-native virtualization to 2.2.0 before upgrading OpenShift Container Platform. Upgrading OpenShift Container Platform before upgrading container-native virtualization might trigger virtual machine deletion. (BZ#1785661)
-
The
masqueradebinding method for virtual machines cannot be used in clusters with RHEL 7 compute nodes. (BZ#1741626) After migration, a virtual machine is assigned a new IP address. However, the commands
oc get vmiandoc describe vmistill generate output containing the obsolete IP address. (BZ#1686208)As a workaround, view the correct IP address by running the following command:
oc get pod -o wide
$ oc get pod -o wideCopy to Clipboard Copied! Toggle word wrap Toggle overflow
- Some resources are improperly retained when removing container-native virtualization. You must manually remove these resources in order to reinstall container-native virtualization. (BZ#1712429)
Users without administrator privileges cannot add a network interface to a project in an L2 network using the virtual machine wizard. This issue is caused by missing permissions that allow users to load network attachment definitions. (BZ#1743985)
As a workaround, provide the user with permissions to load the network attachment definitions.
Define
ClusterRoleandClusterRoleBindingobjects to the YAML configuration file, using the following examples:Copy to Clipboard Copied! Toggle word wrap Toggle overflow Copy to Clipboard Copied! Toggle word wrap Toggle overflow As a
cluster-adminuser, run the following command to create theClusterRoleandClusterRoleBindingobjects you defined:oc create -f <filename>.yaml
$ oc create -f <filename>.yamlCopy to Clipboard Copied! Toggle word wrap Toggle overflow
Live migration fails when nodes have different CPU models. Even in cases where nodes have the same physical CPU model, differences introduced by microcode updates have the same effect. This is because the default settings trigger host CPU passthrough behavior, which is incompatible with live migration. (BZ#1760028)
As a workaround, set the default CPU model in the
kubevirt-configConfigMap, as shown in the following example:NoteYou must make this change before starting the virtual machines that support live migration.
Open the
kubevirt-configConfigMap for editing by running the following command:oc edit configmap kubevirt-config -n openshift-cnv
$ oc edit configmap kubevirt-config -n openshift-cnvCopy to Clipboard Copied! Toggle word wrap Toggle overflow Edit the ConfigMap:
kind: ConfigMap metadata: name: kubevirt-config data: default-cpu-model: "<cpu-model>"
kind: ConfigMap metadata: name: kubevirt-config data: default-cpu-model: "<cpu-model>"1 Copy to Clipboard Copied! Toggle word wrap Toggle overflow - 1
- Replace
<cpu-model>with the actual CPU model value. You can determine this value by runningoc describe node <node>for all nodes and looking at thecpu-model-<name>labels. Select the CPU model that is present on all of your nodes.
When running
virtctl image-uploadto upload large VM disk images inqcow2format, an end-of-file (EOF) error may be reported after the data is transmitted, even though the upload is either progressing normally or completed. (BZ#1789093)Run the following command to check the status of an upload on a given PVC:
oc describe pvc <pvc-name> | grep cdi.kubevirt.io/storage.pod.phase
$ oc describe pvc <pvc-name> | grep cdi.kubevirt.io/storage.pod.phaseCopy to Clipboard Copied! Toggle word wrap Toggle overflow When attempting to create and launch a virtual machine using a Haswell CPU, the launch of the virtual machine can fail due to incorrectly labeled nodes. This is a change in behavior from previous versions of container-native virtualization, where virtual machines could be successfully launched on Haswell hosts. (BZ#1781497)
As a workaround, select a different CPU model, if possible.
- If you select a directory that shares space with your operating system, you can potentially exhaust the space on the partition, causing the node to be non-functional. Instead, create a separate partition and point the hostpath provisioner to that partition so it will not interfere with your operating system. (BZ#1793132)
- The container-native virtualization upgrade process occasionally fails due to an interruption from the Operator Lifecycle Manager (OLM). This issue is caused by the limitations associated with using a declarative API to track the state of container-native virtualization Operators. Enabling automatic updates during installation decreases the risk of encountering this issue. (BZ#1759612)
-
Container-native virtualization cannot reliably identify node drains that are triggered by running either
oc adm drainorkubectl drain. Do not run these commands on the nodes of any clusters where container-native virtualization is deployed. The nodes might not drain if there are virtual machines running on top of them. The current solution is to put nodes into maintenance. (BZ#1707427) If you navigate to the Subscription tab on the Operators → Installed Operators page and click the current upgrade channel to edit it, there might be no visible results. If this occurs, there are no visible errors. (BZ#1796410)
As a workaround, trigger the upgrade process to container-native virtualization 2.2 from the CLI by running the following
ocpatch command:export TARGET_NAMESPACE=openshift-cnv CNV_CHANNEL=2.2 && oc patch -n "${TARGET_NAMESPACE}" $(oc get subscription -n ${TARGET_NAMESPACE} --no-headers -o name) --type='json' -p='[{"op": "replace", "path": "/spec/channel", "value":"'${CNV_CHANNEL}'"}, {"op": "replace", "path": "/spec/installPlanApproval", "value":"Automatic"}]'$ export TARGET_NAMESPACE=openshift-cnv CNV_CHANNEL=2.2 && oc patch -n "${TARGET_NAMESPACE}" $(oc get subscription -n ${TARGET_NAMESPACE} --no-headers -o name) --type='json' -p='[{"op": "replace", "path": "/spec/channel", "value":"'${CNV_CHANNEL}'"}, {"op": "replace", "path": "/spec/installPlanApproval", "value":"Automatic"}]'Copy to Clipboard Copied! Toggle word wrap Toggle overflow This command points your subscription to upgrade channel
2.2and enables automatic updates.
Chapter 3. Container-native virtualization installation 링크 복사링크가 클립보드에 복사되었습니다!
3.1. Configuring your cluster for container-native virtualization 링크 복사링크가 클립보드에 복사되었습니다!
To obtain an evaluation version of OpenShift Container Platform, download a trial from the OpenShift Container Platform home page.
Container-native virtualization works with OpenShift Container Platform by default, however the following installation configurations are recommended:
- The OpenShift Container Platform cluster is installed on bare metal. Manage your Compute nodes in accordance with the number and size of the virtual machines to host in the cluster.
- Monitoring is configured in the cluster.
3.2. Installing container-native virtualization 링크 복사링크가 클립보드에 복사되었습니다!
Install container-native virtualization to add virtualization functionality to your OpenShift Container Platform cluster.
You can use the OpenShift Container Platform 4.3 web console to subscribe to and deploy the container-native virtualization Operators.
3.2.1. Prerequisites 링크 복사링크가 클립보드에 복사되었습니다!
- OpenShift Container Platform 4.3
Container-native virtualization is a Technology Preview feature only. Technology Preview features are not supported with Red Hat production service level agreements (SLAs) and might not be functionally complete. Red Hat does not recommend using them in production. These features provide early access to upcoming product features, enabling customers to test functionality and provide feedback during the development process.
For more information about the support scope of Red Hat Technology Preview features, see https://access.redhat.com/support/offerings/techpreview/.
3.2.2. Preparing to install container-native virtualization 링크 복사링크가 클립보드에 복사되었습니다!
Before installing container-native virtualization, create a namespace that is named openshift-cnv.
Prerequisites
-
User with
cluster-adminprivileges
Procedure
- From the OpenShift Container Platform web console, navigate to the Administration → Namespaces page.
- Click Create Namespace.
-
In the Name field, type
openshift-cnv. - Click Create.
3.2.3. Subscribing to the Container-native virtualization catalog 링크 복사링크가 클립보드에 복사되었습니다!
Before you install container-native virtualization, subscribe to the Container-native virtualization catalog from the OpenShift Container Platform web console. Subscribing gives the openshift-cnv namespace access to the container-native virtualization Operators.
Prerequisites
-
Create a namespace that is named
openshift-cnv.
Procedure
- Open a browser window and log in to the OpenShift Container Platform web console.
- Navigate to the Operators → OperatorHub page.
- Search for Container-native virtualization and then select it.
- Read the information about the Operator and click Install.
On the Create Operator Subscription page:
Select A specific namespace on the cluster from the Installation Mode list and choose the
openshift-cnvnamespace.Warning-
All namespaces on the cluster (default) installs the Operator in the default
openshift-operatorsnamespace to watch and be made available to all namespaces in the cluster. This option is not supported for use with container-native virtualization. You must only install the Operator in theopenshift-cnvnamespace.
-
All namespaces on the cluster (default) installs the Operator in the default
- Select 2.2 from the list of available Update Channel options.
- For Approval Strategy, ensure that Automatic, which is the default value, is selected. Container-native virtualization automatically updates when a new z-stream release is available.
-
Click Subscribe to make the Operator available to the
openshift-cnvnamespace.
3.2.4. Deploying container-native virtualization 링크 복사링크가 클립보드에 복사되었습니다!
After subscribing to the Container-native virtualization catalog, create the KubeVirt HyperConverged Cluster Operator Deployment custom resource to deploy container-native virtualization.
Prerequisites
-
An active subscription to the Container-native virtualization catalog in the
openshift-cnvnamespace
Procedure
- Navigate to the Operators → Installed Operators page.
- Click Container-native virtualization.
Click the KubeVirt HyperConverged Cluster Operator Deployment tab and click Create HyperConverged Cluster.
WarningTo avoid deployment errors, do not rename the custom resource. Before you proceed to the next step, ensure that the custom resource is named the default
kubevirt-hyperconverged.- Click Create to launch container-native virtualization.
- Navigate to the Workloads → Pods page and monitor the container-native virtualization Pods until they are all Running. After all the Pods display the Running state, you can access container-native virtualization.
3.3. Installing the virtctl client 링크 복사링크가 클립보드에 복사되었습니다!
The virtctl client is a command-line utility for managing container-native virtualization resources.
Install the client to your system by enabling the container-native virtualization repository and installing the kubevirt-virtctl package.
3.3.1. Enabling container-native virtualization repositories 링크 복사링크가 클립보드에 복사되었습니다!
Red Hat offers container-native virtualization repositories for both Red Hat Enterprise Linux 8 and Red Hat Enterprise Linux 7:
-
Red Hat Enterprise Linux 8 repository:
cnv-2.2-for-rhel-8-x86_64-rpms -
Red Hat Enterprise Linux 7 repository:
rhel-7-server-cnv-2.2-rpms
The process for enabling the repository in subscription-manager is the same in both platforms.
Procedure
Use
subscription managerto enable the appropriate container-native virtualization repository for your system:subscription-manager repos --enable <repository>
# subscription-manager repos --enable <repository>Copy to Clipboard Copied! Toggle word wrap Toggle overflow
3.3.2. Installing the virtctl client 링크 복사링크가 클립보드에 복사되었습니다!
Install the virtctl client from the kubevirt-virtctl package.
Procedure
Install the
kubevirt-virtctlpackage:yum install kubevirt-virtctl
# yum install kubevirt-virtctlCopy to Clipboard Copied! Toggle word wrap Toggle overflow
See also: Using the CLI tools for container-native virtualization.
3.4. Uninstalling container-native virtualization 링크 복사링크가 클립보드에 복사되었습니다!
You can uninstall container-native virtualization by using the OpenShift Container Platform web console.
3.4.1. Prerequisites 링크 복사링크가 클립보드에 복사되었습니다!
- Container-native virtualization 2.2
3.4.2. Deleting the KubeVirt HyperConverged Cluster Operator Deployment custom resource 링크 복사링크가 클립보드에 복사되었습니다!
To uninstall container-native virtualization, you must first delete the KubeVirt HyperConverged Cluster Operator Deployment custom resource.
Prerequisites
- An active KubeVirt HyperConverged Cluster Operator Deployment custom resource
Procedure
-
From the OpenShift Container Platform web console, select
openshift-cnvfrom the Projects list. - Navigate to the Operators → Installed Operators page.
- Click Container-native virtualization.
- Click the KubeVirt HyperConverged Cluster Operator Deployment tab.
-
Click the Options menu
in the row containing the kubevirt-hyperconverged custom resource. In the expanded menu, click Delete HyperConverged Cluster.
- Click Delete in the confirmation window.
- Navigate to the Workloads → Pods page to verify that only the Operator Pods are running.
Open a terminal window and clean up the remaining resources by running the following command:
oc delete apiservices v1alpha3.subresources.kubevirt.io -n openshift-cnv
$ oc delete apiservices v1alpha3.subresources.kubevirt.io -n openshift-cnvCopy to Clipboard Copied! Toggle word wrap Toggle overflow
3.4.3. Deleting the Container-native virtualization catalog subscription 링크 복사링크가 클립보드에 복사되었습니다!
To finish uninstalling container-native virtualization, delete the Container-native virtualization catalog subscription.
Prerequisites
- An active subscription to the Container-native virtualization catalog
Procedure
- Navigate to the Operators → OperatorHub page.
- Search for Container-native virtualization and then select it.
- Click Uninstall.
You can now delete the openshift-cnv namespace.
3.4.4. Deleting a namespace using the web console 링크 복사링크가 클립보드에 복사되었습니다!
You can delete a namespace by using the OpenShift Container Platform web console.
If you do not have permissions to delete the namespace, the Delete Namespace option is not available.
Procedure
- Navigate to Administration → Namespaces.
- Locate the namespace that you want to delete in the list of namespaces.
-
On the far right side of the namespace listing, select Delete Namespace from the Options menu
.
- When the Delete Namespace pane opens, enter the name of the namespace that you want to delete in the field.
- Click Delete.
Chapter 4. Upgrading container-native virtualization 링크 복사링크가 클립보드에 복사되었습니다!
You can manually upgrade to the next minor version of container-native virtualization and monitor the status of an update by using the web console.
Container-native virtualization is a Technology Preview feature only. Technology Preview features are not supported with Red Hat production service level agreements (SLAs) and might not be functionally complete. Red Hat does not recommend using them in production. These features provide early access to upcoming product features, enabling customers to test functionality and provide feedback during the development process.
For more information about the support scope of Red Hat Technology Preview features, see https://access.redhat.com/support/offerings/techpreview/.
4.1. About upgrading container-native virtualization 링크 복사링크가 클립보드에 복사되었습니다!
4.1.1. How container-native virtualization upgrades work 링크 복사링크가 클립보드에 복사되었습니다!
- You can upgrade to the next minor version of container-native virtualization by using the OpenShift Container Platform web console to change the channel of your Operator subscription.
- You can enable automatic z-stream updates during container-native virtualization installation.
- Updates are delivered via the Marketplace Operator, which is deployed during OpenShift Container Platform installation. The Marketplace Operator makes external Operators available to your cluster.
- The amount of time an update takes to complete depends on your network connection. Most automatic updates complete within fifteen minutes.
4.1.2. How container-native virtualization upgrades affect your cluster 링크 복사링크가 클립보드에 복사되었습니다!
Upgrading does not interrupt virtual machine workloads.
Virtual machine Pods are not restarted or migrated during an upgrade. If you need to update the
virt-launcherPod, you must restart or live migrate the virtual machine.NoteEach virtual machine has a
virt-launcherPod that runs the virtual machine instance. Thevirt-launcherPod runs an instance oflibvirt, which is used to manage the virtual machine process.
- Upgrading does not interrupt network connections.
- DataVolumes and their associated PersistentVolumeClaims are preserved during upgrade.
4.2. Upgrading container-native virtualization to the next minor version 링크 복사링크가 클립보드에 복사되었습니다!
You can manually upgrade container-native virtualization to the next minor version by using the OpenShift Container Platform web console to change the channel of your Operator subscription.
Prerequisites
-
Access to the cluster as a user with the
cluster-adminrole.
Procedure
- Access the OpenShift Container Platform web console and navigate to Operators → Installed Operators.
- Click Container-native virtualization to open the Operator Details page.
- Click the Subscription tab to open the Subscription Overview page.
- In the Channel pane, click the pencil icon on the right side of the version number to open the Change Subscription Update Channel window.
- Select the next minor version. For example, if you want to upgrade to container-native virtualization 2.2, select 2.2.
- Click Save.
Check the status of the upgrade by navigating to Operators → Installed Operators. You can also check the status by running the following
occommand:oc get csv -n openshift-cnv
$ oc get csv -n openshift-cnvCopy to Clipboard Copied! Toggle word wrap Toggle overflow
4.3. Monitoring upgrade status 링크 복사링크가 클립보드에 복사되었습니다!
The best way to monitor container-native virtualization upgrade status is to watch the ClusterServiceVersion (CSV) PHASE. You can also monitor the CSV conditions in the web console or by running the command provided here.
The PHASE and conditions values are approximations that are based on available information.
Prerequisites
-
Access to the cluster as a user with the
cluster-adminrole. -
Install the OpenShift CLI (
oc).
Procedure
Run the following command:
oc get csv
$ oc get csvCopy to Clipboard Copied! Toggle word wrap Toggle overflow Review the output, checking the
PHASEfield. For example:VERSION REPLACES PHASE 2.2.1 kubevirt-hyperconverged-operator.v2.2.0 Installing 2.2.0 Replacing
VERSION REPLACES PHASE 2.2.1 kubevirt-hyperconverged-operator.v2.2.0 Installing 2.2.0 ReplacingCopy to Clipboard Copied! Toggle word wrap Toggle overflow Optional: Monitor the aggregated status of all container-native virtualization component conditions by running the following command:
oc get hco -n openshift-cnv kubevirt-hyperconverged \ -o=jsonpath='{range .status.conditions[*]}{.type}{"\t"}{.status}{"\t"}{.message}{"\n"}{end}'$ oc get hco -n openshift-cnv kubevirt-hyperconverged \ -o=jsonpath='{range .status.conditions[*]}{.type}{"\t"}{.status}{"\t"}{.message}{"\n"}{end}'Copy to Clipboard Copied! Toggle word wrap Toggle overflow A successful upgrade results in the following output:
ReconcileComplete True Reconcile completed successfully Available True Reconcile completed successfully Progressing False Reconcile completed successfully Degraded False Reconcile completed successfully Upgradeable True Reconcile completed successfully
ReconcileComplete True Reconcile completed successfully Available True Reconcile completed successfully Progressing False Reconcile completed successfully Degraded False Reconcile completed successfully Upgradeable True Reconcile completed successfullyCopy to Clipboard Copied! Toggle word wrap Toggle overflow
Additional information
Chapter 5. Using the CLI tools 링크 복사링크가 클립보드에 복사되었습니다!
The two primary CLI tools used for managing resources in the cluster are:
-
The container-native virtualization
virtctlclient -
The OpenShift Container Platform
occlient
5.1. Prerequisites 링크 복사링크가 클립보드에 복사되었습니다!
-
You must install the
virtctlclient.
5.2. Virtctl client commands 링크 복사링크가 클립보드에 복사되었습니다!
The virtctl client is a command-line utility for managing container-native virtualization resources. The following table contains the virtctl commands used throughout the container-native virtualization documentation.
| Command | Description |
|---|---|
|
| Start a virtual machine. |
|
| Stop a virtual machine. |
|
| Restart a virtual machine. |
|
| Create a service that forwards a designated port of a virtual machine or virtual machine instance and expose the service on the specified port of the node. |
|
| Connect to a serial console of a virtual machine instance. |
|
| Open a VNC connection to a virtual machine instance. |
|
| Upload a virtual machine image to a PersistentVolumeClaim. |
5.3. OpenShift Container Platform client commands 링크 복사링크가 클립보드에 복사되었습니다!
The OpenShift Container Platform oc client is a command-line utility for managing OpenShift Container Platform resources. The following table contains the oc commands used throughout the container-native virtualization documentation.
| Command | Description |
|---|---|
|
|
Log in to the OpenShift Container Platform cluster as |
|
| Display a list of objects for the specified object type in the project. |
|
| Display details of the specific resource in the project. |
|
| Create a resource in the project from a filename or from stdin. |
|
| Edit a resource in the project. |
|
| Delete a resource in the project. |
For more comprehensive information on oc client commands, see the OpenShift Container Platform CLI tools documentation.
Chapter 6. Virtual machines 링크 복사링크가 클립보드에 복사되었습니다!
6.1. Creating virtual machines 링크 복사링크가 클립보드에 복사되었습니다!
Use one of these procedures to create a virtual machine:
- Running the virtual machine wizard
- Pasting a pre-configured YAML file with the virtual machine wizard
- Using the CLI
- Importing a VMware virtual machine or template with the virtual machine wizard
Do not create virtual machines in openshift-* namespaces. Instead, create a new namespace or use an existing namespace without the openshift prefix.
6.1.1. Running the virtual machine wizard to create a virtual machine 링크 복사링크가 클립보드에 복사되었습니다!
The web console features an interactive wizard that guides you through General, Networking, Storage, Advanced, and Review steps to simplify the process of creating virtual machines. All required fields are marked by a *. When the required fields are completed, you can review and create your virtual machine.
Network Interface Cards (NICs) and storage disks can be created and attached to virtual machines after they have been created.
Bootable Disk
If either URL or Container are selected as the Source in the General step, a rootdisk disk is created and attached to the virtual machine as the Bootable Disk. You can modify the rootdisk but you cannot remove it.
A Bootable Disk is not required for virtual machines provisioned from a PXE source if there are no disks attached to the virtual machine. If one or more disks are attached to the virtual machine, you must select one as the Bootable Disk.
Prerequisites
- When you create your virtual machine using the wizard, your virtual machine’s storage medium must support Read-Write-Many (RWM) PVCs.
Procedure
- Click Workloads → Virtual Machines from the side menu.
- Click Create Virtual Machine and select New with Wizard.
- Fill in all required fields in the General step. Selecting a Template automatically fills in these fields.
Click Next to progress to the Networking step. A
nic0NIC is attached by default.- (Optional) Click Add Network Interface to create additional NICs.
-
(Optional) You can remove any or all NICs by clicking the Options menu
and selecting Delete. A virtual machine does not need a NIC attached to be created. NICs can be created after the virtual machine has been created.
Click Next to progress to the Storage screen.
-
(Optional) Click Add Disk to create additional disks. These disks can be removed by clicking the Options menu
and selecting Delete.
-
(Optional) Click the Options menu
to edit the disk and save your changes.
-
(Optional) Click Add Disk to create additional disks. These disks can be removed by clicking the Options menu
- Click Review and Create. The Results screen displays the JSON configuration file for the virtual machine.
The virtual machine is listed in Workloads → Virtual Machines.
Refer to the virtual machine wizard fields section when running the web console wizard.
6.1.1.1. Virtual machine wizard fields 링크 복사링크가 클립보드에 복사되었습니다!
| Name | Parameter | Description |
|---|---|---|
| Template | Template from which to create the virtual machine. Selecting a template will automatically complete other fields. | |
| Source | PXE | Provision virtual machine from PXE menu. Requires a PXE-capable NIC in the cluster. |
| URL | Provision virtual machine from an image available from an HTTP or S3 endpoint. | |
| Container |
Provision virtual machine from a bootable operating system container located in a registry accessible from the cluster. Example: | |
| Disk | Provision virtual machine from a disk. | |
| Attach disk | Attach an existing disk that was previously cloned or created and made available in the PersistentVolumeClaims. When this option is selected, you must manually complete the Operating System, Flavor, and Workload Profile fields. | |
| Operating System | The primary operating system that is selected for the virtual machine. | |
| Flavor | small, medium, large, tiny, Custom | Presets that determine the amount of CPU and memory allocated to the virtual machine. The presets displayed for Flavor are determined by the operating system. |
| Workload Profile | High Performance | A virtual machine configuration that is optimized for high-performance workloads. |
| Server | A profile optimized to run server workloads. | |
| Desktop | A virtual machine configuration for use on a desktop. | |
| Name |
The name can contain lowercase letters ( | |
| Description | Optional description field. | |
| Start virtual machine on creation | Select to automatically start the virtual machine upon creation. |
6.1.1.2. Cloud-init fields 링크 복사링크가 클립보드에 복사되었습니다!
| Name | Description |
|---|---|
| Hostname | Sets a specific host name for the virtual machine. |
| Authenticated SSH Keys | The user’s public key that is copied to ~/.ssh/authorized_keys on the virtual machine. |
| Use custom script | Replaces other options with a field in which you paste a custom cloud-init script. |
6.1.1.3. Networking fields 링크 복사링크가 클립보드에 복사되었습니다!
| Name | Description |
|---|---|
| Name | Name for the Network Interface Card. |
| Model | Driver for the Network Interface Card or model for the Network Interface Card. |
| Network | List of available NetworkAttachmentDefinition objects. |
| Type |
List of available binding methods. For the default Pod network, |
| MAC Address | MAC address for the Network Interface Card. If a MAC address is not specified, an ephemeral address is generated for the session. |
6.1.1.4. Storage fields 링크 복사링크가 클립보드에 복사되었습니다!
| Name | Description |
|---|---|
| Source | Select a blank disk for the virtual machine or choose from the options available: PXE, Container, URL or Disk. To select an existing disk and attach it to the virtual machine, choose Attach Disk from a list of available PersistentVolumeClaims (PVCs) or from a cloned disk. |
| Name |
Name of the disk. The name can contain lowercase letters ( |
| Size (GiB) | Size, in GiB, of the disk. |
| Interface | Name of the interface. |
| Storage class |
Name of the underlying |
6.1.2. Pasting in a pre-configured YAML file to create a virtual machine 링크 복사링크가 클립보드에 복사되었습니다!
Create a virtual machine by writing or pasting a YAML configuration file in the web console in the Workloads → Virtual Machines screen. A valid example virtual machine configuration is provided by default whenever you open the YAML edit screen.
If your YAML configuration is invalid when you click Create, an error message indicates the parameter in which the error occurs. Only one error is shown at a time.
Navigating away from the YAML screen while editing cancels any changes to the configuration you have made.
Procedure
- Click Workloads → Virtual Machines from the side menu.
- Click Create Virtual Machine and select New from YAML.
Write or paste your virtual machine configuration in the editable window.
-
Alternatively, use the
examplevirtual machine provided by default in the YAML screen.
-
Alternatively, use the
- (Optional) Click Download to download the YAML configuration file in its present state.
- Click Create to create the virtual machine.
The virtual machine is listed in Workloads → Virtual Machines.
6.1.3. Using the CLI to create a virtual machine 링크 복사링크가 클립보드에 복사되었습니다!
Procedure
The spec object of the VirtualMachine configuration file references the virtual machine settings, such as the number of cores and the amount of memory, the disk type, and the volumes to use.
-
Attach the virtual machine disk to the virtual machine by referencing the relevant PVC
claimNameas a volume. To create a virtual machine with the OpenShift Container Platform client, run this command:
oc create -f <vm.yaml>
$ oc create -f <vm.yaml>Copy to Clipboard Copied! Toggle word wrap Toggle overflow - Since virtual machines are created in a Stopped state, run a virtual machine instance by starting it.
A ReplicaSet’s purpose is often used to guarantee the availability of a specified number of identical Pods. ReplicaSet is not currently supported in container-native virtualization.
| Setting | Description |
|---|---|
| Cores | The number of cores inside the virtual machine. Must be a value greater than or equal to 1. |
| Memory | The amount of RAM that is allocated to the virtual machine by the node. Specify a value in M for Megabyte or Gi for Gigabyte. |
| Disks: name | The name of the volume that is referenced. Must match the name of a volume. |
| Setting | Description |
|---|---|
| Name | The name of the volume, which must be a DNS label and unique within the virtual machine. |
| PersistentVolumeClaim |
The PVC to attach to the virtual machine. The |
6.1.4. Virtual machine storage volume types 링크 복사링크가 클립보드에 복사되었습니다!
Virtual machine storage volume types are listed, as well as domain and volume settings. See the kubevirt API Reference for a definitive list of virtual machine settings.
| ephemeral | A local copy-on-write (COW) image that uses a network volume as a read-only backing store. The backing volume must be a PersistentVolumeClaim. The ephemeral image is created when the virtual machine starts and stores all writes locally. The ephemeral image is discarded when the virtual machine is stopped, restarted, or deleted. The backing volume (PVC) is not mutated in any way. |
| persistentVolumeClaim | Attaches an available PV to a virtual machine. Attaching a PV allows for the virtual machine data to persist between sessions. Importing an existing virtual machine disk into a PVC by using CDI and attaching the PVC to a virtual machine instance is the recommended method for importing existing virtual machines into OpenShift Container Platform. There are some requirements for the disk to be used within a PVC. |
| dataVolume | DataVolumes build on the persistentVolumeClaim disk type by managing the process of preparing the virtual machine disk via an import, clone, or upload operation. VMs that use this volume type are guaranteed not to start until the volume is ready. |
| cloudInitNoCloud | Attaches a disk that contains the referenced cloud-init NoCloud data source, providing user data and metadata to the virtual machine. A cloud-init installation is required inside the virtual machine disk. |
| containerDisk | References an image, such as a virtual machine disk, that is stored in the container image registry. The image is pulled from the registry and embedded in a volume when the virtual machine is created. A containerDisk volume is ephemeral. It is discarded when the virtual machine is stopped, restarted, or deleted. Container disks are not limited to a single virtual machine and are useful for creating large numbers of virtual machine clones that do not require persistent storage. Only RAW and QCOW2 formats are supported disk types for the container image registry. QCOW2 is recommended for reduced image size. |
| emptyDisk | Creates an additional sparse QCOW2 disk that is tied to the life-cycle of the virtual machine interface. The data survives guest-initiated reboots in the virtual machine but is discarded when the virtual machine stops or is restarted from the web console. The empty disk is used to store application dependencies and data that otherwise exceeds the limited temporary file system of an ephemeral disk. The disk capacity size must also be provided. |
See the kubevirt API Reference for a definitive list of virtual machine settings.
6.2. Editing virtual machines 링크 복사링크가 클립보드에 복사되었습니다!
You can update a virtual machine configuration using either the YAML editor in the web console or the OpenShift client on the command line. You can also update a subset of the parameters in the Virtual Machine Overview of the web console.
6.2.1. Editing a virtual machine in the web console 링크 복사링크가 클립보드에 복사되었습니다!
Edit select values of a virtual machine in the Virtual Machine Overview screen of the web console by clicking on the pencil icon next to the relevant field. Other values can be edited using the CLI.
Procedure
- Click Workloads → Virtual Machines from the side menu.
- Select a virtual machine to open the Virtual Machine Overview screen.
- Click the pencil icon to make that field editable.
- Make the relevant changes and click Save.
If the virtual machine is running, changes will not take effect until you reboot the virtual machine.
6.2.2. Editing a virtual machine YAML configuration using the web console 링크 복사링크가 클립보드에 복사되었습니다!
Using the web console, edit the YAML configuration of a virtual machine.
Not all parameters can be updated. If you edit values that cannot be changed and click Save, an error message indicates the parameter that was not able to be updated.
The YAML configuration can be edited while the virtual machine is Running, however the changes will only take effect after the virtual machine has been stopped and started again.
Navigating away from the YAML screen while editing cancels any changes to the configuration you have made.
Procedure
- Click Workloads → Virtual Machine from the side menu.
- Select a virtual machine.
- Click the YAML tab to display the editable configuration.
- Optional: You can click Download to download the YAML file locally in its current state.
- Edit the file and click Save.
A confirmation message shows that the modification has been successful and includes the updated version number for the object.
6.2.3. Editing a virtual machine YAML configuration using the CLI 링크 복사링크가 클립보드에 복사되었습니다!
Use this procedure to edit a virtual machine YAML configuration using the CLI.
Prerequisites
- You configured a virtual machine with a YAML object configuration file.
-
You installed the
ocCLI.
Procedure
Run the following command to update the virtual machine configuration.
oc edit <object_type> <object_ID>
$ oc edit <object_type> <object_ID>Copy to Clipboard Copied! Toggle word wrap Toggle overflow - Open the object configuration.
- Edit the YAML.
If you edit a running virtual machine, you need to do one of the following:
- Restart the virtual machine
Run the following command for the new configuration to take effect.
oc apply <object_type> <object_ID>
$ oc apply <object_type> <object_ID>Copy to Clipboard Copied! Toggle word wrap Toggle overflow
6.2.4. Adding a virtual disk to a virtual machine 링크 복사링크가 클립보드에 복사되었습니다!
Use this procedure to add a virtual disk to a virtual machine.
Procedure
- From the Virtual Machines tab, select your virtual machine.
- Select the Disks tab.
- Click Add Disks to open the Add Disk window.
- In the Add Disk window, specify Source, Name, Size, Interface, and Storage Class.
- Use the drop-down lists and check boxes to edit the disk configuration.
- Click OK.
6.2.5. Adding a network interface to a virtual machine 링크 복사링크가 클립보드에 복사되었습니다!
Use this procedure to add a network interface to a virtual machine.
Procedure
- From the Virtual Machines tab, select the virtual machine.
- Select the Network Interfaces tab.
- Click Add Network Interface.
- In the Add Network Interface window, specify the Name, Model, Network, Type, and MAC Address of the network interface.
- Click Add to add the network interface.
- Restart the virtual machine to enable access.
- Edit the drop-down lists and check boxes to configure the network interface.
- Click Save Changes.
- Click OK.
The new network interface displays at the top of the Create Network Interface list until the user restarts it.
The new network interface has a Pending VM restart Link State until you restart the virtual machine. Hover over the Link State to display more detailed information.
The Link State is set to Up by default when the network interface card is defined on the virtual machine and connected to the network.
6.2.6. Editing CD-ROMs for Virtual Machines 링크 복사링크가 클립보드에 복사되었습니다!
Use the following procedure to configure CD-ROMs for virtual machines.
Procedure
- From the Virtual Machines tab, select your virtual machine.
- Select the Overview tab.
Click the pencil icon to the right of the CD-ROMs label to open the Edit CD-ROM window.
- If no CD-ROMs are available for editing, the CD label initially displays blank.
- If there are CD-ROMs available, you can eject the CD-ROM by clicking Eject CD-ROM or remove it by clicking -.
In the Edit CD-ROM window, do the following:
Select the type of CD-ROM configuration in the Media Type field.
- On Windows systems, Windows guest tools is attached by default in the Media Type field.
- Complete the required information for each Media Type.
- Click Add CD-ROM.
- When all CD-ROMs are added, click Save.
6.3. Deleting virtual machines 링크 복사링크가 클립보드에 복사되었습니다!
Use one of these procedures to delete a virtual machine:
- Using the web console
- Using the CLI
6.3.1. Deleting a virtual machine using the web console 링크 복사링크가 클립보드에 복사되었습니다!
Deleting a virtual machine permanently removes it from the cluster.
Delete a virtual machine using the ⋮ button of the virtual machine in the Workloads → Virtual Machines list, or using the Actions control of the Virtual Machine Details screen.
Procedure
- In the container-native virtualization console, click Workloads → Virtual Machines from the side menu.
Click the ⋮ button of the virtual machine to delete and select Delete Virtual Machine.
- Alternatively, click the virtual machine name to open the Virtual Machine Details screen and click Actions → Delete Virtual Machine.
- In the confirmation pop-up window, click Delete to permanently delete the virtual machine.
6.3.2. Deleting a virtual machine and its DataVolume using the CLI 링크 복사링크가 클립보드에 복사되었습니다!
When you delete a virtual machine, the DataVolume it uses is not automatically deleted.
Deleting the DataVolume is recommended in order to maintain a clean environment and avoid possible confusion.
Procedure
Run these commands to delete the virtual machine and the DataVolume.
You can delete objects only in the project you are currently working in, unless you specify the -n <project_name> option.
Run the following command to delete the virtual machine:
oc delete vm <fedora-vm>
$ oc delete vm <fedora-vm>Copy to Clipboard Copied! Toggle word wrap Toggle overflow Run the following command to delete the DataVolume:
oc delete dv <datavolume-name>
$ oc delete dv <datavolume-name>Copy to Clipboard Copied! Toggle word wrap Toggle overflow
6.4. Controlling virtual machines states 링크 복사링크가 클립보드에 복사되었습니다!
With container-native virtualization, you can stop, start, and restart virtual machines from both the web console and the command-line interface (CLI).
6.4.1. Controlling virtual machines from the web console 링크 복사링크가 클립보드에 복사되었습니다!
You can also stop, start, and restart virtual machines from the web console.
6.4.1.1. Starting a virtual machine 링크 복사링크가 클립보드에 복사되었습니다!
You can start a virtual machine from the web console.
Procedure
- In the container-native virtualization console, click Workloads → Virtual Machines.
Start the virtual machine from this screen, which makes it easier to perform actions on multiple virtual machines in the one screen, or from the Virtual Machine Details screen where you can view comprehensive details of the selected virtual machine:
-
Click the Options menu
at the end of virtual machine and select Start Virtual Machine.
- Click the virtual machine name to open the Virtual Machine Details screen and click Actions and select Start Virtual Machine.
-
Click the Options menu
- In the confirmation window, click Start to start the virtual machine.
When you start virtual machine that is provisioned from a URL source for the first time, the virtual machine is in the Importing state while container-native virtualization imports the container from the URL endpoint. Depending on the size of the image, this process might take several minutes.
6.4.1.2. Restarting a virtual machine 링크 복사링크가 클립보드에 복사되었습니다!
You can restart a running virtual machine from the web console.
Do not restart a virtual machine while it has a status of Importing. Restarting the virtual machine causes an error for it.
Procedure
- In the container-native virtualization console, click Workloads → Virtual Machines.
Restart the virtual machine from this screen, which makes it easier to perform actions on multiple virtual machines in the one screen, or from the Virtual Machine Details screen where you can view comprehensive details of the selected virtual machine:
-
Click the Options menu
at the end of virtual machine and select Restart Virtual Machine.
- Click the virtual machine name to open the Virtual Machine Details screen and click Actions and select Restart Virtual Machine.
-
Click the Options menu
- In the confirmation window, click Restart to restart the virtual machine.
6.4.1.3. Stopping a virtual machine 링크 복사링크가 클립보드에 복사되었습니다!
You can stop a virtual machine from the web console.
Procedure
- In the container-native virtualization console, click Workloads → Virtual Machines.
Stop the virtual machine from this screen, which makes it easier to perform actions on multiple virtual machines in the one screen, or from the Virtual Machine Details screen where you can view comprehensive details of the selected virtual machine:
-
Click the Options menu
at the end of virtual machine and select Stop Virtual Machine.
- Click the virtual machine name to open the Virtual Machine Details screen and click Actions and select Stop Virtual Machine.
-
Click the Options menu
- In the confirmation window, click Stop to stop the virtual machine.
6.4.2. CLI reference for controlling virtual machines 링크 복사링크가 클립보드에 복사되었습니다!
Use the following virtctl client utility and oc commands to change the state of the virtual machines and display lists of the virtual machines and the virtual machine instances that represent them.
When you run virtctl commands, you modify the virtual machines themselves, not the virtual machine instances that represent them in the web console.
6.4.2.1. start 링크 복사링크가 클립보드에 복사되었습니다!
Start a virtual machine.
Example: Start a virtual machine in the current project
virtctl start <example-vm>
$ virtctl start <example-vm>
Example: Start a virtual machine in a specific project
virtctl start <example-vm> -n <project_name>
$ virtctl start <example-vm> -n <project_name>
6.4.2.2. restart 링크 복사링크가 클립보드에 복사되었습니다!
Restart a running virtual machine.
Example: Restart a virtual machine in the current project
virtctl restart <example-vm>
$ virtctl restart <example-vm>
Example: Restart a virtual machine in a specific project
virtctl restart <example-vm> -n <project_name>
$ virtctl restart <example-vm> -n <project_name>
6.4.2.3. stop 링크 복사링크가 클립보드에 복사되었습니다!
Stop a running virtual machine.
Example: Stop a virtual machine in the current project
virtctl stop <example-vm>
$ virtctl stop <example-vm>
Example: Stop a virtual machine in a specific project
virtctl stop <example-vm> -n <project_name>
$ virtctl stop <example-vm> -n <project_name>
6.4.2.4. list 링크 복사링크가 클립보드에 복사되었습니다!
List the virtual machines or virtual machine instances in a project. The virtual machine instances are abstractions that represent the virtual machines themselves.
Example: List the virtual machines in the current project
oc get vm
$ oc get vm
Example: List the virtual machines in a specific project
oc get vm -n <project_name>
$ oc get vm -n <project_name>
Example: List the running virtual machine instances in the current project
oc get vmi
$ oc get vmi
Example: List the running virtual machine instances in a specific project
oc get vmi -n <project_name>
$ oc get vmi -n <project_name>
6.5. Accessing virtual machine consoles 링크 복사링크가 클립보드에 복사되었습니다!
Container-native virtualization provides different virtual machine consoles that you can use to accomplish different product tasks. You can access these consoles through the web console and by using CLI commands.
6.5.1. Virtual machine console sessions 링크 복사링크가 클립보드에 복사되었습니다!
You can connect to the VNC and serial consoles of a running virtual machine from the Consoles tab in the Virtual Machine Details screen of the web console.
There are two consoles available: the graphical VNC Console and the Serial Console. The VNC Console opens by default whenever you navigate to the Consoles tab. You can switch between the consoles using the VNC Console Serial Console list.
Console sessions remain active in the background unless they are disconnected. When the Disconnect before switching checkbox is active and you switch consoles, the current console session is disconnected and a new session with the selected console connects to the virtual machine. This ensures only one console session is open at a time.
Options for the VNC Console
The Send Key button lists key combinations to send to the virtual machine.
Options for the Serial Console
Use the Disconnect button to manually disconnect the Serial Console session from the virtual machine.
Use the Reconnect button to manually open a Serial Console session to the virtual machine.
6.5.2. Connecting to the virtual machine with the web console 링크 복사링크가 클립보드에 복사되었습니다!
6.5.2.1. Connecting to the terminal 링크 복사링크가 클립보드에 복사되었습니다!
You can connect to a virtual machine by using the web console.
Procedure
- Ensure you are in the correct project. If not, click the Project list and select the appropriate project.
- Click Workloads → Virtual Machines to display the virtual machines in the project.
- Select a virtual machine.
-
In the Overview tab, click the
virt-launcher-<vm-name>Pod. - Click the Terminal tab. If the terminal is blank, select the terminal and press any key to initiate connection.
6.5.2.2. Connecting to the serial console 링크 복사링크가 클립보드에 복사되었습니다!
Connect to the Serial Console of a running virtual machine from the Consoles tab in the Virtual Machine Details screen of the web console.
Procedure
- In the container-native virtualization console, click Workloads → Virtual Machines.
- Select a virtual machine.
- Click Consoles. The VNC console opens by default.
- Click the VNC Console drop-down list and select Serial Console.
6.5.2.3. Connecting to the VNC console 링크 복사링크가 클립보드에 복사되었습니다!
Connect to the VNC console of a running virtual machine from the Consoles tab in the Virtual Machine Details screen of the web console.
Procedure
- In the container-native virtualization console, click Workloads → Virtual Machines.
- Select a virtual machine.
- Click Consoles. The VNC console opens by default.
6.5.2.4. Connecting to the RDP console 링크 복사링크가 클립보드에 복사되었습니다!
The desktop viewer console, which utilizes the Remote Desktop Protocol (RDP), provides a better console experience for connecting to Windows virtual machines.
To connect to a Windows virtual machine with RDP, download the console.rdp file for the virtual machine from the Consoles tab in the Virtual Machine Details screen of the web console and supply it to your preferred RDP client.
Prerequisites
-
A running Windows virtual machine with the QEMU guest agent installed. The
qemu-guest-agentis included in the VirtIO drivers. - A layer-2 NIC attached to the virtual machine.
- An RDP client installed on a machine on the same network as the Windows virtual machine.
Procedure
- In the container-native virtualization console, click Workloads → Virtual Machines.
- Select a Windows virtual machine.
- Click the Consoles tab.
- Click the Consoles list and select Desktop Viewer.
- In the Network Interface list, select the layer-2 NIC.
-
Click Launch Remote Desktop to download the
console.rdpfile. Open an RDP client and reference the
console.rdpfile. For example, using remmina:remmina --connect /path/to/console.rdp
$ remmina --connect /path/to/console.rdpCopy to Clipboard Copied! Toggle word wrap Toggle overflow - Enter the Administrator user name and password to connect to the Windows virtual machine.
6.5.3. Accessing virtual machine consoles by using CLI commands 링크 복사링크가 클립보드에 복사되었습니다!
6.5.3.1. Accessing a virtual machine instance via SSH 링크 복사링크가 클립보드에 복사되었습니다!
You can use SSH to access a virtual machine after you expose port 22 on it.
The virtctl expose command forwards a virtual machine instance port to a node port and creates a service for enabled access. The following example creates the fedora-vm-ssh service that forwards port 22 of the <fedora-vm> virtual machine to a port on the node:
Prerequisites
-
The virtual machine instance you want to access must be connected to the default Pod network by using the
masqueradebinding method. - The virtual machine instance you want to access must be running.
-
Install the OpenShift CLI (
oc).
Procedure
Run the following command to create the
fedora-vm-sshservice:virtctl expose vm <fedora-vm> --port=20022 --target-port=22 --name=fedora-vm-ssh --type=NodePort
$ virtctl expose vm <fedora-vm> --port=20022 --target-port=22 --name=fedora-vm-ssh --type=NodePort1 Copy to Clipboard Copied! Toggle word wrap Toggle overflow - 1
<fedora-vm>is the name of the virtual machine that you run thefedora-vm-sshservice on.
Check the service to find out which port the service acquired:
oc get svc
$ oc get svc NAME TYPE CLUSTER-IP EXTERNAL-IP PORT(S) AGE fedora-vm-ssh NodePort 127.0.0.1 <none> 20022:32551/TCP 6sCopy to Clipboard Copied! Toggle word wrap Toggle overflow In this example, the service acquired the
32551port.Log in to the virtual machine instance via SSH. Use the
ipAddressof the node and the port that you found in the previous step:ssh username@<node_IP_address> -p 32551
$ ssh username@<node_IP_address> -p 32551Copy to Clipboard Copied! Toggle word wrap Toggle overflow
6.5.3.2. Accessing the serial console of a virtual machine instance 링크 복사링크가 클립보드에 복사되었습니다!
The virtctl console command opens a serial console to the specified virtual machine instance.
Prerequisites
-
The
virt-viewerpackage must be installed. - The virtual machine instance you want to access must be running.
Procedure
Connect to the serial console with
virtctl:virtctl console <VMI>
$ virtctl console <VMI>Copy to Clipboard Copied! Toggle word wrap Toggle overflow
6.5.3.3. Accessing the graphical console of a virtual machine instances with VNC 링크 복사링크가 클립보드에 복사되었습니다!
The virtctl client utility can use the remote-viewer function to open a graphical console to a running virtual machine instance. This capability is included in the virt-viewer package.
Prerequisites
-
The
virt-viewerpackage must be installed. - The virtual machine instance you want to access must be running.
If you use virtctl via SSH on a remote machine, you must forward the X session to your machine.
Procedure
Connect to the graphical interface with the
virtctlutility:virtctl vnc <VMI>
$ virtctl vnc <VMI>Copy to Clipboard Copied! Toggle word wrap Toggle overflow If the command failed, try using the
-vflag to collect troubleshooting information:virtctl vnc <VMI> -v 4
$ virtctl vnc <VMI> -v 4Copy to Clipboard Copied! Toggle word wrap Toggle overflow
6.5.3.4. Connecting to a Windows virtual machine with an RDP console 링크 복사링크가 클립보드에 복사되었습니다!
The Remote Desktop Protocol (RDP) provides a better console experience for connecting to Windows virtual machines.
To connect to a Windows virtual machine with RDP, specify the IP address of the attached L2 NIC to your RDP client.
Prerequisites
-
A running Windows virtual machine with the QEMU guest agent installed. The
qemu-guest-agentis included in the VirtIO drivers. - A layer 2 NIC attached to the virtual machine.
- An RDP client installed on a machine on the same network as the Windows virtual machine.
Procedure
Log in to the container-native virtualization cluster through the
ocCLI tool as a user with an access token.oc login -u <user> https://<cluster.example.com>:8443
$ oc login -u <user> https://<cluster.example.com>:8443Copy to Clipboard Copied! Toggle word wrap Toggle overflow Use
oc describe vmito display the configuration of the running Windows virtual machine.oc describe vmi <windows-vmi-name>
$ oc describe vmi <windows-vmi-name>Copy to Clipboard Copied! Toggle word wrap Toggle overflow Copy to Clipboard Copied! Toggle word wrap Toggle overflow -
Identify and copy the IP address of the layer 2 network interface. This is
192.0.2.0in the above example, or2001:db8::if you prefer IPv6. - Open an RDP client and use the IP address copied in the previous step for the connection.
- Enter the Administrator user name and password to connect to the Windows virtual machine.
6.6. Installing VirtIO driver on an existing Windows virtual machine 링크 복사링크가 클립보드에 복사되었습니다!
6.6.1. Understanding VirtIO drivers 링크 복사링크가 클립보드에 복사되었습니다!
VirtIO drivers are paravirtualized device drivers required for Microsoft Windows virtual machines to run in container-native virtualization. The supported drivers are available in the container-native-virtualization/virtio-win container disk of the Red Hat Container Catalog.
The container-native-virtualization/virtio-win container disk must be attached to the virtual machine as a SATA CD drive to enable driver installation. You can install VirtIO drivers during Windows installation on the virtual machine or added to an existing Windows installation.
After the drivers are installed, the container-native-virtualization/virtio-win container disk can be removed from the virtual machine.
See also: Installing Virtio drivers on a new Windows virtual machine.
6.6.2. Supported VirtIO drivers for Microsoft Windows virtual machines 링크 복사링크가 클립보드에 복사되었습니다!
| Driver name | Hardware ID | Description |
|---|---|---|
| viostor |
VEN_1AF4&DEV_1001 | The block driver. Sometimes displays as an SCSI Controller in the Other devices group. |
| viorng |
VEN_1AF4&DEV_1005 | The entropy source driver. Sometimes displays as a PCI Device in the Other devices group. |
| NetKVM |
VEN_1AF4&DEV_1000 | The network driver. Sometimes displays as an Ethernet Controller in the Other devices group. Available only if a VirtIO NIC is configured. |
6.6.3. Adding VirtIO drivers container disk to a virtual machine 링크 복사링크가 클립보드에 복사되었습니다!
Container-native virtualization distributes VirtIO drivers for Microsoft Windows as a container disk, which is available from the Red Hat Container Catalog. To install these drivers to a Windows virtual machine, attach the container-native-virtualization/virtio-win container disk to the virtual machine as a SATA CD drive in the virtual machine configuration file.
Prerequisites
-
Download the
container-native-virtualization/virtio-wincontainer disk from the Red Hat Container Catalog. This is not mandatory, because the container disk will be downloaded from the Red Hat registry if it not already present in the cluster, but it can reduce installation time.
Procedure
Add the
container-native-virtualization/virtio-wincontainer disk as acdromdisk in the Windows virtual machine configuration file. The container disk will be downloaded from the registry if it is not already present in the cluster.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - 1
- Container-native virtualization boots virtual machine disks in the order defined in the
VirtualMachineconfiguration file. You can either define other disks for the virtual machine before thecontainer-native-virtualization/virtio-wincontainer disk or use the optionalbootOrderparameter to ensure the virtual machine boots from the correct disk. If you specify thebootOrderfor a disk, it must be specified for all disks in the configuration.
The disk is available once the virtual machine has started:
-
If you add the container disk to a running virtual machine, use
oc apply -f <vm.yaml>in the CLI or reboot the virtual machine for the changes to take effect. -
If the virtual machine is not running, use
virtctl start <vm>.
-
If you add the container disk to a running virtual machine, use
After the virtual machine has started, the VirtIO drivers can be installed from the attached SATA CD drive.
6.6.4. Installing VirtIO drivers on an existing Windows virtual machine 링크 복사링크가 클립보드에 복사되었습니다!
Install the VirtIO drivers from the attached SATA CD drive to an existing Windows virtual machine.
This procedure uses a generic approach to adding drivers to Windows. The process might differ slightly between versions of Windows. Refer to the installation documentation for your version of Windows for specific installation steps.
Procedure
- Start the virtual machine and connect to a graphical console.
- Log in to a Windows user session.
Open Device Manager and expand Other devices to list any Unknown device.
-
Open the
Device Propertiesto identify the unknown device. Right-click the device and select Properties. - Click the Details tab and select Hardware Ids in the Property list.
- Compare the Value for the Hardware Ids with the supported VirtIO drivers.
-
Open the
- Right-click the device and select Update Driver Software.
- Click Browse my computer for driver software and browse to the attached SATA CD drive, where the VirtIO drivers are located. The drivers are arranged hierarchically according to their driver type, operating system, and CPU architecture.
- Click Next to install the driver.
- Repeat this process for all the necessary VirtIO drivers.
- After the driver installs, click Close to close the window.
- Reboot the virtual machine to complete the driver installation.
6.6.5. Removing the VirtIO container disk from a virtual machine 링크 복사링크가 클립보드에 복사되었습니다!
After installing all required VirtIO drivers to the virtual machine, the container-native-virtualization/virtio-win container disk no longer needs to be attached to the virtual machine. Remove the container-native-virtualization/virtio-win container disk from the virtual machine configuration file.
Procedure
Edit the configuration file and remove the
diskand thevolume.oc edit vm <vm-name>
$ oc edit vm <vm-name>Copy to Clipboard Copied! Toggle word wrap Toggle overflow Copy to Clipboard Copied! Toggle word wrap Toggle overflow - Reboot the virtual machine for the changes to take effect.
6.7. Installing VirtIO driver on a new Windows virtual machine 링크 복사링크가 클립보드에 복사되었습니다!
6.7.1. Prerequisites 링크 복사링크가 클립보드에 복사되었습니다!
- Windows installation media accessible by the virtual machine, such as importing an ISO into a data volume and attaching it to the virtual machine.
6.7.2. Understanding VirtIO drivers 링크 복사링크가 클립보드에 복사되었습니다!
VirtIO drivers are paravirtualized device drivers required for Microsoft Windows virtual machines to run in container-native virtualization. The supported drivers are available in the container-native-virtualization/virtio-win container disk of the Red Hat Container Catalog.
The container-native-virtualization/virtio-win container disk must be attached to the virtual machine as a SATA CD drive to enable driver installation. You can install VirtIO drivers during Windows installation on the virtual machine or added to an existing Windows installation.
After the drivers are installed, the container-native-virtualization/virtio-win container disk can be removed from the virtual machine.
See also: Installing VirtIO driver on an existing Windows virtual machine.
6.7.3. Supported VirtIO drivers for Microsoft Windows virtual machines 링크 복사링크가 클립보드에 복사되었습니다!
| Driver name | Hardware ID | Description |
|---|---|---|
| viostor |
VEN_1AF4&DEV_1001 | The block driver. Sometimes displays as an SCSI Controller in the Other devices group. |
| viorng |
VEN_1AF4&DEV_1005 | The entropy source driver. Sometimes displays as a PCI Device in the Other devices group. |
| NetKVM |
VEN_1AF4&DEV_1000 | The network driver. Sometimes displays as an Ethernet Controller in the Other devices group. Available only if a VirtIO NIC is configured. |
6.7.4. Adding VirtIO drivers container disk to a virtual machine 링크 복사링크가 클립보드에 복사되었습니다!
Container-native virtualization distributes VirtIO drivers for Microsoft Windows as a container disk, which is available from the Red Hat Container Catalog. To install these drivers to a Windows virtual machine, attach the container-native-virtualization/virtio-win container disk to the virtual machine as a SATA CD drive in the virtual machine configuration file.
Prerequisites
-
Download the
container-native-virtualization/virtio-wincontainer disk from the Red Hat Container Catalog. This is not mandatory, because the container disk will be downloaded from the Red Hat registry if it not already present in the cluster, but it can reduce installation time.
Procedure
Add the
container-native-virtualization/virtio-wincontainer disk as acdromdisk in the Windows virtual machine configuration file. The container disk will be downloaded from the registry if it is not already present in the cluster.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - 1
- Container-native virtualization boots virtual machine disks in the order defined in the
VirtualMachineconfiguration file. You can either define other disks for the virtual machine before thecontainer-native-virtualization/virtio-wincontainer disk or use the optionalbootOrderparameter to ensure the virtual machine boots from the correct disk. If you specify thebootOrderfor a disk, it must be specified for all disks in the configuration.
The disk is available once the virtual machine has started:
-
If you add the container disk to a running virtual machine, use
oc apply -f <vm.yaml>in the CLI or reboot the virtual machine for the changes to take effect. -
If the virtual machine is not running, use
virtctl start <vm>.
-
If you add the container disk to a running virtual machine, use
After the virtual machine has started, the VirtIO drivers can be installed from the attached SATA CD drive.
6.7.5. Installing VirtIO drivers during Windows installation 링크 복사링크가 클립보드에 복사되었습니다!
Install the VirtIO drivers from the attached SATA CD driver during Windows installation.
This procedure uses a generic approach to the Windows installation and the installation method might differ between versions of Windows. Refer to the documentation for the version of Windows that you are installing.
Procedure
- Start the virtual machine and connect to a graphical console.
- Begin the Windows installation process.
- Select the Advanced installation.
-
The storage destination will not be recognized until the driver is loaded. Click
Load driver. - The drivers are attached as a SATA CD drive. Click OK and browse the CD drive for the storage driver to load. The drivers are arranged hierarchically according to their driver type, operating system, and CPU architecture.
- Repeat the previous two steps for all required drivers.
- Complete the Windows installation.
6.7.6. Removing the VirtIO container disk from a virtual machine 링크 복사링크가 클립보드에 복사되었습니다!
After installing all required VirtIO drivers to the virtual machine, the container-native-virtualization/virtio-win container disk no longer needs to be attached to the virtual machine. Remove the container-native-virtualization/virtio-win container disk from the virtual machine configuration file.
Procedure
Edit the configuration file and remove the
diskand thevolume.oc edit vm <vm-name>
$ oc edit vm <vm-name>Copy to Clipboard Copied! Toggle word wrap Toggle overflow Copy to Clipboard Copied! Toggle word wrap Toggle overflow - Reboot the virtual machine for the changes to take effect.
6.8. Advanced virtual machine management 링크 복사링크가 클립보드에 복사되었습니다!
6.8.1. Automating management tasks 링크 복사링크가 클립보드에 복사되었습니다!
You can automate container-native virtualization management tasks by using Red Hat Ansible Automation Platform. Learn the basics by using an Ansible Playbook to create a new virtual machine.
6.8.1.1. About Red Hat Ansible Automation 링크 복사링크가 클립보드에 복사되었습니다!
Ansible is an automation tool used to configure systems, deploy software, and perform rolling updates. Ansible includes support for container-native virtualization, and Ansible modules enable you to automate cluster management tasks such as template, persistent volume claim, and virtual machine operations.
Ansible provides a way to automate container-native virtualization management, which you can also accomplish by using the oc CLI tool or APIs. Ansible is unique because it allows you to integrate KubeVirt modules with other Ansible modules.
6.8.1.2. Automating virtual machine creation 링크 복사링크가 클립보드에 복사되었습니다!
You can use the kubevirt_vm Ansible Playbook to create virtual machines in your OpenShift Container Platform cluster using Red Hat Ansible Automation Platform.
Prerequisites
- Red Hat Ansible Engine version 2.8 or newer
Procedure
Edit an Ansible Playbook YAML file so that it includes the
kubevirt_vmtask:Copy to Clipboard Copied! Toggle word wrap Toggle overflow NoteThis snippet only includes the
kubevirt_vmportion of the playbook.Edit the values to reflect the virtual machine you want to create, including the
namespace, the number ofcpu_cores, thememory, and thedisks. For example:Copy to Clipboard Copied! Toggle word wrap Toggle overflow If you want the virtual machine to boot immediately after creation, add
state: runningto the YAML file. For example:kubevirt_vm: namespace: default name: vm1 state: running cpu_cores: 1kubevirt_vm: namespace: default name: vm1 state: running1 cpu_cores: 1Copy to Clipboard Copied! Toggle word wrap Toggle overflow - 1
- Changing this value to
state: absentdeletes the virtual machine, if it already exists.
Run the
ansible-playbookcommand, using your playbook’s file name as the only argument:ansible-playbook create-vm.yaml
$ ansible-playbook create-vm.yamlCopy to Clipboard Copied! Toggle word wrap Toggle overflow Review the output to determine if the play was successful:
Copy to Clipboard Copied! Toggle word wrap Toggle overflow If you did not include
state: runningin your playbook file and you want to boot the VM now, edit the file so that it includesstate: runningand run the playbook again:ansible-playbook create-vm.yaml
$ ansible-playbook create-vm.yamlCopy to Clipboard Copied! Toggle word wrap Toggle overflow
To verify that the virtual machine was created, try to access the VM console.
6.8.1.3. Example: Ansible Playbook for creating virtual machines 링크 복사링크가 클립보드에 복사되었습니다!
You can use the kubevirt_vm Ansible Playbook to automate virtual machine creation.
The following YAML file is an example of the kubevirt_vm playbook. It includes sample values that you must replace with your own information if you run the playbook.
Additional information
6.8.2. Configuring PXE booting for virtual machines 링크 복사링크가 클립보드에 복사되었습니다!
PXE booting, or network booting, is available in container-native virtualization. Network booting allows a computer to boot and load an operating system or other program without requiring a locally attached storage device. For example, you can use it to choose your desired OS image from a PXE server when deploying a new host.
6.8.2.1. Prerequisites 링크 복사링크가 클립보드에 복사되었습니다!
- A Linux bridge must be connected.
- The PXE server must be connected to the same VLAN as the bridge.
6.8.2.2. Container-native virtualization networking glossary 링크 복사링크가 클립보드에 복사되었습니다!
Container-native virtualization provides advanced networking functionality by using custom resources and plug-ins.
The following terms are used throughout container-native virtualization documentation:
- Container Network Interface (CNI)
- a Cloud Native Computing Foundation project, focused on container network connectivity. Container-native virtualization uses CNI plug-ins to build upon the basic Kubernetes networking functionality.
- Multus
- a "meta" CNI plug-in that allows multiple CNIs to exist so that a Pod or virtual machine can use the interfaces it needs.
- Custom Resource Definition (CRD)
- a Kubernetes API resource that allows you to define custom resources, or an object defined by using the CRD API resource.
- NetworkAttachmentDefinition
- a CRD introduced by the Multus project that allows you to attach Pods, virtual machines, and virtual machine instances to one or more networks.
- Preboot eXecution Environment (PXE)
- an interface that enables an administrator to boot a client machine from a server over the network. Network booting allows you to remotely load operating systems and other software onto the client.
6.8.2.3. PXE booting with a specified MAC address 링크 복사링크가 클립보드에 복사되었습니다!
As an administrator, you can boot a client over the network by first creating a NetworkAttachmentDefinition object for your PXE network. Then, reference the NetworkAttachmentDefinition in your virtual machine instance configuration file before you start the virtual machine instance. You can also specify a MAC address in the virtual machine instance configuration file, if required by the PXE server.
Prerequisites
- A Linux bridge must be connected.
- The PXE server must be connected to the same VLAN as the bridge.
Procedure
Configure a PXE network on the cluster:
Create the NetworkAttachmentDefinition file for PXE network
pxe-net-conf:Copy to Clipboard Copied! Toggle word wrap Toggle overflow - 1
- The
cnv-tuningplug-in provides support for custom MAC addresses.
NoteThe virtual machine instance will be attached to the bridge
br1through an access port with the requested VLAN.
Create the NetworkAttachmentDefinition object by using the file you created in the previous step:
oc create -f pxe-net-conf.yaml
$ oc create -f pxe-net-conf.yamlCopy to Clipboard Copied! Toggle word wrap Toggle overflow Edit the virtual machine instance configuration file to include the details of the interface and network.
Specify the network and MAC address, if required by the PXE server. If the MAC address is not specified, a value is assigned automatically. However, note that at this time, MAC addresses assigned automatically are not persistent.
Ensure that
bootOrderis set to1so that the interface boots first. In this example, the interface is connected to a network called<pxe-net>:Copy to Clipboard Copied! Toggle word wrap Toggle overflow NoteBoot order is global for interfaces and disks.
Assign a boot device number to the disk to ensure proper booting after operating system provisioning.
Set the disk
bootOrdervalue to2:Copy to Clipboard Copied! Toggle word wrap Toggle overflow Specify that the network is connected to the previously created NetworkAttachmentDefinition. In this scenario,
<pxe-net>is connected to the NetworkAttachmentDefinition called<pxe-net-conf>:Copy to Clipboard Copied! Toggle word wrap Toggle overflow
Create the virtual machine instance:
oc create -f vmi-pxe-boot.yaml
$ oc create -f vmi-pxe-boot.yaml virtualmachineinstance.kubevirt.io "vmi-pxe-boot" createdCopy to Clipboard Copied! Toggle word wrap Toggle overflow Wait for the virtual machine instance to run:
oc get vmi vmi-pxe-boot -o yaml | grep -i phase
$ oc get vmi vmi-pxe-boot -o yaml | grep -i phase phase: RunningCopy to Clipboard Copied! Toggle word wrap Toggle overflow View the virtual machine instance using VNC:
virtctl vnc vmi-pxe-boot
$ virtctl vnc vmi-pxe-bootCopy to Clipboard Copied! Toggle word wrap Toggle overflow - Watch the boot screen to verify that the PXE boot is successful.
Log in to the virtual machine instance:
virtctl console vmi-pxe-boot
$ virtctl console vmi-pxe-bootCopy to Clipboard Copied! Toggle word wrap Toggle overflow Verify the interfaces and MAC address on the virtual machine and that the interface connected to the bridge has the specified MAC address. In this case, we used
eth1for the PXE boot, without an IP address. The other interface,eth0, got an IP address from OpenShift Container Platform.ip addr
$ ip addr ... 3. eth1: <BROADCAST,MULTICAST> mtu 1500 qdisc noop state DOWN group default qlen 1000 link/ether de:00:00:00:00:de brd ff:ff:ff:ff:ff:ffCopy to Clipboard Copied! Toggle word wrap Toggle overflow
6.8.2.4. Template: virtual machine instance configuration file for PXE booting 링크 복사링크가 클립보드에 복사되었습니다!
6.8.3. Managing guest memory 링크 복사링크가 클립보드에 복사되었습니다!
If you want to adjust guest memory settings to suit a specific use case, you can do so by editing the guest’s YAML configuration file. Container-native virtualization allows you to configure guest memory overcommitment and disable guest memory overhead accounting.
Both of these procedures carry some degree of risk. Proceed only if you are an experienced administrator.
6.8.3.1. Configuring guest memory overcommitment 링크 복사링크가 클립보드에 복사되었습니다!
If your virtual workload requires more memory than available, you can use memory overcommitment to allocate all or most of the host’s memory to your virtual machine instances. Enabling memory overcommitment means you can maximize resources that are normally reserved for the host.
For example, if the host has 32 GB RAM, you can use memory overcommitment to fit 8 virtual machines with 4 GB RAM each. This allocation works under the assumption that the virtual machines will not use all of their memory at the same time.
Procedure
To explicitly tell the virtual machine instance that it has more memory available than was requested from the cluster, edit the virtual machine configuration file and set
spec.domain.memory.guestto a higher value thanspec.domain.resources.requests.memory. This process is called memory overcommitment.In this example,
1024Mis requested from the cluster, but the virtual machine instance is told that it has2048Mavailable. As long as there is enough free memory available on the node, the virtual machine instance will consume up to 2048M.Copy to Clipboard Copied! Toggle word wrap Toggle overflow NoteThe same eviction rules as those for Pods apply to the virtual machine instance if the node is under memory pressure.
Create the virtual machine:
oc create -f <file name>.yaml
$ oc create -f <file name>.yamlCopy to Clipboard Copied! Toggle word wrap Toggle overflow
6.8.3.2. Disabling guest memory overhead accounting 링크 복사링크가 클립보드에 복사되었습니다!
This procedure is only useful in certain use-cases and must only be attempted by advanced users.
A small amount of memory is requested by each virtual machine instance in addition to the amount that you request. This additional memory is used for the infrastructure that wraps each VirtualMachineInstance process.
Though it is not usually advisable, it is possible to increase the virtual machine instance density on the node by disabling guest memory overhead accounting.
Procedure
To disable guest memory overhead accounting, edit the YAML configuration file and set the
overcommitGuestOverheadvalue totrue. This parameter is disabled by default.Copy to Clipboard Copied! Toggle word wrap Toggle overflow NoteIf
overcommitGuestOverheadis enabled, it adds the guest overhead to memory limits, if present.Create the virtual machine:
oc create -f <file name>.yaml
$ oc create -f <file name>.yamlCopy to Clipboard Copied! Toggle word wrap Toggle overflow
6.9. Importing virtual machines 링크 복사링크가 클립보드에 복사되었습니다!
6.9.1. TLS certificates for DataVolume imports 링크 복사링크가 클립보드에 복사되었습니다!
6.9.1.1. Adding TLS certificates for authenticating DataVolume imports 링크 복사링크가 클립보드에 복사되었습니다!
TLS certificates for registry or HTTPS endpoints must be added to a ConfigMap in order to import data from these sources. This ConfigMap must be present in the namespace of the destination DataVolume.
Create the ConfigMap by referencing the relative file path for the TLS certificate.
Procedure
Ensure you are in the correct namespace. The ConfigMap can only be referenced by DataVolumes if it is in the same namespace.
oc get ns
$ oc get nsCopy to Clipboard Copied! Toggle word wrap Toggle overflow Create the ConfigMap:
oc create configmap <configmap-name> --from-file=</path/to/file/ca.pem>
$ oc create configmap <configmap-name> --from-file=</path/to/file/ca.pem>Copy to Clipboard Copied! Toggle word wrap Toggle overflow
6.9.1.2. Example: ConfigMap created from a TLS certificate 링크 복사링크가 클립보드에 복사되었습니다!
The following example is of a ConfigMap created from ca.pem TLS certificate.
6.9.2. Importing virtual machine images with DataVolumes 링크 복사링크가 클립보드에 복사되었습니다!
You can import an existing virtual machine image into your OpenShift Container Platform cluster. Container-native virtualization uses DataVolumes to automate the import of data and the creation of an underlying PersistentVolumeClaim (PVC).
When you import a disk image into a PVC, the disk image is expanded to use the full storage capacity that is requested in the PVC. To use this space, the disk partitions and file system(s) in the virtual machine might need to be expanded.
The resizing procedure varies based on the operating system installed on the VM. Refer to the operating system documentation for details.
6.9.2.1. Prerequisites 링크 복사링크가 클립보드에 복사되었습니다!
- If the endpoint requires a TLS certificate, the certificate must be included in a ConfigMap in the same namespace as the DataVolume and referenced in the DataVolume configuration.
- You may need to define a StorageClass or prepare CDI scratch space for this operation to complete successfully.
6.9.2.2. CDI supported operations matrix 링크 복사링크가 클립보드에 복사되었습니다!
This matrix shows the supported CDI operations for content types against endpoints, and which of these operations requires scratch space.
| Content types | HTTP | HTTPS | HTTP basic auth | Registry | Upload |
|---|---|---|---|---|---|
| KubeVirt(QCOW2) |
✓ QCOW2 |
✓ QCOW2** |
✓ QCOW2 |
✓ QCOW2* |
✓ QCOW2* |
| KubeVirt (RAW) |
✓ RAW |
✓ RAW |
✓ RAW |
✓ RAW* |
✓ RAW* |
| Archive+ | ✓ TAR | ✓ TAR | ✓ TAR | □ TAR | □ TAR |
✓ Supported operation
□ Unsupported operation
* Requires scratch space
** Requires scratch space if a custom certificate authority is required
+ Archive does not support block mode DVs
6.9.2.3. About DataVolumes 링크 복사링크가 클립보드에 복사되었습니다!
DataVolume objects are custom resources that are provided by the Containerized Data Importer (CDI) project. DataVolumes orchestrate import, clone, and upload operations that are associated with an underlying PersistentVolumeClaim (PVC). DataVolumes are integrated with KubeVirt, and they prevent a virtual machine from being started before the PVC has been prepared.
6.9.2.4. Importing a virtual machine image into an object with DataVolumes 링크 복사링크가 클립보드에 복사되었습니다!
To create a virtual machine from an imported image, specify the image location in the VirtualMachine configuration file before you create the virtual machine.
Prerequisites
-
Install the OpenShift CLI (
oc). -
A virtual machine disk image, in RAW, ISO, or QCOW2 format, optionally compressed by using
xzorgz -
An
HTTPendpoint where the image is hosted, along with any authentication credentials needed to access the data source - At least one available PersistentVolume
Procedure
Identify an
HTTPfile server that hosts the virtual disk image that you want to import. You need the complete URL in the correct format:If your data source requires authentication credentials, edit the
endpoint-secret.yamlfile, and apply the updated configuration to the cluster:Copy to Clipboard Copied! Toggle word wrap Toggle overflow oc apply -f endpoint-secret.yaml
$ oc apply -f endpoint-secret.yamlCopy to Clipboard Copied! Toggle word wrap Toggle overflow Edit the virtual machine configuration file, specifying the data source for the image you want to import. In this example, a Fedora image is imported:
Copy to Clipboard Copied! Toggle word wrap Toggle overflow - 1
- The
HTTPsource of the image you want to import. - 2
- The
secretRefparameter is optional. - 3
- The
certConfigMapis required for communicating with servers that use self-signed certificates or certificates not signed by the system CA bundle. The referenced ConfigMap must be in the same namespace as the DataVolume.
Create the virtual machine:
oc create -f vm-<name>-datavolume.yaml
$ oc create -f vm-<name>-datavolume.yamlCopy to Clipboard Copied! Toggle word wrap Toggle overflow NoteThe
oc createcommand creates the DataVolume and the virtual machine. The CDI controller creates an underlying PVC with the correct annotation, and the import process begins. When the import completes, the DataVolume status changes toSucceeded, and the virtual machine is allowed to start.DataVolume provisioning happens in the background, so there is no need to monitor it. You can start the virtual machine, and it will not run until the import is complete.
Optional verification steps
-
Run
oc get podsand look for the importer Pod. This Pod downloads the image from the specified URL and stores it on the provisioned PV. Monitor the DataVolume status until it shows
Succeeded.oc describe dv <data-label>
$ oc describe dv <data-label>1 Copy to Clipboard Copied! Toggle word wrap Toggle overflow - 1
- The data label for the DataVolume specified in the virtual machine configuration file.
To verify that provisioning is complete and that the VMI has started, try accessing its serial console:
virtctl console <vm-fedora-datavolume>
$ virtctl console <vm-fedora-datavolume>Copy to Clipboard Copied! Toggle word wrap Toggle overflow
6.9.2.5. Template: DataVolume virtual machine configuration file 링크 복사링크가 클립보드에 복사되었습니다!
example-dv-vm.yaml
- 1
- The
HTTPsource of the image you want to import, if applicable.
6.9.2.6. Template: DataVolume import configuration file 링크 복사링크가 클립보드에 복사되었습니다!
example-import-dv.yaml
6.9.3. Importing virtual machine images to block storage with DataVolumes 링크 복사링크가 클립보드에 복사되었습니다!
You can import an existing virtual machine image into your OpenShift Container Platform cluster. Container-native virtualization uses DataVolumes to automate the import of data and the creation of an underlying PersistentVolumeClaim (PVC).
When you import a disk image into a PVC, the disk image is expanded to use the full storage capacity that is requested in the PVC. To use this space, the disk partitions and file system(s) in the virtual machine might need to be expanded.
The resizing procedure varies based on the operating system that is installed on the virtual machine. Refer to the operating system documentation for details.
6.9.3.1. Prerequisites 링크 복사링크가 클립보드에 복사되었습니다!
- If you require scratch space according to the CDI supported operations matrix, you must first define a StorageClass or prepare CDI scratch space for this operation to complete successfully.
6.9.3.2. About DataVolumes 링크 복사링크가 클립보드에 복사되었습니다!
DataVolume objects are custom resources that are provided by the Containerized Data Importer (CDI) project. DataVolumes orchestrate import, clone, and upload operations that are associated with an underlying PersistentVolumeClaim (PVC). DataVolumes are integrated with KubeVirt, and they prevent a virtual machine from being started before the PVC has been prepared.
6.9.3.3. About block PersistentVolumes 링크 복사링크가 클립보드에 복사되었습니다!
A block PersistentVolume (PV) is a PV that is backed by a raw block device. These volumes do not have a filesystem and can provide performance benefits for virtual machines that either write to the disk directly or implement their own storage service.
Raw block volumes are provisioned by specifying volumeMode: Block in the PV and PersistentVolumeClaim (PVC) specification.
6.9.3.4. Creating a local block PersistentVolume 링크 복사링크가 클립보드에 복사되었습니다!
Create a local block PersistentVolume (PV) on a node by populating a file and mounting it as a loop device. You can then reference this loop device in a PV configuration as a Block volume and use it as a block device for a virtual machine image.
Procedure
-
Log in as
rootto the node on which to create the local PV. This procedure usesnode01for its examples. Create a file and populate it with null characters so that it can be used as a block device. The following example creates a file
loop10with a size of 2Gb (20 100Mb blocks):dd if=/dev/zero of=<loop10> bs=100M count=20
$ dd if=/dev/zero of=<loop10> bs=100M count=20Copy to Clipboard Copied! Toggle word wrap Toggle overflow Mount the
loop10file as a loop device.losetup </dev/loop10>d3 <loop10>
$ losetup </dev/loop10>d3 <loop10>1 2 Copy to Clipboard Copied! Toggle word wrap Toggle overflow Create a
PersistentVolumeconfiguration that references the mounted loop device.Copy to Clipboard Copied! Toggle word wrap Toggle overflow Create the block PV.
oc create -f <local-block-pv10.yaml>
# oc create -f <local-block-pv10.yaml>1 Copy to Clipboard Copied! Toggle word wrap Toggle overflow - 1
- The filename of the PersistentVolume created in the previous step.
6.9.3.5. Importing a virtual machine image to a block PersistentVolume using DataVolumes 링크 복사링크가 클립보드에 복사되었습니다!
You can import an existing virtual machine image into your OpenShift Container Platform cluster. Container-native virtualization uses DataVolumes to automate the importing data and the creation of an underlying PersistentVolumeClaim (PVC). You can then reference the DataVolume in a virtual machine configuration.
Prerequisites
-
A virtual machine disk image, in RAW, ISO, or QCOW2 format, optionally compressed by using
xzorgz. -
An
HTTPors3endpoint where the image is hosted, along with any authentication credentials needed to access the data source - At least one available block PV.
Procedure
If your data source requires authentication credentials, edit the
endpoint-secret.yamlfile, and apply the updated configuration to the cluster.Edit the
endpoint-secret.yamlfile with your preferred text editor:Copy to Clipboard Copied! Toggle word wrap Toggle overflow Update the secret:
oc apply -f endpoint-secret.yaml
$ oc apply -f endpoint-secret.yamlCopy to Clipboard Copied! Toggle word wrap Toggle overflow
Create a
DataVolumeconfiguration that specifies the data source for the image you want to import andvolumeMode: Blockso that an available block PV is used.Copy to Clipboard Copied! Toggle word wrap Toggle overflow Create the DataVolume to import the virtual machine image.
oc create -f <import-pv-datavolume.yaml>
$ oc create -f <import-pv-datavolume.yaml>1 Copy to Clipboard Copied! Toggle word wrap Toggle overflow - 1
- The filename DataVolume created in the previous step.
6.9.3.6. CDI supported operations matrix 링크 복사링크가 클립보드에 복사되었습니다!
This matrix shows the supported CDI operations for content types against endpoints, and which of these operations requires scratch space.
| Content types | HTTP | HTTPS | HTTP basic auth | Registry | Upload |
|---|---|---|---|---|---|
| KubeVirt(QCOW2) |
✓ QCOW2 |
✓ QCOW2** |
✓ QCOW2 |
✓ QCOW2* |
✓ QCOW2* |
| KubeVirt (RAW) |
✓ RAW |
✓ RAW |
✓ RAW |
✓ RAW* |
✓ RAW* |
| Archive+ | ✓ TAR | ✓ TAR | ✓ TAR | □ TAR | □ TAR |
✓ Supported operation
□ Unsupported operation
* Requires scratch space
** Requires scratch space if a custom certificate authority is required
+ Archive does not support block mode DVs
6.9.4. Importing a VMware virtual machine or template 링크 복사링크가 클립보드에 복사되었습니다!
You can import a single VMware virtual machine or template into your OpenShift Container Platform cluster.
If you import a VMware template, the wizard creates a virtual machine based on the template.
Importing a VMware virtual machine or template is a Technology Preview feature only. Technology Preview features are not supported with Red Hat production service level agreements (SLAs) and might not be functionally complete. Red Hat does not recommend using them in production. These features provide early access to upcoming product features, enabling customers to test functionality and provide feedback during the development process.
For more information about the support scope of Red Hat Technology Preview features, see https://access.redhat.com/support/offerings/techpreview/.
The import process uses the VMware Virtual Disk Development Kit (VDDK) to copy the VMware virtual disk. You can download the VDDK SDK, build a VDDK image, upload image to your image registry, and add it to the v2v-vmware ConfigMap.
You can import the VMware VM with the virtual machine wizard and then update the virtual machine’s network name.
6.9.4.1. Configuring an image registry for the VDDK image 링크 복사링크가 클립보드에 복사되었습니다!
You can configure either an internal OpenShift Container Platform image registry or a secure external image registry for the VDDK image.
Storing the VDDK image in a public registry might violate the terms of the VMware license.
6.9.4.1.1. Configuring an internal image registry 링크 복사링크가 클립보드에 복사되었습니다!
You can configure the internal OpenShift Container Platform image registry on bare metal by updating the Image Registry Operator configuration.
6.9.4.1.1.1. Changing the image registry’s management state 링크 복사링크가 클립보드에 복사되었습니다!
To start the image registry, you must change the Image Registry Operator configuration’s managementState from Removed to Managed.
Procedure
Change
managementStateImage Registry Operator configuration fromRemovedtoManaged. For example:oc patch configs.imageregistry.operator.openshift.io cluster --type merge --patch '{"spec":{"managementState":"Managed"}}'$ oc patch configs.imageregistry.operator.openshift.io cluster --type merge --patch '{"spec":{"managementState":"Managed"}}'Copy to Clipboard Copied! Toggle word wrap Toggle overflow
6.9.4.1.1.2. Configuring registry storage for bare metal 링크 복사링크가 클립보드에 복사되었습니다!
As a cluster administrator, following installation you must configure your registry to use storage.
Prerequisites
- Cluster administrator permissions.
- A cluster on bare metal.
- Provision persistent storage for your cluster, such as Red Hat OpenShift Container Storage. To deploy a private image registry, your storage must provide ReadWriteMany access mode.
- Must have "100Gi" capacity.
Procedure
To configure your registry to use storage, change the
spec.storage.pvcin theconfigs.imageregistry/clusterresource.NoteWhen using shared storage such as NFS, it is strongly recommended to use the
supplementalGroupsstrategy, which dictates the allowable supplemental groups for the Security Context, rather than thefsGroupID. Refer to the NFS Group IDs documentation for details.Verify you do not have a registry Pod:
oc get pod -n openshift-image-registry
$ oc get pod -n openshift-image-registryCopy to Clipboard Copied! Toggle word wrap Toggle overflow Note-
If the storage type is
emptyDIR, the replica number cannot be greater than1. If the storage type is
NFS, you must enable theno_wdelayandroot_squashmount options. For example:cat /etc/exports /mnt/data *(rw,sync,no_wdelay,root_squash,insecure,fsid=0) sh-4.3# exportfs -rv exporting *:/mnt/data
# cat /etc/exports /mnt/data *(rw,sync,no_wdelay,root_squash,insecure,fsid=0) sh-4.3# exportfs -rv exporting *:/mnt/dataCopy to Clipboard Copied! Toggle word wrap Toggle overflow
-
If the storage type is
Check the registry configuration:
oc edit configs.imageregistry.operator.openshift.io
$ oc edit configs.imageregistry.operator.openshift.io storage: pvc: claim:Copy to Clipboard Copied! Toggle word wrap Toggle overflow Leave the
claimfield blank to allow the automatic creation of animage-registry-storagePVC.Check the
clusteroperatorstatus:oc get clusteroperator image-registry
$ oc get clusteroperator image-registryCopy to Clipboard Copied! Toggle word wrap Toggle overflow
6.9.4.1.2. Configuring access to an internal image registry 링크 복사링크가 클립보드에 복사되었습니다!
You can access the OpenShift Container Platform internal registry directly, from within the cluster, or externally, by exposing the registry with a route.
6.9.4.1.2.1. Accessing registry directly from the cluster 링크 복사링크가 클립보드에 복사되었습니다!
You can access the registry from inside the cluster.
Procedure
Access the registry from the cluster by using internal routes:
Access the node by getting the node’s address:
oc get nodes oc debug nodes/<node_address>
$ oc get nodes $ oc debug nodes/<node_address>Copy to Clipboard Copied! Toggle word wrap Toggle overflow In order to have access to tools such as
ocandpodmanon the node, run the following command:chroot /host
sh-4.2# chroot /hostCopy to Clipboard Copied! Toggle word wrap Toggle overflow Log in to the container image registry by using your access token:
oc login -u kubeadmin -p <password_from_install_log> https://api-int.<cluster_name>.<base_domain>:6443 podman login -u kubeadmin -p $(oc whoami -t) image-registry.openshift-image-registry.svc:5000
sh-4.4# oc login -u kubeadmin -p <password_from_install_log> https://api-int.<cluster_name>.<base_domain>:6443 sh-4.4# podman login -u kubeadmin -p $(oc whoami -t) image-registry.openshift-image-registry.svc:5000Copy to Clipboard Copied! Toggle word wrap Toggle overflow You should see a message confirming login, such as:
Login Succeeded!
Login Succeeded!Copy to Clipboard Copied! Toggle word wrap Toggle overflow NoteYou can pass any value for the user name; the token contains all necessary information. Passing a user name that contains colons will result in a login failure.
Since the Image Registry Operator creates the route, it will likely be similar to
default-route-openshift-image-registry.<cluster_name>.Perform
podman pullandpodman pushoperations against your registry:ImportantYou can pull arbitrary images, but if you have the system:registry role added, you can only push images to the registry in your project.
In the following examples, use:
Expand Component Value <registry_ip>
172.30.124.220<port>
5000<project>
openshift<image>
image<tag>
omitted (defaults to
latest)Pull an arbitrary image:
podman pull name.io/image
$ podman pull name.io/imageCopy to Clipboard Copied! Toggle word wrap Toggle overflow Tag the new image with the form
<registry_ip>:<port>/<project>/<image>. The project name must appear in this pull specification for OpenShift Container Platform to correctly place and later access the image in the registry:podman tag name.io/image image-registry.openshift-image-registry.svc:5000/openshift/image
$ podman tag name.io/image image-registry.openshift-image-registry.svc:5000/openshift/imageCopy to Clipboard Copied! Toggle word wrap Toggle overflow NoteYou must have the
system:image-builderrole for the specified project, which allows the user to write or push an image. Otherwise, thepodman pushin the next step will fail. To test, you can create a new project to push the image.Push the newly-tagged image to your registry:
podman push image-registry.openshift-image-registry.svc:5000/openshift/image
$ podman push image-registry.openshift-image-registry.svc:5000/openshift/imageCopy to Clipboard Copied! Toggle word wrap Toggle overflow
6.9.4.1.2.2. Exposing a secure registry manually 링크 복사링크가 클립보드에 복사되었습니다!
Instead of logging in to the OpenShift Container Platform registry from within the cluster, you can gain external access to it by exposing it with a route. This allows you to log in to the registry from outside the cluster using the route address, and to tag and push images using the route host.
Prerequisites:
The following prerequisites are automatically performed:
- Deploy the Registry Operator.
- Deploy the Ingress Operator.
Procedure
You can expose the route by using DefaultRoute parameter in the configs.imageregistry.operator.openshift.io resource or by using custom routes.
To expose the registry using DefaultRoute:
Set
DefaultRoutetoTrue:oc patch configs.imageregistry.operator.openshift.io/cluster --patch '{"spec":{"defaultRoute":true}}' --type=merge$ oc patch configs.imageregistry.operator.openshift.io/cluster --patch '{"spec":{"defaultRoute":true}}' --type=mergeCopy to Clipboard Copied! Toggle word wrap Toggle overflow Log in with Podman:
HOST=$(oc get route default-route -n openshift-image-registry --template='{{ .spec.host }}') podman login -u $(oc whoami) -p $(oc whoami -t) --tls-verify=false $HOST$ HOST=$(oc get route default-route -n openshift-image-registry --template='{{ .spec.host }}') $ podman login -u $(oc whoami) -p $(oc whoami -t) --tls-verify=false $HOST1 Copy to Clipboard Copied! Toggle word wrap Toggle overflow - 1
--tls-verify=falseis needed if the cluster’s default certificate for routes is untrusted. You can set a custom, trusted certificate as the default certificate with the Ingress Operator.
To expose the registry using custom routes:
Create a secret with your route’s TLS keys:
oc create secret tls public-route-tls \ -n openshift-image-registry \ --cert=</path/to/tls.crt> \ --key=</path/to/tls.key>$ oc create secret tls public-route-tls \ -n openshift-image-registry \ --cert=</path/to/tls.crt> \ --key=</path/to/tls.key>Copy to Clipboard Copied! Toggle word wrap Toggle overflow This step is optional. If you do not create a secret, the route uses the default TLS configuration from the Ingress Operator.
On the Registry Operator:
Copy to Clipboard Copied! Toggle word wrap Toggle overflow Only set
secretNameif you are providing a custom TLS configuration for the registry’s route.
6.9.4.1.3. Configuring access to an external image registry 링크 복사링크가 클립보드에 복사되었습니다!
If you use an external image registry for the VDDK image, you can add the external image registry’s certificate authorities to the OpenShift Container Platform cluster.
Optionally, you can create a pull secret from your Docker credentials and add it to your service account.
6.9.4.1.3.1. Adding certificate authorities to the cluster 링크 복사링크가 클립보드에 복사되었습니다!
You can add certificate authorities (CAs) to the cluster for use when pushing and pulling images via the following procedure.
Prerequisites
- You must have cluster administrator privileges.
-
You must have access to the registry’s public certificates, usually a
hostname/ca.crtfile located in the/etc/docker/certs.d/directory.
Procedure
Create a ConfigMap in the
openshift-confignamespace containing the trusted certificates for the registries that use self-signed certificates. For each CA file, ensure the key in the ConfigMap is the registry’s hostname in thehostname[..port]format:oc create configmap registry-cas -n openshift-config \ --from-file=myregistry.corp.com..5000=/etc/docker/certs.d/myregistry.corp.com:5000/ca.crt \ --from-file=otherregistry.com=/etc/docker/certs.d/otherregistry.com/ca.crt
$ oc create configmap registry-cas -n openshift-config \ --from-file=myregistry.corp.com..5000=/etc/docker/certs.d/myregistry.corp.com:5000/ca.crt \ --from-file=otherregistry.com=/etc/docker/certs.d/otherregistry.com/ca.crtCopy to Clipboard Copied! Toggle word wrap Toggle overflow Update the cluster image configuration:
oc patch image.config.openshift.io/cluster --patch '{"spec":{"additionalTrustedCA":{"name":"registry-cas"}}}' --type=merge$ oc patch image.config.openshift.io/cluster --patch '{"spec":{"additionalTrustedCA":{"name":"registry-cas"}}}' --type=mergeCopy to Clipboard Copied! Toggle word wrap Toggle overflow
6.9.4.1.3.2. Allowing Pods to reference images from other secured registries 링크 복사링크가 클립보드에 복사되었습니다!
The .dockercfg $HOME/.docker/config.json file for Docker clients is a Docker credentials file that stores your authentication information if you have previously logged into a secured or insecure registry.
To pull a secured container image that is not from OpenShift Container Platform’s internal registry, you must create a pull secret from your Docker credentials and add it to your service account.
Procedure
If you already have a
.dockercfgfile for the secured registry, you can create a secret from that file by running:oc create secret generic <pull_secret_name> \ --from-file=.dockercfg=<path/to/.dockercfg> \ --type=kubernetes.io/dockercfg$ oc create secret generic <pull_secret_name> \ --from-file=.dockercfg=<path/to/.dockercfg> \ --type=kubernetes.io/dockercfgCopy to Clipboard Copied! Toggle word wrap Toggle overflow Or if you have a
$HOME/.docker/config.jsonfile:oc create secret generic <pull_secret_name> \ --from-file=.dockerconfigjson=<path/to/.docker/config.json> \ --type=kubernetes.io/dockerconfigjson$ oc create secret generic <pull_secret_name> \ --from-file=.dockerconfigjson=<path/to/.docker/config.json> \ --type=kubernetes.io/dockerconfigjsonCopy to Clipboard Copied! Toggle word wrap Toggle overflow If you do not already have a Docker credentials file for the secured registry, you can create a secret by running:
oc create secret docker-registry <pull_secret_name> \ --docker-server=<registry_server> \ --docker-username=<user_name> \ --docker-password=<password> \ --docker-email=<email>$ oc create secret docker-registry <pull_secret_name> \ --docker-server=<registry_server> \ --docker-username=<user_name> \ --docker-password=<password> \ --docker-email=<email>Copy to Clipboard Copied! Toggle word wrap Toggle overflow To use a secret for pulling images for Pods, you must add the secret to your service account. The name of the service account in this example should match the name of the service account the Pod uses.
defaultis the default service account:oc secrets link default <pull_secret_name> --for=pull
$ oc secrets link default <pull_secret_name> --for=pullCopy to Clipboard Copied! Toggle word wrap Toggle overflow To use a secret for pushing and pulling build images, the secret must be mountable inside of a Pod. You can do this by running:
oc secrets link builder <pull_secret_name>
$ oc secrets link builder <pull_secret_name>Copy to Clipboard Copied! Toggle word wrap Toggle overflow
6.9.4.2. Creating and using a VDDK image 링크 복사링크가 클립보드에 복사되었습니다!
You can download the VMware Virtual Disk Development Kit (VDDK), build a VDDK image, and push the VDDK image to your image registry. You then add the VDDK image to the v2v-vmware ConfigMap.
Prerequisites
- You must have access to an OpenShift Container Platform internal image registry or a secure external registry.
Procedure
Create and navigate to a temporary directory:
mkdir /tmp/<dir_name> && cd /tmp/<dir_name>
$ mkdir /tmp/<dir_name> && cd /tmp/<dir_name>Copy to Clipboard Copied! Toggle word wrap Toggle overflow - In a browser, navigate to VMware code and click SDKs.
- Under Compute Virtualization, click Virtual Disk Development Kit (VDDK).
- Select the latest VDDK release, click Download, and then save the VDDK archive in the temporary directory.
Extract the VDDK archive:
tar -xzf VMware-vix-disklib-<version>.x86_64.tar.gz
$ tar -xzf VMware-vix-disklib-<version>.x86_64.tar.gzCopy to Clipboard Copied! Toggle word wrap Toggle overflow Create a
Dockerfile:Copy to Clipboard Copied! Toggle word wrap Toggle overflow Build the image:
podman build . -t <registry_route_or_server_path>/vddk:<tag>
$ podman build . -t <registry_route_or_server_path>/vddk:<tag>1 Copy to Clipboard Copied! Toggle word wrap Toggle overflow - 1
- Specify your image registry:
-
For an internal OpenShift Container Platform registry, use the internal registry route, for example,
image-registry.openshift-image-registry.svc:5000/openshift/vddk:<tag>. -
For an external registry, specify the server name, path, and tag, for example,
server.example.com:5000/vddk:<tag>.
-
For an internal OpenShift Container Platform registry, use the internal registry route, for example,
Push the image to the registry:
podman push <registry_route_or_server_path>/vddk:<tag>
$ podman push <registry_route_or_server_path>/vddk:<tag>Copy to Clipboard Copied! Toggle word wrap Toggle overflow - Ensure that the image is accessible to your OpenShift Container Platform environment.
Edit the
v2v-vmwareConfigMap in the openshift-cnv project:oc edit configmap v2v-vmware -n openshift-cnv
$ oc edit configmap v2v-vmware -n openshift-cnvCopy to Clipboard Copied! Toggle word wrap Toggle overflow Add the
vddk-init-imageparameter to thedatastanza:... data: vddk-init-image: <registry_route_or_server_path>/vddk:<tag>
... data: vddk-init-image: <registry_route_or_server_path>/vddk:<tag>Copy to Clipboard Copied! Toggle word wrap Toggle overflow
6.9.4.3. Importing a VMware virtual machine or template with the virtual machine wizard 링크 복사링크가 클립보드에 복사되었습니다!
You can import a VMware virtual machine or template using the virtual machine wizard.
Prerequisites
-
You must create a VDDK image, push it to an image registry, and add it to the
v2v-vmwareConfigMap. There must be sufficient storage space for the imported disk.
WarningIf you try to import a virtual machine whose disk size is larger than the available storage space, the operation cannot complete. You will not be able to import another virtual machine or to clean up the storage because there are insufficient resources to support object deletion. To resolve this situation, you must add more object storage devices to the storage backend.
- The VMware virtual machine must be powered off.
Procedure
- In the container-native virtualization web console, click Workloads → Virtual Machines.
- Click Create Virtual Machine and select Import with Wizard.
In the General screen, perform the following steps:
- Select VMware from the Provider list.
Select Connect to New Instance or a saved vCenter instance from the vCenter instance list.
- If you select Connect to New Instance, fill in the vCenter hostname, Username, and Password.
- If you select a saved vCenter instance, the wizard connects to the vCenter instance using the saved credentials.
- Select a virtual machine or a template to import from the VM or Template to Import list.
- Select an operating system.
Select an existing flavor or Custom from the Flavor list.
If you select Custom, specify the Memory (GB) and the CPUs.
- Select a Workload Profile.
- If the virtual machine name is already being used by another virtual machine in the namespace, update the name.
- Click Next.
In the Networking screen, perform the following steps:
-
Click the Options menu
of a network interface and select Edit.
Enter a valid network interface name.
The name can contain lowercase letters (
a-z), numbers (0-9), and hyphens (-), up to a maximum of 253 characters. The first and last characters must be alphanumeric. The name must not contain uppercase letters, spaces, periods (.), or special characters.- Select the network interface model.
- Select the network definition.
- Select the network interface type.
- Enter the MAC address.
- Click Save and then click Next.
-
Click the Options menu
In the Storage screen, perform the following steps:
-
Click the Options menu
of a disk and select Edit.
Enter a valid name.
The name can contain lowercase letters (
a-z), numbers (0-9), and hyphens (-), up to a maximum of 253 characters. The first and last characters must be alphanumeric. The name must not contain uppercase letters, spaces, periods (.), or special characters.- Select the interface type.
Select the storage class.
If you do not select a storage class, container-native virtualization uses the default storage class to create the virtual machine.
- Click Save and then click Next.
-
Click the Options menu
-
In the Advanced screen, enter the Hostname and Authorized SSH Keys if you are using
cloud-init. - Click Next.
Review your settings and click Create Virtual Machine.
A Successfully created virtual machine message and a list of resources created for the virtual machine are displayed. The powered off virtual machine appears in Workloads → Virtual Machines.
Click See virtual machine details to view the dashboard of the imported virtual machine.
If an error occurs, perform the following steps:
- Click Workloads → Pods.
-
Click the Conversion Pod, for example,
kubevirt-v2v-conversion-rhel7-mini-1-27b9h. - Click Logs and check for error messages.
See virtual machine wizard fields for more information on the wizard fields.
6.9.4.4. Updating the imported VMware virtual machine’s NIC name 링크 복사링크가 클립보드에 복사되었습니다!
You must update the NIC name of a virtual machine imported from VMware to conform to container-native virtualization naming conventions.
Procedure
- Log in to the virtual machine.
-
Go to the
/etc/sysconfig/network-scriptsdirectory. Change the network configuration file name to
ifcfg-eth0:mv vmnic0 ifcfg-eth0
$ mv vmnic0 ifcfg-eth01 Copy to Clipboard Copied! Toggle word wrap Toggle overflow - 1
- Additional network configuration files are numbered sequentially, for example,
ifcfg-eth1,ifcfg-eth2.
Update the
NAMEandDEVICEparameters in the network configuration file:NAME=eth0 DEVICE=eth0
NAME=eth0 DEVICE=eth0Copy to Clipboard Copied! Toggle word wrap Toggle overflow Restart the network:
systemctl restart network
$ systemctl restart networkCopy to Clipboard Copied! Toggle word wrap Toggle overflow
6.9.4.5. Troubleshooting a VMware virtual machine import 링크 복사링크가 클립보드에 복사되었습니다!
If an imported virtual machine’s status is Import error: (VMware), you can check the Conversion Pod log for errors:
Obtain the Conversion Pod name:
oc get pods -n <project> | grep v2v
$ oc get pods -n <project> | grep v2v1 kubevirt-v2v-conversion-f66f7d-zqkz7 1/1 Running 0 4h49mCopy to Clipboard Copied! Toggle word wrap Toggle overflow - 1
- Specify the project of your imported virtual machine.
Obtain the Conversion Pod log:
oc logs kubevirt-v2v-conversion-f66f7d-zqkz7 -f -n <project>
$ oc logs kubevirt-v2v-conversion-f66f7d-zqkz7 -f -n <project>Copy to Clipboard Copied! Toggle word wrap Toggle overflow
6.9.4.5.1. Error messages 링크 복사링크가 클립보드에 복사되었습니다!
If an imported virtual machine event displays the error message,
Readiness probe failed, the following error message appears in the Conversion Pod log:INFO - have error: ('virt-v2v error: internal error: invalid argument: libvirt domain ‘v2v_migration_vm_1’ is running or paused. It must be shut down in order to perform virt-v2v conversion',)"INFO - have error: ('virt-v2v error: internal error: invalid argument: libvirt domain ‘v2v_migration_vm_1’ is running or paused. It must be shut down in order to perform virt-v2v conversion',)"Copy to Clipboard Copied! Toggle word wrap Toggle overflow You must ensure that the virtual machine is shut down before importing it.
6.9.4.5.2. Known issues 링크 복사링크가 클립보드에 복사되었습니다!
Your OpenShift Container Platform environment must have sufficient storage space for the imported disk.
If you try to import a virtual machine whose disk size is larger than the available storage space, the operation cannot complete. You will not be able to import another virtual machine or to clean up the storage because there are insufficient resources to support object deletion. To resolve this situation, you must add more object storage devices to the storage backend. (BZ#1721504)
- If you use NFS-backed storage for the 2 GB disk that is attached to the Conversion Pod, you must configure a hostPath volume. (BZ#1814611)
6.9.4.6. Virtual machine wizard fields 링크 복사링크가 클립보드에 복사되었습니다!
6.9.4.6.1. Virtual machine wizard fields 링크 복사링크가 클립보드에 복사되었습니다!
| Name | Parameter | Description |
|---|---|---|
| Template | Template from which to create the virtual machine. Selecting a template will automatically complete other fields. | |
| Source | PXE | Provision virtual machine from PXE menu. Requires a PXE-capable NIC in the cluster. |
| URL | Provision virtual machine from an image available from an HTTP or S3 endpoint. | |
| Container |
Provision virtual machine from a bootable operating system container located in a registry accessible from the cluster. Example: | |
| Disk | Provision virtual machine from a disk. | |
| Attach disk | Attach an existing disk that was previously cloned or created and made available in the PersistentVolumeClaims. When this option is selected, you must manually complete the Operating System, Flavor, and Workload Profile fields. | |
| Operating System | The primary operating system that is selected for the virtual machine. | |
| Flavor | small, medium, large, tiny, Custom | Presets that determine the amount of CPU and memory allocated to the virtual machine. The presets displayed for Flavor are determined by the operating system. |
| Workload Profile | High Performance | A virtual machine configuration that is optimized for high-performance workloads. |
| Server | A profile optimized to run server workloads. | |
| Desktop | A virtual machine configuration for use on a desktop. | |
| Name |
The name can contain lowercase letters ( | |
| Description | Optional description field. | |
| Start virtual machine on creation | Select to automatically start the virtual machine upon creation. |
6.9.4.6.2. Cloud-init fields 링크 복사링크가 클립보드에 복사되었습니다!
| Name | Description |
|---|---|
| Hostname | Sets a specific host name for the virtual machine. |
| Authenticated SSH Keys | The user’s public key that is copied to ~/.ssh/authorized_keys on the virtual machine. |
| Use custom script | Replaces other options with a field in which you paste a custom cloud-init script. |
6.9.4.6.3. Networking fields 링크 복사링크가 클립보드에 복사되었습니다!
| Name | Description |
|---|---|
| Name | Name for the Network Interface Card. |
| Model | Driver for the Network Interface Card or model for the Network Interface Card. |
| Network | List of available NetworkAttachmentDefinition objects. |
| Type |
List of available binding methods. For the default Pod network, |
| MAC Address | MAC address for the Network Interface Card. If a MAC address is not specified, an ephemeral address is generated for the session. |
6.9.4.6.4. Storage fields 링크 복사링크가 클립보드에 복사되었습니다!
| Name | Description |
|---|---|
| Source | Select a blank disk for the virtual machine or choose from the options available: PXE, Container, URL or Disk. To select an existing disk and attach it to the virtual machine, choose Attach Disk from a list of available PersistentVolumeClaims (PVCs) or from a cloned disk. |
| Name |
Name of the disk. The name can contain lowercase letters ( |
| Size (GiB) | Size, in GiB, of the disk. |
| Interface | Name of the interface. |
| Storage class |
Name of the underlying |
6.10. Cloning virtual machines 링크 복사링크가 클립보드에 복사되었습니다!
6.10.1. Enabling user permissions to clone DataVolumes across namespaces 링크 복사링크가 클립보드에 복사되었습니다!
The isolating nature of namespaces means that users cannot by default clone resources between namespaces.
To enable a user to clone a virtual machine to another namespace, a user with the cluster-admin role must create a new ClusterRole. Bind this ClusterRole to a user to enable them to clone virtual machines to the destination namespace.
6.10.1.1. Prerequisites 링크 복사링크가 클립보드에 복사되었습니다!
-
Only a user with the
cluster-adminrole can create ClusterRoles.
6.10.1.2. About DataVolumes 링크 복사링크가 클립보드에 복사되었습니다!
DataVolume objects are custom resources that are provided by the Containerized Data Importer (CDI) project. DataVolumes orchestrate import, clone, and upload operations that are associated with an underlying PersistentVolumeClaim (PVC). DataVolumes are integrated with KubeVirt, and they prevent a virtual machine from being started before the PVC has been prepared.
6.10.1.3. Creating RBAC resources for cloning DataVolumes 링크 복사링크가 클립보드에 복사되었습니다!
Create a new ClusterRole that enables permissions for all actions for the datavolumes resource.
Procedure
Create a ClusterRole manifest:
Copy to Clipboard Copied! Toggle word wrap Toggle overflow - 1
- Unique name for the ClusterRole.
Create the ClusterRole in the cluster:
oc create -f <datavolume-cloner.yaml>
$ oc create -f <datavolume-cloner.yaml>1 Copy to Clipboard Copied! Toggle word wrap Toggle overflow - 1
- The file name of the ClusterRole manifest created in the previous step.
Create a RoleBinding manifest that applies to both the source and destination namespaces and references the ClusterRole created in the previous step.
Copy to Clipboard Copied! Toggle word wrap Toggle overflow Create the RoleBinding in the cluster:
oc create -f <datavolume-cloner.yaml>
$ oc create -f <datavolume-cloner.yaml>1 Copy to Clipboard Copied! Toggle word wrap Toggle overflow - 1
- The file name of the RoleBinding manifest created in the previous step.
6.10.2. Cloning a virtual machine disk into a new DataVolume 링크 복사링크가 클립보드에 복사되었습니다!
You can clone the PersistentVolumeClaim (PVC) of a virtual machine disk into a new DataVolume by referencing the source PVC in your DataVolume configuration file.
6.10.2.1. Prerequisites 링크 복사링크가 클립보드에 복사되었습니다!
- You may need to define a StorageClass or prepare CDI scratch space for this operation to complete successfully. The CDI supported operations matrix shows the conditions that require scratch space.
- Users need additional permissions to clone the PVC of a virtual machine disk into another namespace.
6.10.2.2. About DataVolumes 링크 복사링크가 클립보드에 복사되었습니다!
DataVolume objects are custom resources that are provided by the Containerized Data Importer (CDI) project. DataVolumes orchestrate import, clone, and upload operations that are associated with an underlying PersistentVolumeClaim (PVC). DataVolumes are integrated with KubeVirt, and they prevent a virtual machine from being started before the PVC has been prepared.
6.10.2.3. Cloning the PersistentVolumeClaim of a virtual machine disk into a new DataVolume 링크 복사링크가 클립보드에 복사되었습니다!
You can clone a PersistentVolumeClaim (PVC) of an existing virtual machine disk into a new DataVolume. The new DataVolume can then be used for a new virtual machine.
When a DataVolume is created independently of a virtual machine, the lifecycle of the DataVolume is independent of the virtual machine. If the virtual machine is deleted, neither the DataVolume nor its associated PVC is deleted.
Prerequisites
- Determine the PVC of an existing virtual machine disk to use. You must power down the virtual machine that is associated with the PVC before you can clone it.
-
Install the OpenShift CLI (
oc).
Procedure
- Examine the virtual machine disk you want to clone to identify the name and namespace of the associated PVC.
Create a YAML file for a DataVolume object that specifies the name of the new DataVolume, the name and namespace of the source PVC, and the size of the new DataVolume.
For example:
Copy to Clipboard Copied! Toggle word wrap Toggle overflow Start cloning the PVC by creating the DataVolume:
oc create -f <cloner-datavolume>.yaml
$ oc create -f <cloner-datavolume>.yamlCopy to Clipboard Copied! Toggle word wrap Toggle overflow NoteDataVolumes prevent a virtual machine from starting before the PVC is prepared, so you can create a virtual machine that references the new DataVolume while the PVC clones.
6.10.2.4. Template: DataVolume clone configuration file 링크 복사링크가 클립보드에 복사되었습니다!
example-clone-dv.yaml
6.10.2.5. CDI supported operations matrix 링크 복사링크가 클립보드에 복사되었습니다!
This matrix shows the supported CDI operations for content types against endpoints, and which of these operations requires scratch space.
| Content types | HTTP | HTTPS | HTTP basic auth | Registry | Upload |
|---|---|---|---|---|---|
| KubeVirt(QCOW2) |
✓ QCOW2 |
✓ QCOW2** |
✓ QCOW2 |
✓ QCOW2* |
✓ QCOW2* |
| KubeVirt (RAW) |
✓ RAW |
✓ RAW |
✓ RAW |
✓ RAW* |
✓ RAW* |
| Archive+ | ✓ TAR | ✓ TAR | ✓ TAR | □ TAR | □ TAR |
✓ Supported operation
□ Unsupported operation
* Requires scratch space
** Requires scratch space if a custom certificate authority is required
+ Archive does not support block mode DVs
6.10.3. Cloning a virtual machine by using a DataVolumeTemplate 링크 복사링크가 클립보드에 복사되었습니다!
You can create a new virtual machine by cloning the PersistentVolumeClaim (PVC) of an existing VM. By including a dataVolumeTemplate in your virtual machine configuration file, you create a new DataVolume from the original PVC.
6.10.3.1. Prerequisites 링크 복사링크가 클립보드에 복사되었습니다!
- You may need to define a StorageClass or prepare CDI scratch space for this operation to complete successfully. The CDI supported operations matrix shows the conditions that require scratch space.
- Users need additional permissions to clone the PVC of a virtual machine disk into another namespace.
6.10.3.2. About DataVolumes 링크 복사링크가 클립보드에 복사되었습니다!
DataVolume objects are custom resources that are provided by the Containerized Data Importer (CDI) project. DataVolumes orchestrate import, clone, and upload operations that are associated with an underlying PersistentVolumeClaim (PVC). DataVolumes are integrated with KubeVirt, and they prevent a virtual machine from being started before the PVC has been prepared.
You can create a virtual machine that clones the PersistentVolumeClaim (PVC) of an existing virtual machine into a DataVolume. By referencing a dataVolumeTemplate in the virtual machine spec, the source PVC is cloned to a DataVolume, which is then automatically used for the creation of the virtual machine.
When a DataVolume is created as part of the DataVolumeTemplate of a virtual machine, the lifecycle of the DataVolume is then dependent on the virtual machine. If the virtual machine is deleted, the DataVolume and associated PVC are also deleted.
Prerequisites
- Determine the PVC of an existing virtual machine disk to use. You must power down the virtual machine that is associated with the PVC before you can clone it.
-
Install the OpenShift CLI (
oc).
Procedure
- Examine the virtual machine you want to clone to identify the name and namespace of the associated PVC.
Create a YAML file for a
VirtualMachineobject. The following virtual machine example clonesmy-favorite-vm-disk, which is located in thesource-namespacenamespace. The2GiDataVolume calledfavorite-cloneis created frommy-favorite-vm-disk.For example:
Copy to Clipboard Copied! Toggle word wrap Toggle overflow - 1
- The virtual machine to create.
Create the virtual machine with the PVC-cloned DataVolume:
oc create -f <vm-clone-datavolumetemplate>.yaml
$ oc create -f <vm-clone-datavolumetemplate>.yamlCopy to Clipboard Copied! Toggle word wrap Toggle overflow
6.10.3.4. Template: DataVolume virtual machine configuration file 링크 복사링크가 클립보드에 복사되었습니다!
example-dv-vm.yaml
- 1
- The
HTTPsource of the image you want to import, if applicable.
6.10.3.5. CDI supported operations matrix 링크 복사링크가 클립보드에 복사되었습니다!
This matrix shows the supported CDI operations for content types against endpoints, and which of these operations requires scratch space.
| Content types | HTTP | HTTPS | HTTP basic auth | Registry | Upload |
|---|---|---|---|---|---|
| KubeVirt(QCOW2) |
✓ QCOW2 |
✓ QCOW2** |
✓ QCOW2 |
✓ QCOW2* |
✓ QCOW2* |
| KubeVirt (RAW) |
✓ RAW |
✓ RAW |
✓ RAW |
✓ RAW* |
✓ RAW* |
| Archive+ | ✓ TAR | ✓ TAR | ✓ TAR | □ TAR | □ TAR |
✓ Supported operation
□ Unsupported operation
* Requires scratch space
** Requires scratch space if a custom certificate authority is required
+ Archive does not support block mode DVs
6.10.4. Cloning a virtual machine disk into a new block storage DataVolume 링크 복사링크가 클립보드에 복사되었습니다!
You can clone the PersistentVolumeClaim (PVC) of a virtual machine disk into a new block DataVolume by referencing the source PVC in your DataVolume configuration file.
6.10.4.1. Prerequisites 링크 복사링크가 클립보드에 복사되었습니다!
- If you require scratch space according to the CDI supported operations matrix, you must first define a StorageClass or prepare CDI scratch space for this operation to complete successfully.
- Users need additional permissions to clone the PVC of a virtual machine disk into another namespace.
6.10.4.2. About DataVolumes 링크 복사링크가 클립보드에 복사되었습니다!
DataVolume objects are custom resources that are provided by the Containerized Data Importer (CDI) project. DataVolumes orchestrate import, clone, and upload operations that are associated with an underlying PersistentVolumeClaim (PVC). DataVolumes are integrated with KubeVirt, and they prevent a virtual machine from being started before the PVC has been prepared.
6.10.4.3. About block PersistentVolumes 링크 복사링크가 클립보드에 복사되었습니다!
A block PersistentVolume (PV) is a PV that is backed by a raw block device. These volumes do not have a filesystem and can provide performance benefits for virtual machines that either write to the disk directly or implement their own storage service.
Raw block volumes are provisioned by specifying volumeMode: Block in the PV and PersistentVolumeClaim (PVC) specification.
6.10.4.4. Creating a local block PersistentVolume 링크 복사링크가 클립보드에 복사되었습니다!
Create a local block PersistentVolume (PV) on a node by populating a file and mounting it as a loop device. You can then reference this loop device in a PV configuration as a Block volume and use it as a block device for a virtual machine image.
Procedure
-
Log in as
rootto the node on which to create the local PV. This procedure usesnode01for its examples. Create a file and populate it with null characters so that it can be used as a block device. The following example creates a file
loop10with a size of 2Gb (20 100Mb blocks):dd if=/dev/zero of=<loop10> bs=100M count=20
$ dd if=/dev/zero of=<loop10> bs=100M count=20Copy to Clipboard Copied! Toggle word wrap Toggle overflow Mount the
loop10file as a loop device.losetup </dev/loop10>d3 <loop10>
$ losetup </dev/loop10>d3 <loop10>1 2 Copy to Clipboard Copied! Toggle word wrap Toggle overflow Create a
PersistentVolumeconfiguration that references the mounted loop device.Copy to Clipboard Copied! Toggle word wrap Toggle overflow Create the block PV.
oc create -f <local-block-pv10.yaml>
# oc create -f <local-block-pv10.yaml>1 Copy to Clipboard Copied! Toggle word wrap Toggle overflow - 1
- The filename of the PersistentVolume created in the previous step.
6.10.4.5. Cloning the PersistentVolumeClaim of a virtual machine disk into a new DataVolume 링크 복사링크가 클립보드에 복사되었습니다!
You can clone a PersistentVolumeClaim (PVC) of an existing virtual machine disk into a new DataVolume. The new DataVolume can then be used for a new virtual machine.
When a DataVolume is created independently of a virtual machine, the lifecycle of the DataVolume is independent of the virtual machine. If the virtual machine is deleted, neither the DataVolume nor its associated PVC is deleted.
Prerequisites
- Determine the PVC of an existing virtual machine disk to use. You must power down the virtual machine that is associated with the PVC before you can clone it.
-
Install the OpenShift CLI (
oc). - At least one available block PersistentVolume (PV) that is the same size as or larger than the source PVC.
Procedure
- Examine the virtual machine disk you want to clone to identify the name and namespace of the associated PVC.
Create a YAML file for a DataVolume object that specifies the name of the new DataVolume, the name and namespace of the source PVC,
volumeMode: Blockso that an available block PV is used, and the size of the new DataVolume.For example:
Copy to Clipboard Copied! Toggle word wrap Toggle overflow - 1
- The name of the new DataVolume.
- 2
- The namespace where the source PVC exists.
- 3
- The name of the source PVC.
- 4
- The size of the new DataVolume. You must allocate enough space, or the cloning operation fails. The size must be the same as or larger than the source PVC.
- 5
- Specifies that the destination is a block PV
Start cloning the PVC by creating the DataVolume:
oc create -f <cloner-datavolume>.yaml
$ oc create -f <cloner-datavolume>.yamlCopy to Clipboard Copied! Toggle word wrap Toggle overflow NoteDataVolumes prevent a virtual machine from starting before the PVC is prepared, so you can create a virtual machine that references the new DataVolume while the PVC clones.
6.10.4.6. CDI supported operations matrix 링크 복사링크가 클립보드에 복사되었습니다!
This matrix shows the supported CDI operations for content types against endpoints, and which of these operations requires scratch space.
| Content types | HTTP | HTTPS | HTTP basic auth | Registry | Upload |
|---|---|---|---|---|---|
| KubeVirt(QCOW2) |
✓ QCOW2 |
✓ QCOW2** |
✓ QCOW2 |
✓ QCOW2* |
✓ QCOW2* |
| KubeVirt (RAW) |
✓ RAW |
✓ RAW |
✓ RAW |
✓ RAW* |
✓ RAW* |
| Archive+ | ✓ TAR | ✓ TAR | ✓ TAR | □ TAR | □ TAR |
✓ Supported operation
□ Unsupported operation
* Requires scratch space
** Requires scratch space if a custom certificate authority is required
+ Archive does not support block mode DVs
6.11. Virtual machine networking 링크 복사링크가 클립보드에 복사되었습니다!
6.11.1. Using the default Pod network for virtual machines 링크 복사링크가 클립보드에 복사되었습니다!
You can use the default Pod network with container-native virtualization. To do so, you must use the masquerade binding method. It is the only recommended binding method for use with the default Pod network. Do not use masquerade mode with non-default networks.
For secondary networks, use the bridge binding method.
6.11.1.1. Configuring masquerade mode from the command line 링크 복사링크가 클립보드에 복사되었습니다!
You can use masquerade mode to hide a virtual machine’s outgoing traffic behind the Pod IP address. Masquerade mode uses Network Address Translation (NAT) to connect virtual machines to the Pod network backend through a Linux bridge.
Enable masquerade mode and allow traffic to enter the virtual machine by editing your virtual machine configuration file.
Prerequisites
- The virtual machine must be configured to use DHCP to acquire IPv4 addresses. The examples below are configured to use DHCP.
Procedure
Edit the
interfacesspec of your virtual machine configuration file:Copy to Clipboard Copied! Toggle word wrap Toggle overflow Create the virtual machine:
oc create -f <vm-name>.yaml
$ oc create -f <vm-name>.yamlCopy to Clipboard Copied! Toggle word wrap Toggle overflow
6.11.1.2. Selecting binding method 링크 복사링크가 클립보드에 복사되었습니다!
If you create a virtual machine from the container-native virtualization web console wizard, select the required binding method from the Networking screen.
6.11.1.2.1. Networking fields 링크 복사링크가 클립보드에 복사되었습니다!
| Name | Description |
|---|---|
| Name | Name for the Network Interface Card. |
| Model | Driver for the Network Interface Card or model for the Network Interface Card. |
| Network | List of available NetworkAttachmentDefinition objects. |
| Type |
List of available binding methods. For the default Pod network, |
| MAC Address | MAC address for the Network Interface Card. If a MAC address is not specified, an ephemeral address is generated for the session. |
6.11.1.3. Virtual machine configuration examples for the default network 링크 복사링크가 클립보드에 복사되었습니다!
6.11.1.3.1. Template: virtual machine configuration file 링크 복사링크가 클립보드에 복사되었습니다!
6.11.1.3.2. Template: Windows virtual machine instance configuration file 링크 복사링크가 클립보드에 복사되었습니다!
6.11.2. Attaching a virtual machine to multiple networks 링크 복사링크가 클립보드에 복사되었습니다!
Container-native virtualization provides layer-2 networking capabilities that allow you to connect virtual machines to multiple networks. You can import virtual machines with existing workloads that depend on access to multiple interfaces. You can also configure a PXE network so that you can boot machines over the network.
To get started, a network administrator configures a bridge NetworkAttachmentDefinition for a namespace in the web console or CLI. Users can then create a NIC to attach Pods and virtual machines in that namespace to the bridge network.
6.11.2.1. Container-native virtualization networking glossary 링크 복사링크가 클립보드에 복사되었습니다!
Container-native virtualization provides advanced networking functionality by using custom resources and plug-ins.
The following terms are used throughout container-native virtualization documentation:
- Container Network Interface (CNI)
- a Cloud Native Computing Foundation project, focused on container network connectivity. Container-native virtualization uses CNI plug-ins to build upon the basic Kubernetes networking functionality.
- Multus
- a "meta" CNI plug-in that allows multiple CNIs to exist so that a Pod or virtual machine can use the interfaces it needs.
- Custom Resource Definition (CRD)
- a Kubernetes API resource that allows you to define custom resources, or an object defined by using the CRD API resource.
- NetworkAttachmentDefinition
- a CRD introduced by the Multus project that allows you to attach Pods, virtual machines, and virtual machine instances to one or more networks.
- Preboot eXecution Environment (PXE)
- an interface that enables an administrator to boot a client machine from a server over the network. Network booting allows you to remotely load operating systems and other software onto the client.
6.11.2.2. Creating a NetworkAttachmentDefinition 링크 복사링크가 클립보드에 복사되었습니다!
6.11.2.2.1. Creating a Linux bridge NetworkAttachmentDefinition in the web console 링크 복사링크가 클립보드에 복사되었습니다!
The NetworkAttachmentDefinition is a custom resource that exposes layer-2 devices to a specific namespace in your container-native virtualization cluster.
Network administrators can create NetworkAttachmentDefinitions to provide existing layer-2 networking to Pods and virtual machines.
Prerequisites
- Container-native virtualization 2.2 or above installed on your cluster.
- A Linux bridge must be configured and attached to the correct Network Interface Card (NIC) on every node.
-
If you use VLANs,
vlan_filteringmust be enabled on the bridge. The NIC must be tagged to all relevant VLANs.
-
For example:
bridge vlan add dev bond0 vid 1-4095 master
-
For example:
Procedure
- In the web console, click Networking → Network Attachment Definitions.
- Click Create Network Attachment Definition .
- Enter a unique Name and optional Description.
- Click the Network Type list and select CNV Linux bridge.
- Enter the name of the bridge in the Bridge Name field.
- (Optional) If the resource has VLAN IDs configured, enter the ID numbers in the VLAN Tag Number field.
- Click Create.
6.11.2.2.2. Creating a Linux bridge NetworkAttachmentDefinition in the CLI 링크 복사링크가 클립보드에 복사되었습니다!
As a network administrator, you can configure a NetworkAttachmentDefinition of type cnv-bridge to provide Layer-2 networking to Pods and virtual machines.
The NetworkAttachmentDefinition must be in the same namespace as the Pod or virtual machine.
Prerequisites
- Container-native virtualization 2.0 or newer
- A Linux bridge must be configured and attached to the correct Network Interface Card on every node.
-
If you use VLANs,
vlan_filteringmust be enabled on the bridge. The NIC must be tagged to all relevant VLANs.
-
For example:
bridge vlan add dev bond0 vid 1-4095 master
-
For example:
Procedure
Create a new file for the NetworkAttachmentDefinition in any local directory. The file must have the following contents, modified to match your configuration:
Copy to Clipboard Copied! Toggle word wrap Toggle overflow - 1
- If you add this annotation to your NetworkAttachmentDefinition, your virtual machine instances will only run on nodes that have the
br0bridge connected. - 2
- Required. A name for the configuration.
- 3
- The actual name of the Container Network Interface (CNI) plug-in that provides the network for this NetworkAttachmentDefinition. Do not change this field unless you want to use a different CNI.
- 4
- You must substitute the actual name of the bridge, if it is not
br0. - 5
- Required. This allows the MAC pool manager to assign a unique MAC address to the connection.
oc create -f <resource_spec.yaml>
$ oc create -f <resource_spec.yaml>Copy to Clipboard Copied! Toggle word wrap Toggle overflow Edit the configuration file of a virtual machine or virtual machine instance that you want to connect to the bridge network:
Copy to Clipboard Copied! Toggle word wrap Toggle overflow - 1
- You must substitute the actual
namevalue from the NetworkAttachmentDefinition.
Apply the configuration file to the resource:
oc create -f <local/path/to/network-attachment-definition.yaml>
$ oc create -f <local/path/to/network-attachment-definition.yaml>Copy to Clipboard Copied! Toggle word wrap Toggle overflow
When defining the NIC in the next section, ensure that the NETWORK value is the bridge network name from the NetworkAttachmentDefinition you created in the previous section.
6.11.2.3. Creating a NIC for a virtual machine 링크 복사링크가 클립보드에 복사되었습니다!
Create and attach additional NICs to a virtual machine from the web console.
Procedure
- In the correct project in the container-native virtualization console, click Workloads → Virtual Machines.
- Select a virtual machine.
- Click Network Interfaces to display the NICs already attached to the virtual machine.
- Click Create Network Interface to create a new slot in the list.
- Fill in the Name, Model, Network, Type, and MAC Address for the new NIC.
- Click the ✓ button to save and attach the NIC to the virtual machine.
6.11.2.4. Networking fields 링크 복사링크가 클립보드에 복사되었습니다!
| Name | Description |
|---|---|
| Name | Name for the Network Interface Card. |
| Model | Driver for the Network Interface Card or model for the Network Interface Card. |
| Network | List of available NetworkAttachmentDefinition objects. |
| Type |
List of available binding methods. For the default Pod network, |
| MAC Address | MAC address for the Network Interface Card. If a MAC address is not specified, an ephemeral address is generated for the session. |
Install the optional QEMU guest agent on the virtual machine so that the host can display relevant information about the additional networks.
6.11.3. Installing the QEMU guest agent on virtual machines 링크 복사링크가 클립보드에 복사되었습니다!
The QEMU guest agent is a daemon that runs on the virtual machine. The agent passes network information on the virtual machine, notably the IP address of additional networks, to the host.
6.11.3.1. Prerequisites 링크 복사링크가 클립보드에 복사되었습니다!
Verify that the guest agent is installed and running by entering the following command:
systemctl status qemu-guest-agent
$ systemctl status qemu-guest-agentCopy to Clipboard Copied! Toggle word wrap Toggle overflow
6.11.3.2. Installing QEMU guest agent on a Linux virtual machine 링크 복사링크가 클립보드에 복사되었습니다!
The qemu-guest-agent is widely available and available by default in Red Hat virtual machines. Install the agent and start the service
Procedure
- Access the virtual machine command line through one of the consoles or by SSH.
Install the QEMU guest agent on the virtual machine:
yum install -y qemu-guest-agent
$ yum install -y qemu-guest-agentCopy to Clipboard Copied! Toggle word wrap Toggle overflow Start the QEMU guest agent service:
systemctl start qemu-guest-agent
$ systemctl start qemu-guest-agentCopy to Clipboard Copied! Toggle word wrap Toggle overflow Ensure the service is persistent:
systemctl enable qemu-guest-agent
$ systemctl enable qemu-guest-agentCopy to Clipboard Copied! Toggle word wrap Toggle overflow
You can also install and start the QEMU guest agent by using the custom script field in the cloud-init section of the wizard when creating either virtual machines or virtual machines templates in the web console.
6.11.3.3. Installing QEMU guest agent on a Windows virtual machine 링크 복사링크가 클립보드에 복사되었습니다!
For Windows virtual machines, the QEMU guest agent is included in the VirtIO drivers, which can be installed using one of the following procedures:
6.11.3.3.1. Installing VirtIO drivers on an existing Windows virtual machine 링크 복사링크가 클립보드에 복사되었습니다!
Install the VirtIO drivers from the attached SATA CD drive to an existing Windows virtual machine.
This procedure uses a generic approach to adding drivers to Windows. The process might differ slightly between versions of Windows. Refer to the installation documentation for your version of Windows for specific installation steps.
Procedure
- Start the virtual machine and connect to a graphical console.
- Log in to a Windows user session.
Open Device Manager and expand Other devices to list any Unknown device.
-
Open the
Device Propertiesto identify the unknown device. Right-click the device and select Properties. - Click the Details tab and select Hardware Ids in the Property list.
- Compare the Value for the Hardware Ids with the supported VirtIO drivers.
-
Open the
- Right-click the device and select Update Driver Software.
- Click Browse my computer for driver software and browse to the attached SATA CD drive, where the VirtIO drivers are located. The drivers are arranged hierarchically according to their driver type, operating system, and CPU architecture.
- Click Next to install the driver.
- Repeat this process for all the necessary VirtIO drivers.
- After the driver installs, click Close to close the window.
- Reboot the virtual machine to complete the driver installation.
6.11.3.3.2. Installing VirtIO drivers during Windows installation 링크 복사링크가 클립보드에 복사되었습니다!
Install the VirtIO drivers from the attached SATA CD driver during Windows installation.
This procedure uses a generic approach to the Windows installation and the installation method might differ between versions of Windows. Refer to the documentation for the version of Windows that you are installing.
Procedure
- Start the virtual machine and connect to a graphical console.
- Begin the Windows installation process.
- Select the Advanced installation.
-
The storage destination will not be recognized until the driver is loaded. Click
Load driver. - The drivers are attached as a SATA CD drive. Click OK and browse the CD drive for the storage driver to load. The drivers are arranged hierarchically according to their driver type, operating system, and CPU architecture.
- Repeat the previous two steps for all required drivers.
- Complete the Windows installation.
6.11.4. Viewing the IP address of NICs on a virtual machine 링크 복사링크가 클립보드에 복사되었습니다!
The QEMU guest agent runs on the virtual machine and passes the IP address of attached NICs to the host, allowing you to view the IP address from both the web console and the oc client.
6.11.4.1. Prerequisites 링크 복사링크가 클립보드에 복사되었습니다!
Verify that the guest agent is installed and running by entering the following command:
systemctl status qemu-guest-agent
$ systemctl status qemu-guest-agentCopy to Clipboard Copied! Toggle word wrap Toggle overflow - If the guest agent is not installed and running, install and run the guest agent on the virtual machine.
6.11.4.2. Viewing the IP address of a virtual machine interface in the CLI 링크 복사링크가 클립보드에 복사되었습니다!
The network interface configuration is included in the oc describe vmi <vmi_name> command.
You can also view the IP address information by running ip addr on the virtual machine, or by running oc get vmi <vmi_name> -o yaml.
Procedure
Use the
oc describecommand to display the virtual machine interface configuration:Copy to Clipboard Copied! Toggle word wrap Toggle overflow
6.11.4.3. Viewing the IP address of a virtual machine interface in the web console 링크 복사링크가 클립보드에 복사되었습니다!
The IP information displays in the Virtual Machine Overview screen for the virtual machine.
Procedure
- In the container-native virtualization console, click Workloads → Virtual Machines.
- Click the virtual machine name to open the Virtual Machine Overview screen.
The information for each attached NIC is displayed under IP ADDRESSES.
6.12. Virtual machine disks 링크 복사링크가 클립보드에 복사되었습니다!
6.12.1. Configuring local storage for virtual machines 링크 복사링크가 클립보드에 복사되었습니다!
You can configure local storage for your virtual machines by using the hostpath provisioner feature.
6.12.1.1. About the hostpath provisioner 링크 복사링크가 클립보드에 복사되었습니다!
The hostpath provisioner is a local storage provisioner designed for container-native virtualization. If you want to configure local storage for virtual machines, you must enable the hostpath provisioner first.
When you install the container-native virtualization Operator, the hostpath provisioner Operator is automatically installed. To use it, you must:
Configure SELinux:
- If you use Red Hat Enterprise Linux CoreOS 8 workers, you must create a MachineConfig object on each node.
-
Otherwise, apply the SELinux label
container_file_tto the PersistentVolume (PV) backing directory on each node.
- Create a HostPathProvisioner custom resource.
- Create a StorageClass object for the hostpath provisioner.
The hostpath provisioner Operator deploys the provisioner as a DaemonSet on each node when you create its custom resource. In the custom resource file, you specify the backing directory for the PersistentVolumes that the hostpath provisioner creates.
You must configure SELinux before you create the HostPathProvisioner custom resource. To configure SELinux on Red Hat Enterprise Linux CoreOS 8 workers, you must create a MachineConfig object on each node.
If you do not use Red Hat Enterprise Linux CoreOS workers, skip this procedure.
Prerequisites
- Create a backing directory on each node for the PersistentVolumes (PVs) that the hostpath provisioner creates.
If you select a directory that shares space with your operating system, you can exhaust the space on that partition, causing the node to stop functioning. To avoid this issue, create a separate partition and point the hostpath provisioner to that directory.
Procedure
Create the MachineConfig file. For example:
touch machineconfig.yaml
$ touch machineconfig.yamlCopy to Clipboard Copied! Toggle word wrap Toggle overflow Edit the file, ensuring that you include the directory where you want the hostpath provisioner to create PVs. For example:
Copy to Clipboard Copied! Toggle word wrap Toggle overflow - 1
- Specify the backing directory where you want the provisioner to create PVs.
Create the MachineConfig object:
oc create -f machineconfig.yaml -n <namespace>
$ oc create -f machineconfig.yaml -n <namespace>Copy to Clipboard Copied! Toggle word wrap Toggle overflow
6.12.1.3. Using the hostpath provisioner to enable local storage 링크 복사링크가 클립보드에 복사되었습니다!
To deploy the hostpath provisioner and enable your virtual machines to use local storage, first create a HostPathProvisioner custom resource.
Prerequisites
- Create a backing directory on each node for the PersistentVolumes (PVs) that the hostpath provisioner creates.
If you select a directory that shares space with your operating system, you can exhaust the space on that partition, causing the node to stop functioning. To avoid this issue, create a separate partition and point the hostpath provisioner to that directory.
Apply the SELinux context
container_file_tto the PV backing directory on each node. For example:sudo chcon -t container_file_t -R </path/to/backing/directory>
$ sudo chcon -t container_file_t -R </path/to/backing/directory>Copy to Clipboard Copied! Toggle word wrap Toggle overflow NoteIf you use Red Hat Enterprise Linux CoreOS 8 workers, you must configure SELinux by using a MachineConfig manifest instead.
Procedure
Create the HostPathProvisioner custom resource file. For example:
touch hostpathprovisioner_cr.yaml
$ touch hostpathprovisioner_cr.yamlCopy to Clipboard Copied! Toggle word wrap Toggle overflow Edit the file, ensuring that the
spec.pathConfig.pathvalue is the directory where you want the hostpath provisioner to create PVs. For example:Copy to Clipboard Copied! Toggle word wrap Toggle overflow NoteIf you did not create the backing directory, the provisioner attempts to create it for you. If you did not apply the
container_file_tSELinux context, this can causePermission deniederrors.Create the custom resource in the
openshift-cnvnamespace:oc create -f hostpathprovisioner_cr.yaml -n openshift-cnv
$ oc create -f hostpathprovisioner_cr.yaml -n openshift-cnvCopy to Clipboard Copied! Toggle word wrap Toggle overflow
6.12.1.4. Creating a StorageClass object 링크 복사링크가 클립보드에 복사되었습니다!
When you create a StorageClass object, you set parameters that affect the dynamic provisioning of PersistentVolumes (PVs) that belong to that storage class.
You cannot update a StorageClass object’s parameters after you create it.
Procedure
Create a YAML file for defining the storage class. For example:
touch storageclass.yaml
$ touch storageclass.yamlCopy to Clipboard Copied! Toggle word wrap Toggle overflow Edit the file. For example:
Copy to Clipboard Copied! Toggle word wrap Toggle overflow - 1
- You can optionally rename the storage class by changing this value.
- 2
- The two possible
reclaimPolicyvalues areDeleteandRetain. If you do not specify a value, the storage class defaults toDelete. - 3
- The
volumeBindingModevalue determines when dynamic provisioning and volume binding occur. SpecifyWaitForFirstConsumerto delay the binding and provisioning of a PV until after a Pod that uses the PersistentVolumeClaim (PVC) is created. This ensures that the PV meets the Pod’s scheduling requirements.
Create the StorageClass object:
oc create -f storageclass.yaml
$ oc create -f storageclass.yamlCopy to Clipboard Copied! Toggle word wrap Toggle overflow
Additional information
6.12.2. Uploading local disk images by using the virtctl tool 링크 복사링크가 클립보드에 복사되었습니다!
You can upload a disk image that is stored locally by using the virtctl command-line utility.
6.12.2.1. Prerequisites 링크 복사링크가 클립보드에 복사되었습니다!
-
Install the
kubevirt-virtctlpackage - You may need to define a StorageClass or prepare CDI scratch space for this operation to complete successfully.
6.12.2.2. CDI supported operations matrix 링크 복사링크가 클립보드에 복사되었습니다!
This matrix shows the supported CDI operations for content types against endpoints, and which of these operations requires scratch space.
| Content types | HTTP | HTTPS | HTTP basic auth | Registry | Upload |
|---|---|---|---|---|---|
| KubeVirt(QCOW2) |
✓ QCOW2 |
✓ QCOW2** |
✓ QCOW2 |
✓ QCOW2* |
✓ QCOW2* |
| KubeVirt (RAW) |
✓ RAW |
✓ RAW |
✓ RAW |
✓ RAW* |
✓ RAW* |
| Archive+ | ✓ TAR | ✓ TAR | ✓ TAR | □ TAR | □ TAR |
✓ Supported operation
□ Unsupported operation
* Requires scratch space
** Requires scratch space if a custom certificate authority is required
+ Archive does not support block mode DVs
6.12.2.3. Uploading a local disk image to a new PersistentVolumeClaim 링크 복사링크가 클립보드에 복사되었습니다!
You can use the virtctl CLI utility to upload a virtual machine disk image from a client machine to your cluster. Uploading the disk image creates a PersistentVolumeClaim (PVC) that you can associate with a virtual machine.
Prerequisites
-
A virtual machine disk image, in RAW, ISO, or QCOW2 format, optionally compressed by using
xzorgz. -
The
kubevirt-virtctlpackage must be installed on the client machine. - The client machine must be configured to trust the OpenShift Container Platform router’s certificate.
Procedure
Identify the following items:
- File location of the VM disk image you want to upload
- Name and size required for the resulting PVC
Run the
virtctl image-uploadcommand to upload your VM image. You must specify the PVC name, PVC size, and file location. For example:virtctl image-upload --pvc-name=<upload-fedora-pvc> --pvc-size=<2Gi> --image-path=</images/fedora.qcow2>
$ virtctl image-upload --pvc-name=<upload-fedora-pvc> --pvc-size=<2Gi> --image-path=</images/fedora.qcow2>Copy to Clipboard Copied! Toggle word wrap Toggle overflow WarningTo allow insecure server connections when using HTTPS, use the
--insecureparameter. Be aware that when you use the--insecureflag, the authenticity of the upload endpoint is not verified.To verify that the PVC was created, view all PVC objects:
oc get pvc
$ oc get pvcCopy to Clipboard Copied! Toggle word wrap Toggle overflow
6.12.3. Uploading a local disk image to a block storage DataVolume 링크 복사링크가 클립보드에 복사되었습니다!
You can upload a local disk image into a block DataVolume by using the virtctl command-line utility.
In this workflow, you create a local block device to use as a PersistentVolume, associate this block volume with an upload DataVolume, and use virtctl to upload the local disk image into the DataVolume.
6.12.3.1. Prerequisites 링크 복사링크가 클립보드에 복사되었습니다!
- If you require scratch space according to the CDI supported operations matrix, you must first define a StorageClass or prepare CDI scratch space for this operation to complete successfully.
6.12.3.2. About DataVolumes 링크 복사링크가 클립보드에 복사되었습니다!
DataVolume objects are custom resources that are provided by the Containerized Data Importer (CDI) project. DataVolumes orchestrate import, clone, and upload operations that are associated with an underlying PersistentVolumeClaim (PVC). DataVolumes are integrated with KubeVirt, and they prevent a virtual machine from being started before the PVC has been prepared.
6.12.3.3. About block PersistentVolumes 링크 복사링크가 클립보드에 복사되었습니다!
A block PersistentVolume (PV) is a PV that is backed by a raw block device. These volumes do not have a filesystem and can provide performance benefits for virtual machines that either write to the disk directly or implement their own storage service.
Raw block volumes are provisioned by specifying volumeMode: Block in the PV and PersistentVolumeClaim (PVC) specification.
6.12.3.4. Creating a local block PersistentVolume 링크 복사링크가 클립보드에 복사되었습니다!
Create a local block PersistentVolume (PV) on a node by populating a file and mounting it as a loop device. You can then reference this loop device in a PV configuration as a Block volume and use it as a block device for a virtual machine image.
Procedure
-
Log in as
rootto the node on which to create the local PV. This procedure usesnode01for its examples. Create a file and populate it with null characters so that it can be used as a block device. The following example creates a file
loop10with a size of 2Gb (20 100Mb blocks):dd if=/dev/zero of=<loop10> bs=100M count=20
$ dd if=/dev/zero of=<loop10> bs=100M count=20Copy to Clipboard Copied! Toggle word wrap Toggle overflow Mount the
loop10file as a loop device.losetup </dev/loop10>d3 <loop10>
$ losetup </dev/loop10>d3 <loop10>1 2 Copy to Clipboard Copied! Toggle word wrap Toggle overflow Create a
PersistentVolumeconfiguration that references the mounted loop device.Copy to Clipboard Copied! Toggle word wrap Toggle overflow Create the block PV.
oc create -f <local-block-pv10.yaml>
# oc create -f <local-block-pv10.yaml>1 Copy to Clipboard Copied! Toggle word wrap Toggle overflow - 1
- The filename of the PersistentVolume created in the previous step.
6.12.3.5. Creating an upload DataVolume 링크 복사링크가 클립보드에 복사되었습니다!
Create a DataVolume with an upload data source to use for uploading local disk images.
Procedure
Create a DataVolume configuration that specifies
spec: source: upload{}:Copy to Clipboard Copied! Toggle word wrap Toggle overflow Create the DataVolume:
oc create -f <upload-datavolume>.yaml
$ oc create -f <upload-datavolume>.yamlCopy to Clipboard Copied! Toggle word wrap Toggle overflow
6.12.3.6. Uploading a local disk image to a new DataVolume 링크 복사링크가 클립보드에 복사되었습니다!
You can use the virtctl CLI utility to upload a virtual machine disk image from a client machine to a DataVolume (DV) in your cluster. After you upload the image, you can add it to a virtual machine.
Prerequisites
-
A virtual machine disk image, in RAW, ISO, or QCOW2 format, optionally compressed by using
xzorgz. -
The
kubevirt-virtctlpackage must be installed on the client machine. - The client machine must be configured to trust the OpenShift Container Platform router’s certificate.
- A spare DataVolume that is the same size or larger than the disk that you are uploading.
Procedure
Identify the following items:
- File location of the VM disk image that you want to upload
- Name of the DataVolume
Run the
virtctl image-uploadcommand to upload your disk image. You must specify the DV name and file location. For example:virtctl image-upload --dv-name=<upload-datavolume> --image-path=</images/fedora.qcow2>
$ virtctl image-upload --dv-name=<upload-datavolume> --image-path=</images/fedora.qcow2>1 2 Copy to Clipboard Copied! Toggle word wrap Toggle overflow WarningTo allow insecure server connections when using HTTPS, use the
--insecureparameter. Be aware that when you use the--insecureflag, the authenticity of the upload endpoint is not verified.To verify that the DV was created, view all DV objects:
oc get dvs
$ oc get dvsCopy to Clipboard Copied! Toggle word wrap Toggle overflow
6.12.3.7. CDI supported operations matrix 링크 복사링크가 클립보드에 복사되었습니다!
This matrix shows the supported CDI operations for content types against endpoints, and which of these operations requires scratch space.
| Content types | HTTP | HTTPS | HTTP basic auth | Registry | Upload |
|---|---|---|---|---|---|
| KubeVirt(QCOW2) |
✓ QCOW2 |
✓ QCOW2** |
✓ QCOW2 |
✓ QCOW2* |
✓ QCOW2* |
| KubeVirt (RAW) |
✓ RAW |
✓ RAW |
✓ RAW |
✓ RAW* |
✓ RAW* |
| Archive+ | ✓ TAR | ✓ TAR | ✓ TAR | □ TAR | □ TAR |
✓ Supported operation
□ Unsupported operation
* Requires scratch space
** Requires scratch space if a custom certificate authority is required
+ Archive does not support block mode DVs
6.12.4. Moving a local virtual machine disk to a different node 링크 복사링크가 클립보드에 복사되었습니다!
Virtual machines that use local volume storage can be moved so that they run on a specific node.
You might want to move the virtual machine to a specific node for the following reasons:
- The current node has limitations to the local storage configuration.
- The new node is better optimized for the workload of that virtual machine.
To move a virtual machine that uses local storage, you must clone the underlying volume by using a DataVolume. After the cloning operation is complete, you can edit the virtual machine configuration so that it uses the new DataVolume, or add the new DataVolume to another virtual machine.
Users without the cluster-admin role require additional user permissions in order to clone volumes across namespaces.
6.12.4.1. Cloning a local volume to another node 링크 복사링크가 클립보드에 복사되었습니다!
You can move a virtual machine disk so that it runs on a specific node by cloning the underlying PersistentVolumeClaim (PVC).
To ensure the virtual machine disk is cloned to the correct node, you must either create a new PersistentVolume (PV) or identify one on the correct node. Apply a unique label to the PV so that it can be referenced by the DataVolume.
The destination PV must be the same size or larger than the source PVC. If the destination PV is smaller than the source PVC, the cloning operation fails.
Prerequisites
- The virtual machine must not be running. Power down the virtual machine before cloning the virtual machine disk.
Procedure
Either create a new local PV on the node, or identify a local PV already on the node:
Create a local PV that includes the
nodeAffinity.nodeSelectorTermsparameters. The following manifest creates a10Gilocal PV onnode01.Copy to Clipboard Copied! Toggle word wrap Toggle overflow Identify a PV that already exists on the target node. You can identify the node where a PV is provisioned by viewing the
nodeAffinityfield in its configuration:oc get pv <destination-pv> -o yaml
$ oc get pv <destination-pv> -o yamlCopy to Clipboard Copied! Toggle word wrap Toggle overflow The following snippet shows that the PV is on
node01:Copy to Clipboard Copied! Toggle word wrap Toggle overflow
Add a unique label to the PV:
oc label pv <destination-pv> node=node01
$ oc label pv <destination-pv> node=node01Copy to Clipboard Copied! Toggle word wrap Toggle overflow Create a DataVolume manifest that references the following:
- The PVC name and namespace of the virtual machine.
- The label you applied to the PV in the previous step.
The size of the destination PV.
Copy to Clipboard Copied! Toggle word wrap Toggle overflow - 1
- The name of the new DataVolume.
- 2
- The name of the source PVC. If you do not know the PVC name, you can find it in the virtual machine configuration:
spec.volumes.persistentVolumeClaim.claimName. - 3
- The namespace where the source PVC exists.
- 4
- The label that you applied to the PV in the previous step.
- 5
- The size of the destination PV.
Start the cloning operation by applying the DataVolume manifest to your cluster:
oc apply -f <clone-datavolume.yaml>
$ oc apply -f <clone-datavolume.yaml>Copy to Clipboard Copied! Toggle word wrap Toggle overflow
The DataVolume clones the PVC of the virtual machine into the PV on the specific node.
6.12.5. Expanding virtual storage by adding blank disk images 링크 복사링크가 클립보드에 복사되었습니다!
You can increase your storage capacity or create new data partitions by adding blank disk images to container-native virtualization.
6.12.5.1. About DataVolumes 링크 복사링크가 클립보드에 복사되었습니다!
DataVolume objects are custom resources that are provided by the Containerized Data Importer (CDI) project. DataVolumes orchestrate import, clone, and upload operations that are associated with an underlying PersistentVolumeClaim (PVC). DataVolumes are integrated with KubeVirt, and they prevent a virtual machine from being started before the PVC has been prepared.
6.12.5.2. Creating a blank disk image with DataVolumes 링크 복사링크가 클립보드에 복사되었습니다!
You can create a new blank disk image in a PersistentVolumeClaim by customizing and deploying a DataVolume configuration file.
Prerequisites
- At least one available PersistentVolume.
-
Install the OpenShift CLI (
oc).
Procedure
Edit the DataVolume configuration file:
Copy to Clipboard Copied! Toggle word wrap Toggle overflow Create the blank disk image by running the following command:
oc create -f <blank-image-datavolume>.yaml
$ oc create -f <blank-image-datavolume>.yamlCopy to Clipboard Copied! Toggle word wrap Toggle overflow
6.12.5.3. Template: DataVolume configuration file for blank disk images 링크 복사링크가 클립보드에 복사되었습니다!
blank-image-datavolume.yaml
6.12.6. Storage defaults for DataVolumes 링크 복사링크가 클립보드에 복사되었습니다!
The kubevirt-storage-class-defaults ConfigMap provides access mode and volume mode defaults for DataVolumes. You can edit or add storage class defaults to the ConfigMap in order to create DataVolumes in the web console that better match the underlying storage.
6.12.6.1. About storage settings for DataVolumes 링크 복사링크가 클립보드에 복사되었습니다!
DataVolumes require a defined access mode and volume mode to be created in the web console. These storage settings are configured by default with a ReadWriteOnce access mode and Filesystem volume mode.
You can modify these settings by editing the kubevirt-storage-class-defaults ConfigMap in the openshift-cnv namespace. You can also add settings for other storage classes in order to create DataVolumes in the web console for different storage types.
You must configure storage settings that are supported by the underlying storage.
All DataVolumes that you create in the web console use the default storage settings unless you specifya storage class that is also defined in the ConfigMap.
6.12.6.1.1. Access modes 링크 복사링크가 클립보드에 복사되었습니다!
DataVolumes support the following access modes:
-
ReadWriteOnce: The volume can be mounted as read-write by a single node.ReadWriteOncehas greater versatility and is the default setting. -
ReadWriteMany: The volume can be mounted as read-write by many nodes.ReadWriteManyis required for some features, such as live migration of virtual machines between nodes.
ReadWriteMany is recommended if the underlying storage supports it.
6.12.6.1.2. Volume modes 링크 복사링크가 클립보드에 복사되었습니다!
The volume mode defines if a volume is intended to be used with a formatted filesystem or to remain in raw block state. DataVolumes support the following volume modes:
-
Filesystem: Creates a filesystem on the DataVolume. This is the default setting. -
Block: Creates a block DataVolume. Only useBlockif the underlying storage supports it.
6.12.6.2. Editing the kubevirt-storage-class-defaults ConfigMap in the web console 링크 복사링크가 클립보드에 복사되었습니다!
Modify the storage settings for DataVolumes by editing the kubevirt-storage-class-defaults ConfigMap in the openshift-cnv namespace. You can also add settings for other storage classes in order to create DataVolumes in the web console for different storage types.
You must configure storage settings that are supported by the underlying storage.
Procedure
- Click Workloads → Config Maps from the side menu.
- In the Project list, select openshift-cnv.
- Click kubevirt-storage-class-defaults to open the Config Map Overview.
- Click the YAML tab to display the editable configuration.
Update the
datavalues with the storage configuration that is appropriate for your underlying storage:Copy to Clipboard Copied! Toggle word wrap Toggle overflow - 1
- The default accessMode is
ReadWriteOnce. - 2
- The default volumeMode is
Filesystem. - 3
- If you add an access mode for a storage class, replace the
<new>part of the parameter with the storage class name. - 4
- If you add a volume mode for a storage class, replace the
<new>part of the parameter with the storage class name.
- Click Save to update the ConfigMap.
6.12.6.3. Editing the kubevirt-storage-class-defaults ConfigMap in the CLI 링크 복사링크가 클립보드에 복사되었습니다!
Modify the storage settings for DataVolumes by editing the kubevirt-storage-class-defaults ConfigMap in the openshift-cnv namespace. You can also add settings for other storage classes in order to create DataVolumes in the web console for different storage types.
You must configure storage settings that are supported by the underlying storage.
Procedure
Use
oc editto edit the ConfigMap:oc edit configmap kubevirt-storage-class-defaults -n openshift-cnv
$ oc edit configmap kubevirt-storage-class-defaults -n openshift-cnvCopy to Clipboard Copied! Toggle word wrap Toggle overflow Update the
datavalues of the ConfigMap:Copy to Clipboard Copied! Toggle word wrap Toggle overflow - 1
- The default accessMode is
ReadWriteOnce. - 2
- The default volumeMode is
Filesystem. - 3
- If you add an access mode for storage class, replace the
<new>part of the parameter with the storage class name. - 4
- If you add a volume mode for a storage class, replace the
<new>part of the parameter with the storage class name.
- Save and exit the editor to update the ConfigMap.
6.12.6.4. Example of multiple storage class defaults 링크 복사링크가 클립보드에 복사되었습니다!
The following YAML file is an example of a kubevirt-storage-class-defaults ConfigMap that has storage settings configured for two storage classes, migration and block.
Ensure that all settings are supported by your underlying storage before you update the ConfigMap.
6.12.7. Preparing CDI scratch space 링크 복사링크가 클립보드에 복사되었습니다!
6.12.7.1. About DataVolumes 링크 복사링크가 클립보드에 복사되었습니다!
DataVolume objects are custom resources that are provided by the Containerized Data Importer (CDI) project. DataVolumes orchestrate import, clone, and upload operations that are associated with an underlying PersistentVolumeClaim (PVC). DataVolumes are integrated with KubeVirt, and they prevent a virtual machine from being started before the PVC has been prepared.
6.12.7.2. Understanding scratch space 링크 복사링크가 클립보드에 복사되었습니다!
The Containerized Data Importer (CDI) requires scratch space (temporary storage) to complete some operations, such as importing and uploading virtual machine images. During this process, the CDI provisions a scratch space PVC equal to the size of the PVC backing the destination DataVolume (DV). The scratch space PVC is deleted after the operation completes or aborts.
The CDIConfig object allows you to define which StorageClass to use to bind the scratch space PVC by setting the scratchSpaceStorageClass in the spec: section of the CDIConfig object.
If the defined StorageClass does not match a StorageClass in the cluster, then the default StorageClass defined for the cluster is used. If there is no default StorageClass defined in the cluster, the StorageClass used to provision the original DV or PVC is used.
The CDI requires requesting scratch space with a file volume mode, regardless of the PVC backing the origin DataVolume. If the origin PVC is backed by block volume mode, you must define a StorageClass capable of provisioning file volume mode PVCs.
Manual provisioning
If there are no storage classes, the CDI will use any PVCs in the project that match the size requirements for the image. If there are no PVCs that match these requirements, the CDI import Pod will remain in a Pending state until an appropriate PVC is made available or until a timeout function kills the Pod.
6.12.7.3. CDI operations that require scratch space 링크 복사링크가 클립보드에 복사되었습니다!
| Type | Reason |
|---|---|
| Registry imports | The CDI must download the image to a scratch space and extract the layers to find the image file. The image file is then passed to QEMU-IMG for conversion to a raw disk. |
| Upload image | QEMU-IMG does not accept input from STDIN. Instead, the image to upload is saved in scratch space before it can be passed to QEMU-IMG for conversion. |
| HTTP imports of archived images | QEMU-IMG does not know how to handle the archive formats CDI supports. Instead, the image is unarchived and saved into scratch space before it is passed to QEMU-IMG. |
| HTTP imports of authenticated images | QEMU-IMG inadequately handles authentication. Instead, the image is saved to scratch space and authenticated before it is passed to QEMU-IMG. |
| HTTP imports of custom certificates | QEMU-IMG inadequately handles custom certificates of HTTPS endpoints. Instead, the CDI downloads the image to scratch space before passing the file to QEMU-IMG. |
6.12.7.4. Defining a StorageClass in the CDI configuration 링크 복사링크가 클립보드에 복사되었습니다!
Define a StorageClass in the CDI configuration to dynamically provision scratch space for CDI operations.
Procedure
Use the
occlient to edit thecdiconfig/configand add or edit thespec: scratchSpaceStorageClassto match a StorageClass in the cluster.oc edit cdiconfig/config
$ oc edit cdiconfig/configCopy to Clipboard Copied! Toggle word wrap Toggle overflow Copy to Clipboard Copied! Toggle word wrap Toggle overflow
6.12.7.5. CDI supported operations matrix 링크 복사링크가 클립보드에 복사되었습니다!
This matrix shows the supported CDI operations for content types against endpoints, and which of these operations requires scratch space.
| Content types | HTTP | HTTPS | HTTP basic auth | Registry | Upload |
|---|---|---|---|---|---|
| KubeVirt(QCOW2) |
✓ QCOW2 |
✓ QCOW2** |
✓ QCOW2 |
✓ QCOW2* |
✓ QCOW2* |
| KubeVirt (RAW) |
✓ RAW |
✓ RAW |
✓ RAW |
✓ RAW* |
✓ RAW* |
| Archive+ | ✓ TAR | ✓ TAR | ✓ TAR | □ TAR | □ TAR |
✓ Supported operation
□ Unsupported operation
* Requires scratch space
** Requires scratch space if a custom certificate authority is required
+ Archive does not support block mode DVs
Additional resources
- See the Dynamic provisioning section for more information on StorageClasses and how these are defined in the cluster.
Chapter 7. Virtual machine templates 링크 복사링크가 클립보드에 복사되었습니다!
7.1. Creating virtual machine templates 링크 복사링크가 클립보드에 복사되었습니다!
Using Virtual machines templates is an easy way to create multiple virtual machines with similar configuration. After a template is created, reference the template when creating virtual machines.
7.1.1. Creating a virtual machine template with the interactive wizard in the web console 링크 복사링크가 클립보드에 복사되었습니다!
The web console features an interactive wizard that guides you through the General, Networking, Storage, Advanced, and Review steps to simplify the process of creating virtual machine templates. All required fields are marked with a *. The wizard prevents you from moving to the next step until you provide values in the required fields.
Procedure
- In the container-native virtualization console, click Workloads → Virtual Machine Templates.
- Click Create Template and select New with Wizard.
- Fill in all required fields in the General step.
Click Next to progress to the Networking screen. A NIC that is named
nic0is attached by default.- Optional: Click Add Network Interface to create additional NICs.
-
Optional: You can remove any or all NICs by clicking the Options menu
and selecting Delete. Virtual machines created from a template do not need a NIC attached. NICs can be created after a virtual machine has been created.
Click Next to progress to the Storage screen.
- Optional: Click Add Disk to create additional disks.
- Optional: Click a disk to modify available fields. Click the ✓ button to save the changes.
Optional: Click Disk to choose an available disk from the Select Storage list.
NoteIf either URL or Container are selected as the Source in the General step, a
rootdiskdisk is created and attached to virtual machines as the Bootable Disk. You can modify therootdiskbut you cannot remove it.A Bootable Disk is not required for virtual machines provisioned from a PXE source if there are no disks attached to the virtual machine. If one or more disks are attached to the virtual machine, you must select one as the Bootable Disk.
Click Create Virtual Machine Template >. The Results screen displays the JSON configuration file for the virtual machine template.
The template is listed in Workloads → Virtual Machine Templates.
7.1.2. Virtual machine template interactive wizard fields 링크 복사링크가 클립보드에 복사되었습니다!
The following tables describe the fields for the Basic Settings, Networking, and Storage panes in the Create Virtual Machine Template interactive wizard.
7.1.2.1. Virtual machine template wizard fields 링크 복사링크가 클립보드에 복사되었습니다!
| Name | Parameter | Description |
|---|---|---|
| Source | PXE | Provision virtual machine from PXE menu. Requires a PXE-capable NIC in the cluster. |
| URL | Provision virtual machine from an image available from an HTTP or S3 endpoint. | |
| Container |
Provision virtual machine from a bootable operating system container located in a registry accessible from the cluster. Example: | |
| Disk | Provision virtual machine from a disk. | |
| Attach disk | Attach an existing disk that was previously cloned or created and made available in the PersistentVolumeClaims. When this option is selected, you must manually complete the Operating System, Flavor, and Workload Profile fields. | |
| Operating System | The primary operating system that is selected for the virtual machine. | |
| Flavor | small, medium, large, tiny, Custom | Presets that determine the amount of CPU and memory allocated to the virtual machine. The presets displayed for Flavor are determined by the operating system. |
| Workload Profile | High Performance | A virtual machine configuration that is optimized for high-performance workloads. |
| Server | A profile optimized to run server workloads. | |
| Desktop | A virtual machine configuration for use on a desktop. | |
| Name |
The name can contain lowercase letters ( | |
| Description | Optional description field. |
7.1.2.2. Cloud-init fields 링크 복사링크가 클립보드에 복사되었습니다!
| Name | Description |
|---|---|
| Hostname | Sets a specific host name for the virtual machine. |
| Authenticated SSH Keys | The user’s public key that is copied to ~/.ssh/authorized_keys on the virtual machine. |
| Use custom script | Replaces other options with a field in which you paste a custom cloud-init script. |
7.1.2.3. Networking fields 링크 복사링크가 클립보드에 복사되었습니다!
| Name | Description |
|---|---|
| Name | Name for the Network Interface Card. |
| Model | Driver for the Network Interface Card or model for the Network Interface Card. |
| Network | List of available NetworkAttachmentDefinition objects. |
| Type |
List of available binding methods. For the default Pod network, |
| MAC Address | MAC address for the Network Interface Card. If a MAC address is not specified, an ephemeral address is generated for the session. |
7.1.2.4. Storage fields 링크 복사링크가 클립보드에 복사되었습니다!
| Name | Description |
|---|---|
| Source | Select a blank disk for the virtual machine or choose from the options available: PXE, Container, URL or Disk. To select an existing disk and attach it to the virtual machine, choose Attach Disk from a list of available PersistentVolumeClaims (PVCs) or from a cloned disk. |
| Name |
Name of the disk. The name can contain lowercase letters ( |
| Size (GiB) | Size, in GiB, of the disk. |
| Interface | Name of the interface. |
| Storage class |
Name of the underlying |
7.2. Editing virtual machine templates 링크 복사링크가 클립보드에 복사되었습니다!
You can update a virtual machine template in the web console, either by editing the full configuration in the YAML editor or by editing a subset of the parameters in the Virtual Machine Template Overview screen.
7.2.1. Editing a virtual machine template in the web console 링크 복사링크가 클립보드에 복사되었습니다!
Edit select values of a virtual machine template in the Virtual Machine Template Overview screen of the web console by clicking on the pencil icon next to the relevant field. Other values can be edited using the CLI.
Procedure
- Click Workloads → Virtual Machine Templates from the side menu.
- Select a virtual machine template to open the Virtual Machine Template Overview screen.
- Click the pencil icon to make that field editable.
- Make the relevant changes and click Save.
Editing a virtual machine template will not affect virtual machines already created from that template.
7.2.2. Editing virtual machine template YAML configuration in the web console 링크 복사링크가 클립보드에 복사되었습니다!
You can edit the YAML configuration of a virtual machine template from the web console.
Not all parameters can be modified. If you click Save with an invalid configuration, an error message indicates the parameter that cannot be modified.
Navigating away from the YAML screen while editing cancels any changes to the configuration that you made.
Procedure
- In the container-native virtualization console, click Workloads → Virtual Machine Templates.
- Select a template.
- Click the YAML tab to display the editable configuration.
- Edit the file and click Save.
A confirmation message, which includes the updated version number for the object, shows the modification has been successful.
7.2.3. Adding a virtual disk to a virtual machine template 링크 복사링크가 클립보드에 복사되었습니다!
Use this procedure to add a virtual disk to a virtual machine template.
Procedure
- From the Virtual Machine Templates tab, select your virtual machine template.
- Select the Disks tab.
- Click Add Disks to open the Add Disk window.
- In the Add Disk window, specify Source, Name, Size, Interface, and Storage Class.
- Use the drop-down lists and check boxes to edit the disk configuration.
- Click OK.
7.2.4. Adding a network interface to a virtual machine template 링크 복사링크가 클립보드에 복사되었습니다!
Use this procedure to add a network interface to a virtual machine template.
Procedure
- From the Virtual Machine Templates tab, select the virtual machine template.
- Select the Network Interfaces tab.
- Click Add Network Interface.
- In the Add Network Interface window, specify the Name, Model, Network, Type, and MAC Address of the network interface.
- Click Add to add the network interface.
- Restart the virtual machine to enable access.
- Edit the drop-down lists and check boxes to configure the network interface.
- Click Save Changes.
- Click OK.
The new network interface displays at the top of the Create Network Interface list until the user restarts it.
The new network interface has a Pending VM restart Link State until you restart the virtual machine. Hover over the Link State to display more detailed information.
The Link State is set to Up by default when the network interface card is defined on the virtual machine and connected to the network.
7.2.5. Editing CD-ROMs for Virtual Machine Templates 링크 복사링크가 클립보드에 복사되었습니다!
Use the following procedure to configure CD-ROMs for virtual machines.
Procedure
- From the Virtual Machine Templates tab, select your virtual machine template.
- Select the Overview tab.
Click the pencil icon to the right of the CD-ROMs label to open the Edit CD-ROM window.
- If no CD-ROMs are available for editing, the CD label initially displays blank.
- If there are CD-ROMs available, you can eject the CD-ROM by clicking Eject CD-ROM or remove it by clicking -.
In the Edit CD-ROM window, do the following:
Select the type of CD-ROM configuration in the Media Type field.
- On Windows systems, Windows guest tools is attached by default in the Media Type field.
- Complete the required information for each Media Type.
- Click Add CD-ROM.
- When all CD-ROMs are added, click Save.
7.3. Deleting a virtual machine template 링크 복사링크가 클립보드에 복사되었습니다!
You can delete a virtual machine template in the web console.
7.3.1. Deleting a virtual machine template in the web console 링크 복사링크가 클립보드에 복사되었습니다!
Deleting a virtual machine template permanently removes it from the cluster.
Procedure
- In the container-native virtualization console, click Workloads → Virtual Machine Templates.
You can delete the virtual machine template from this pane, which makes it easier to perform actions on multiple templates in the one pane, or from the Virtual Machine Template Details pane where you can view comprehensive details of the selected template:
-
Click the Options menu
of the template to delete and select Delete Template.
- Click the template name to open the Virtual Machine Template Details pane and click Actions → Delete Template.
-
Click the Options menu
- In the confirmation pop-up window, click Delete to permanently delete the template.
Chapter 8. Live migration 링크 복사링크가 클립보드에 복사되었습니다!
8.1. Virtual machine live migration 링크 복사링크가 클립보드에 복사되었습니다!
8.1.1. Understanding live migration 링크 복사링크가 클립보드에 복사되었습니다!
Live migration is the process of moving a running virtual machine instance to another node in the cluster without interruption to the virtual workload or access. This can be a manual process, if you select a virtual machine instance to migrate to another node, or an automatic process, if the virtual machine instance has a LiveMigrate eviction strategy and the node on which it is running is placed into maintenance.
Virtual machines must have a PersistentVolumeClaim (PVC) with a shared ReadWriteMany (RWX) access mode to be live migrated.
Additional resources:
8.2. Live migration limits and timeouts 링크 복사링크가 클립보드에 복사되었습니다!
Live migration limits and timeouts are applied so that migration processes do not overwhelm the cluster. Configure these settings by editing the kubevirt-config configuration file.
8.2.1. Configuring live migration limits and timeouts 링크 복사링크가 클립보드에 복사되었습니다!
Configure live migration limits and timeouts for the cluster by adding updated key:value fields to the kubevirt-config configuration file, which is located in the openshift-cnv namespace.
Procedure
Edit the
kubevirt-configconfiguration file and add the necessary live migration parameters. The following example shows the default values:oc edit configmap kubevirt-config -n openshift-cnv
$ oc edit configmap kubevirt-config -n openshift-cnvCopy to Clipboard Copied! Toggle word wrap Toggle overflow Copy to Clipboard Copied! Toggle word wrap Toggle overflow
8.2.2. Cluster-wide live migration limits and timeouts 링크 복사링크가 클립보드에 복사되었습니다!
| Parameter | Description | Default |
|---|---|---|
|
| Number of migrations running in parallel in the cluster. | 5 |
|
| Maximum number of outbound migrations per node. | 2 |
|
| Bandwidth limit of each migration, in MiB/s. | 64Mi |
|
|
The migration will be canceled if it has not completed in this time, in seconds per GiB of memory. For example, a virtual machine instance with 6GiB memory will timeout if it has not completed migration in 4800 seconds. If the | 800 |
|
| The migration will be canceled if memory copy fails to make progress in this time, in seconds. | 150 |
8.3. Migrating a virtual machine instance to another node 링크 복사링크가 클립보드에 복사되었습니다!
Manually initiate a live migration of a virtual machine instance to another node using either the web console or the CLI.
8.3.1. Initiating live migration of a virtual machine instance in the web console 링크 복사링크가 클립보드에 복사되었습니다!
Migrate a running virtual machine instance to a different node in the cluster.
The Migrate Virtual Machine action is visible to all users but only admin users can initiate a virtual machine migration.
Procedure
- In the container-native virtualization console, click Workloads → Virtual Machines.
You can initiate the migration from this screen, which makes it easier to perform actions on multiple virtual machines in the one screen, or from the Virtual Machine Details screen where you can view comprehensive details of the selected virtual machine:
-
Click the Options menu
at the end of virtual machine and select Migrate Virtual Machine.
- Click the virtual machine name to open the Virtual Machine Details screen and click Actions → Migrate Virtual Machine.
-
Click the Options menu
- Click Migrate to migrate the virtual machine to another node.
8.3.2. Initiating live migration of a virtual machine instance in the CLI 링크 복사링크가 클립보드에 복사되었습니다!
Initiate a live migration of a running virtual machine instance by creating a VirtualMachineInstanceMigration object in the cluster and referencing the name of the virtual machine instance.
Procedure
Create a
VirtualMachineInstanceMigrationconfiguration file for the virtual machine instance to migrate. For example,vmi-migrate.yaml:Copy to Clipboard Copied! Toggle word wrap Toggle overflow Create the object in the cluster:
oc create -f vmi-migrate.yaml
$ oc create -f vmi-migrate.yamlCopy to Clipboard Copied! Toggle word wrap Toggle overflow
The VirtualMachineInstanceMigration object triggers a live migration of the virtual machine instance. This object exists in the cluster for as long as the virtual machine instance is running, unless manually deleted.
8.4. Monitoring live migration of a virtual machine instance 링크 복사링크가 클립보드에 복사되었습니다!
You can monitor the progress of a live migration of a virtual machine instance from either the web console or the CLI.
8.4.1. Monitoring live migration of a virtual machine instance in the web console 링크 복사링크가 클립보드에 복사되었습니다!
For the duration of the migration, the virtual machine has a status of Migrating. This status is displayed in the Virtual Machines list or in the Virtual Machine Details screen for the migrating virtual machine.
Procedure
- In the container-native virtualization console, click Workloads → Virtual Machines.
8.4.2. Monitoring live migration of a virtual machine instance in the CLI 링크 복사링크가 클립보드에 복사되었습니다!
The status of the virtual machine migration is stored in the Status component of the VirtualMachineInstance configuration.
Procedure
Use the
oc describecommand on the migrating virtual machine instance:oc describe vmi vmi-fedora
$ oc describe vmi vmi-fedoraCopy to Clipboard Copied! Toggle word wrap Toggle overflow Copy to Clipboard Copied! Toggle word wrap Toggle overflow
8.5. Cancelling the live migration of a virtual machine instance 링크 복사링크가 클립보드에 복사되었습니다!
Cancel the live migration so that the virtual machine instance remains on the original node.
You can cancel a live migration from either the web console or the CLI.
8.5.1. Cancelling live migration of a virtual machine instance in the web console 링크 복사링크가 클립보드에 복사되었습니다!
A live migration of the virtual machine instance can be cancelled using the Options menu
found on each virtual machine in the Workloads → Virtual Machines screen, or from the Actions menu on the Virtual Machine Details screen.
Procedure
- In the container-native virtualization console, click Workloads → Virtual Machines.
You can cancel the migration from this screen, which makes it easier to perform actions on multiple virtual machines in the one screen, or from the Virtual Machine Details screen where you can view comprehensive details of the selected virtual machine:
-
Click the Options menu
at the end of virtual machine and select Cancel Virtual Machine Migration.
- Click the virtual machine name to open the Virtual Machine Details screen and click Actions → Cancel Virtual Machine Migration.
-
Click the Options menu
- Click Cancel Migration to cancel the virtual machine live migration.
8.5.2. Cancelling live migration of a virtual machine instance in the CLI 링크 복사링크가 클립보드에 복사되었습니다!
Cancel the live migration of a virtual machine instance by deleting the VirtualMachineInstanceMigration object associated with the migration.
Procedure
Delete the
VirtualMachineInstanceMigrationobject that triggered the live migration,migration-jobin this example:oc delete vmim migration-job
$ oc delete vmim migration-jobCopy to Clipboard Copied! Toggle word wrap Toggle overflow
8.6. Configuring virtual machine eviction strategy 링크 복사링크가 클립보드에 복사되었습니다!
The LiveMigrate eviction strategy ensures that a virtual machine instance is not interrupted if the node is placed into maintenance or drained. Virtual machines instances with this eviction strategy will be live migrated to another node.
8.6.1. Configuring custom virtual machines with the LiveMigration eviction strategy 링크 복사링크가 클립보드에 복사되었습니다!
You only need to configure the LiveMigration eviction strategy on custom virtual machines. Common templates have this eviction strategy configured by default.
Procedure
Add the
evictionStrategy: LiveMigrateoption to thespecsection in the virtual machine configuration file. This example usesoc editto update the relevant snippet of theVirtualMachineconfiguration file:oc edit vm <custom-vm> -n <my-namespace>
$ oc edit vm <custom-vm> -n <my-namespace>Copy to Clipboard Copied! Toggle word wrap Toggle overflow Copy to Clipboard Copied! Toggle word wrap Toggle overflow Restart the virtual machine for the update to take effect:
virtctl restart <custom-vm> -n <my-namespace>
$ virtctl restart <custom-vm> -n <my-namespace>Copy to Clipboard Copied! Toggle word wrap Toggle overflow
Chapter 9. Node maintenance 링크 복사링크가 클립보드에 복사되었습니다!
9.1. Manually refreshing TLS certificates 링크 복사링크가 클립보드에 복사되었습니다!
The TLS certificates for container-native virtualization components are created at the time of installation and are valid for one year. You must manually refresh these certificates before they expire.
9.1.1. Refreshing TLS certificates 링크 복사링크가 클립보드에 복사되었습니다!
To refresh the TLS certificates for container-native virtualization, download and run the rotate-certs script. This script is available from the kubevirt/hyperconverged-cluster-operator repository on GitHub.
When refreshing the certificates, the following operations are impacted:
- Migrations are canceled
- Image uploads are canceled
- VNC and console connections are closed
Prerequisites
-
Ensure that you are logged in to the cluster as a user with
cluster-adminprivileges. The script uses your active session to the cluster to refresh certificates in theopenshift-cnvnamespace.
Procedure
Download the
rotate-certs.shscript from GitHub:curl -O https://raw.githubusercontent.com/kubevirt/hyperconverged-cluster-operator/master/tools/rotate-certs.sh
$ curl -O https://raw.githubusercontent.com/kubevirt/hyperconverged-cluster-operator/master/tools/rotate-certs.shCopy to Clipboard Copied! Toggle word wrap Toggle overflow Ensure the script is executable:
chmod +x rotate-certs.sh
$ chmod +x rotate-certs.shCopy to Clipboard Copied! Toggle word wrap Toggle overflow Run the script:
./rotate-certs.sh -n openshift-cnv
$ ./rotate-certs.sh -n openshift-cnvCopy to Clipboard Copied! Toggle word wrap Toggle overflow
The TLS certificates are refreshed and valid for one year.
9.2. Node maintenance mode 링크 복사링크가 클립보드에 복사되었습니다!
9.2.1. Understanding node maintenance mode 링크 복사링크가 클립보드에 복사되었습니다!
Placing a node into maintenance marks the node as unschedulable and drains all the virtual machines and pods from it. Virtual machine instances that have a LiveMigrate eviction strategy are live migrated to another node without loss of service. This eviction strategy is configured by default in virtual machine created from common templates but must be configured manually for custom virtual machines.
Virtual machine instances without an eviction strategy will be deleted on the node and recreated on another node.
Virtual machines must have a PersistentVolumeClaim (PVC) with a shared ReadWriteMany (RWX) access mode to be live migrated.
Additional resources:
9.3. Setting a node to maintenance mode 링크 복사링크가 클립보드에 복사되었습니다!
9.3.1. Understanding node maintenance mode 링크 복사링크가 클립보드에 복사되었습니다!
Placing a node into maintenance marks the node as unschedulable and drains all the virtual machines and pods from it. Virtual machine instances that have a LiveMigrate eviction strategy are live migrated to another node without loss of service. This eviction strategy is configured by default in virtual machine created from common templates but must be configured manually for custom virtual machines.
Virtual machine instances without an eviction strategy will be deleted on the node and recreated on another node.
Virtual machines must have a PersistentVolumeClaim (PVC) with a shared ReadWriteMany (RWX) access mode to be live migrated.
Place a node into maintenance from either the web console or the CLI.
9.3.2. Setting a node to maintenance mode in the web console 링크 복사링크가 클립보드에 복사되었습니다!
Set a node to maintenance mode using the Options menu
found on each node in the Compute → Nodes list, or using the Actions control of the Node Details screen.
Procedure
- In the container-native virtualization console, click Compute → Nodes.
You can set the node to maintenance from this screen, which makes it easier to perform actions on multiple nodes in the one screen or from the Node Details screen where you can view comprehensive details of the selected node:
-
Click the Options menu
at the end of the node and select Start Maintenance.
- Click the node name to open the Node Details screen and click Actions → Start Maintenance.
-
Click the Options menu
- Click Start Maintenance in the confirmation window.
The node will live migrate virtual machine instances that have the liveMigration eviction strategy, and the node is no longer schedulable. All other pods and virtual machines on the node are deleted and recreated on another node.
9.3.3. Setting a node to maintenance mode in the CLI 링크 복사링크가 클립보드에 복사되었습니다!
Set a node to maintenance mode by creating a NodeMaintenance Custom Resource (CR) object that references the node name and the reason for setting it to maintenance mode.
Procedure
Create the node maintenance CR configuration. This example uses a CR that is called
node02-maintenance.yaml:Copy to Clipboard Copied! Toggle word wrap Toggle overflow Create the
NodeMaintenanceobject in the cluster:oc apply -f <node02-maintenance.yaml>
$ oc apply -f <node02-maintenance.yaml>Copy to Clipboard Copied! Toggle word wrap Toggle overflow
The node live migrates virtual machine instances that have the liveMigration eviction strategy, and taint the node so that it is no longer schedulable. All other pods and virtual machines on the node are deleted and recreated on another node.
Additional resources:
9.4. Resuming a node from maintenance mode 링크 복사링크가 클립보드에 복사되었습니다!
Resuming a node brings it out of maintenance mode and schedulable again.
Resume a node from maintenance from either the web console or the CLI.
9.4.1. Resuming a node from maintenance mode in the web console 링크 복사링크가 클립보드에 복사되었습니다!
Resume a node from maintenance mode using the Options menu
found on each node in the Compute → Nodes list, or using the Actions control of the Node Details screen.
Procedure
- In the container-native virtualization console, click Compute → Nodes.
You can resume the node from this screen, which makes it easier to perform actions on multiple nodes in the one screen, or from the Node Details screen where you can view comprehensive details of the selected node:
-
Click the Options menu
at the end of the node and select Stop Maintenance.
- Click the node name to open the Node Details screen and click Actions → Stop Maintenance.
-
Click the Options menu
- Click Stop Maintenance in the confirmation window.
The node becomes schedulable, but virtual machine instances that were running on the node prior to maintenance will not automatically migrate back to this node.
9.4.2. Resuming a node from maintenance mode in the CLI 링크 복사링크가 클립보드에 복사되었습니다!
Resume a node from maintenance mode and make it schedulable again by deleting the NodeMaintenance object for the node.
Procedure
Find the
NodeMaintenanceobject:oc get nodemaintenance
$ oc get nodemaintenanceCopy to Clipboard Copied! Toggle word wrap Toggle overflow Optional: Insepct the
NodeMaintenanceobject to ensure it is associated with the correct node:oc describe nodemaintenance <node02-maintenance>
$ oc describe nodemaintenance <node02-maintenance>Copy to Clipboard Copied! Toggle word wrap Toggle overflow Copy to Clipboard Copied! Toggle word wrap Toggle overflow Delete the
NodeMaintenanceobject:oc delete nodemaintenance <node02-maintenance>
$ oc delete nodemaintenance <node02-maintenance>Copy to Clipboard Copied! Toggle word wrap Toggle overflow
Chapter 10. Logging, events, and monitoring 링크 복사링크가 클립보드에 복사되었습니다!
10.1. Viewing virtual machine logs 링크 복사링크가 클립보드에 복사되었습니다!
10.1.1. Understanding virtual machine logs 링크 복사링크가 클립보드에 복사되었습니다!
Logs are collected for OpenShift Container Platform Builds, Deployments, and Pods. In container-native virtualization, virtual machine logs can be retrieved from the virtual machine launcher Pod in either the web console or the CLI.
The -f option follows the log output in real time, which is useful for monitoring progress and error checking.
If the launcher Pod is failing to start, use the --previous option to see the logs of the last attempt.
ErrImagePull and ImagePullBackOff errors can be caused by an incorrect Deployment configuration or problems with the images that are referenced.
10.1.2. Viewing virtual machine logs in the CLI 링크 복사링크가 클립보드에 복사되었습니다!
Get virtual machine logs from the virtual machine launcher Pod.
Procedure
User the following command:
oc logs <virt-launcher-name>
$ oc logs <virt-launcher-name>Copy to Clipboard Copied! Toggle word wrap Toggle overflow
10.1.3. Viewing virtual machine logs in the web console 링크 복사링크가 클립보드에 복사되었습니다!
Get virtual machine logs from the associated virtual machine launcher Pod.
Procedure
- In the container-native virtualization console, click Workloads → Virtual Machines.
- Click the virtual machine to open the Virtual Machine Details panel.
-
In the Overview tab, click the
virt-launcher-<name>Pod in the POD section. - Click Logs.
10.2. Viewing events 링크 복사링크가 클립보드에 복사되었습니다!
10.2.1. Understanding virtual machine events 링크 복사링크가 클립보드에 복사되었습니다!
OpenShift Container Platform events are records of important life-cycle information in a namespace and are useful for monitoring and troubleshooting resource scheduling, creation, and deletion issues.
Container-native virtualization adds events for virtual machines and virtual machine instances. These can be viewed from either the web console or the CLI.
See also: Viewing system event information in an OpenShift Container Platform cluster.
10.2.2. Viewing the events for a virtual machine in the web console 링크 복사링크가 클립보드에 복사되었습니다!
You can view the stream events for a running a virtual machine from the Virtual Machine Details panel of the web console.
The ▮▮ button pauses the events stream.
The ▶ button continues a paused events stream.
Procedure
- Click Workloads → Virtual Machines from the side menu.
- Select a Virtual Machine.
- Click Events to view all events for the virtual machine.
10.2.3. Viewing namespace events in the CLI 링크 복사링크가 클립보드에 복사되었습니다!
Use the OpenShift Container Platform client to get the events for a namespace.
Procedure
In the namespace, use the
oc getcommand:oc get events
$ oc get eventsCopy to Clipboard Copied! Toggle word wrap Toggle overflow
10.2.4. Viewing resource events in the CLI 링크 복사링크가 클립보드에 복사되었습니다!
Events are included in the resource description, which you can get using the OpenShift Container Platform client.
Procedure
In the namespace, use the
oc describecommand. The following example shows how to get the events for a virtual machine, a virtual machine instance, and the virt-launcher Pod for a virtual machine:oc describe vm <vm> oc describe vmi <vmi> oc describe pod virt-launcher-<name>
$ oc describe vm <vm> $ oc describe vmi <vmi> $ oc describe pod virt-launcher-<name>Copy to Clipboard Copied! Toggle word wrap Toggle overflow
10.3. Viewing information about virtual machine workloads 링크 복사링크가 클립보드에 복사되었습니다!
You can view high-level information about your virtual machines by using the Virtual Machines dashboard in the OpenShift Container Platform web console.
10.3.1. About the Virtual Machines dashboard 링크 복사링크가 클립보드에 복사되었습니다!
Access the Virtual Machines dashboard from the OpenShift Container Platform web console by navigating to the Workloads → Virtual Machines page and selecting a virtual machine.
The dashboard includes the following cards:
Details provides identifying information about the virtual machine, including:
- Name
- Namespace
- Date of creation
- Node name
- IP address
Inventory lists the virtual machine’s resources, including:
- Network interface controllers (NICs)
- Disks
Status includes:
- The current status of the virtual machine
- A note indicating whether or not the QEMU guest agent is installed on the virtual machine
Utilization includes charts that display usage data for:
- CPU
- Memory
- Filesystem
- Network transfer
Use the drop-down list to choose a duration for the utilization data. The available options are 1 Hour, 6 Hours, and 24 Hours.
- Events lists messages about virtual machine activity over the past hour. To view additional events, click View all.
10.4. Monitoring virtual machine health 링크 복사링크가 클립보드에 복사되었습니다!
Use this procedure to create liveness and readiness probes to monitor virtual machine health.
10.4.1. About liveness and readiness probes 링크 복사링크가 클립보드에 복사되었습니다!
When a VirtualMachineInstance (VMI) fails, liveness probes stop the VMI. Controllers such as VirtualMachine then spawn other VMIs, restoring virtual machine responsiveness.
Readiness probes tell services and endpoints that the VirtualMachineInstance is ready to receive traffic from services. If readiness probes fail, the VirtualMachineInstance is removed from applicable endpoints until the probe recovers.
10.4.2. Define an HTTP liveness probe 링크 복사링크가 클립보드에 복사되었습니다!
This procedure provides an example configuration file for defining HTTP liveness probes.
Procedure
Customize a YAML configuration file to create an HTTP liveness probe, using the following code block as an example. In this example:
-
You configure a probe using
spec.livenessProbe.httpGet, which queries port1500of the VirtualMachineInstance, after an initial delay of120seconds. The VirtualMachineInstance installs and runs a minimal HTTP server on port
1500usingcloud-init.Copy to Clipboard Copied! Toggle word wrap Toggle overflow
-
You configure a probe using
Create the VirtualMachineInstance by running the following command:
oc create -f <file name>.yaml
$ oc create -f <file name>.yamlCopy to Clipboard Copied! Toggle word wrap Toggle overflow
10.4.3. Define a TCP liveness probe 링크 복사링크가 클립보드에 복사되었습니다!
This procedure provides an example configuration file for defining TCP liveness probes.
Procedure
Customize a YAML configuration file to create an TCP liveness probe, using this code block as an example. In this example:
-
You configure a probe using
spec.livenessProbe.tcpSocket, which queries port1500of the VirtualMachineInstance, after an initial delay of120seconds. The VirtualMachineInstance installs and runs a minimal HTTP server on port
1500usingcloud-init.Copy to Clipboard Copied! Toggle word wrap Toggle overflow
-
You configure a probe using
Create the VirtualMachineInstance by running the following command:
oc create -f <file name>.yaml
$ oc create -f <file name>.yamlCopy to Clipboard Copied! Toggle word wrap Toggle overflow
10.4.4. Define a readiness probe 링크 복사링크가 클립보드에 복사되었습니다!
This procedure provides an example configuration file for defining readiness probes.
Procedure
Customize a YAML configuration file to create a readiness probe. Readiness probes are configured in a similar manner to liveness probes. However, note the following differences in this example:
-
Readiness probes are saved using a different spec name. For example, you create a readiness probe as
spec.readinessProbeinstead of asspec.livenessProbe.<type-of-probe>. When creating a readiness probe, you optionally set a
failureThresholdand asuccessThresholdto switch betweenreadyandnon-readystates, should the probe succeed or fail multiple times.Copy to Clipboard Copied! Toggle word wrap Toggle overflow
-
Readiness probes are saved using a different spec name. For example, you create a readiness probe as
Create the VirtualMachineInstance by running the following command:
oc create -f <file name>.yaml
$ oc create -f <file name>.yamlCopy to Clipboard Copied! Toggle word wrap Toggle overflow
10.5. Using the OpenShift Container Platform dashboard to get cluster information 링크 복사링크가 클립보드에 복사되었습니다!
Access the OpenShift Container Platform dashboard, which captures high-level information about the cluster, by clicking Home > Dashboards > Overview from the OpenShift Container Platform web console.
The OpenShift Container Platform dashboard provides various cluster information, captured in individual dashboard cards.
10.5.1. About the OpenShift Container Platform dashboards page 링크 복사링크가 클립보드에 복사되었습니다!
The OpenShift Container Platform dashboard consists of the following cards:
Details provides a brief overview of informational cluster details.
Status include ok, error, warning, in progress, and unknown. Resources can add custom status names.
- Cluster ID
- Provider
- Version
Cluster Inventory details number of resources and associated statuses. It is helpful when intervention is required to resolve problems, including information about:
- Number of nodes
- Number of Pods
- Persistent storage volume claims
- Virtual machines (available if container-native virtualization is installed)
- Bare metal hosts in the cluster, listed according to their state (only available in metal3 environment).
- Cluster Health summarizes the current health of the cluster as a whole, including relevant alerts and descriptions. If container-native virtualization is installed, the overall health of container-native virtualization is diagnosed as well. If more than one subsystem is present, click See All to view the status of each subsystem.
Cluster Capacity charts help administrators understand when additional resources are required in the cluster. The charts contain an inner ring that displays current consumption, while an outer ring displays thresholds configured for the resource, including information about:
- CPU time
- Memory allocation
- Storage consumed
- Network resources consumed
- Cluster Utilization shows the capacity of various resources over a specified period of time, to help administrators understand the scale and frequency of high resource consumption.
- Events lists messages related to recent activity in the cluster, such as Pod creation or virtual machine migration to another host.
- Top Consumers helps administrators understand how cluster resources are consumed. Click on a resource to jump to a detailed page listing Pods and nodes that consume the largest amount of the specified cluster resource (CPU, memory, or storage).
10.6. OpenShift Container Platform cluster monitoring, logging, and Telemetry 링크 복사링크가 클립보드에 복사되었습니다!
OpenShift Container Platform provides various resources for monitoring at the cluster level.
10.6.1. About OpenShift Container Platform cluster monitoring 링크 복사링크가 클립보드에 복사되었습니다!
OpenShift Container Platform includes a pre-configured, pre-installed, and self-updating monitoring stack that is based on the Prometheus open source project and its wider eco-system. It provides monitoring of cluster components and includes a set of alerts to immediately notify the cluster administrator about any occurring problems and a set of Grafana dashboards. The cluster monitoring stack is only supported for monitoring OpenShift Container Platform clusters.
To ensure compatibility with future OpenShift Container Platform updates, configuring only the specified monitoring stack options is supported.
10.6.2. Cluster logging components 링크 복사링크가 클립보드에 복사되었습니다!
The cluster logging components are based upon Elasticsearch, Fluentd, and Kibana (EFK). The collector, Fluentd, is deployed to each node in the OpenShift Container Platform cluster. It collects all node and container logs and writes them to Elasticsearch (ES). Kibana is the centralized, web UI where users and administrators can create rich visualizations and dashboards with the aggregated data.
There are currently 5 different types of cluster logging components:
- logStore - This is where the logs will be stored. The current implementation is Elasticsearch.
- collection - This is the component that collects logs from the node, formats them, and stores them in the logStore. The current implementation is Fluentd.
- visualization - This is the UI component used to view logs, graphs, charts, and so forth. The current implementation is Kibana.
- curation - This is the component that trims logs by age. The current implementation is Curator.
For more information on cluster logging, see the OpenShift Container Platform cluster logging documentation.
10.6.3. About Telemetry 링크 복사링크가 클립보드에 복사되었습니다!
Telemetry sends a carefully chosen subset of the cluster monitoring metrics to Red Hat. These metrics are sent continuously and describe:
- The size of an OpenShift Container Platform cluster
- The health and status of OpenShift Container Platform components
- The health and status of any upgrade being performed
- Limited usage information about OpenShift Container Platform components and features
- Summary info about alerts reported by the cluster monitoring component
This continuous stream of data is used by Red Hat to monitor the health of clusters in real time and to react as necessary to problems that impact our customers. It also allows Red Hat to roll out OpenShift Container Platform upgrades to customers so as to minimize service impact and continuously improve the upgrade experience.
This debugging information is available to Red Hat Support and engineering teams with the same restrictions as accessing data reported via support cases. All connected cluster information is used by Red Hat to help make OpenShift Container Platform better and more intuitive to use. None of the information is shared with third parties.
10.6.3.1. Information collected by Telemetry 링크 복사링크가 클립보드에 복사되었습니다!
Primary information collected by Telemetry includes:
- The number of updates available per cluster
- Channel and image repository used for an update
- The number of errors that occurred during an update
- Progress information of running updates
- The number of machines per cluster
- The number of CPU cores and size of RAM of the machines
- The number of members in the etcd cluster and number of objects currently stored in the etcd cluster
- The number of CPU cores and RAM used per machine type - infra or master
- The number of CPU cores and RAM used per cluster
- The number of running virtual machine instances in the cluster
- Use of OpenShift Container Platform framework components per cluster
- The version of the OpenShift Container Platform cluster
- Health, condition, and status for any OpenShift Container Platform framework component that is installed on the cluster, for example Cluster Version Operator, Cluster Monitoring, Image Registry, and Elasticsearch for Logging
- A unique random identifier that is generated during installation
- The name of the platform that OpenShift Container Platform is deployed on, such as Amazon Web Services
Telemetry does not collect identifying information such as user names, passwords, or the names or addresses of user resources.
10.6.4. CLI troubleshooting and debugging commands 링크 복사링크가 클립보드에 복사되었습니다!
For a list of the oc client troubleshooting and debugging commands, see the OpenShift Container Platform CLI tools documentation.
10.7. Collecting container-native virtualization data for Red Hat Support 링크 복사링크가 클립보드에 복사되었습니다!
When opening a support case, it is helpful to provide debugging information about your cluster to Red Hat Support.
The must-gather tool enables you to collect diagnostic information about your OpenShift Container Platform cluster, including virtual machines and other data related to container-native virtualization.
For prompt support, supply diagnostic information for both OpenShift Container Platform and container-native virtualization.
Container-native virtualization is a Technology Preview feature only. Technology Preview features are not supported with Red Hat production service level agreements (SLAs) and might not be functionally complete. Red Hat does not recommend using them in production. These features provide early access to upcoming product features, enabling customers to test functionality and provide feedback during the development process.
For more information about the support scope of Red Hat Technology Preview features, see https://access.redhat.com/support/offerings/techpreview/.
10.7.1. About the must-gather tool 링크 복사링크가 클립보드에 복사되었습니다!
The oc adm must-gather CLI command collects the information from your cluster that is most likely needed for debugging issues, such as:
- Resource definitions
- Audit logs
- Service logs
You can specify one or more images when you run the command by including the --image argument. When you specify an image, the tool collects data related to that feature or product.
When you run oc adm must-gather, a new Pod is created on the cluster. The data is collected on that Pod and saved in a new directory that starts with must-gather.local. This directory is created in the current working directory.
10.7.2. About collecting container-native virtualization data 링크 복사링크가 클립보드에 복사되었습니다!
You can use the oc adm must-gather CLI command to collect information about your cluster, including features and objects associated with container-native virtualization:
- The Hyperconverged Cluster Operator namespaces (and child objects)
- All namespaces (and their child objects) that belong to any container-native virtualization resources
- All container-native virtualization Custom Resource Definitions (CRDs)
- All namespaces that contain virtual machines
- All virtual machine definitions
To collect container-native virtualization data with must-gather, you must specify the container-native virtualization image: --image=registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel8.
10.7.3. Gathering data about specific features 링크 복사링크가 클립보드에 복사되었습니다!
You can gather debugging information about specific features by using the oc adm must-gather CLI command with the --image or --image-stream argument. The must-gather tool supports multiple images, so you can gather data about more than one feature by running a single command.
Prerequisites
-
Access to the cluster as a user with the
cluster-adminrole. -
The OpenShift Container Platform CLI (
oc) installed.
Procedure
-
Navigate to the directory where you want to store the
must-gatherdata. Run the
oc adm must-gathercommand with one or more--imageor--image-streamarguments. For example, the following command gathers both the default cluster data and information specific to container-native virtualization:oc adm must-gather \ --image-stream=openshift/must-gather \ --image=registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel8
$ oc adm must-gather \ --image-stream=openshift/must-gather \1 --image=registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel82 Copy to Clipboard Copied! Toggle word wrap Toggle overflow Create a compressed file from the
must-gatherdirectory that was just created in your working directory. For example, on a computer that uses a Linux operating system, run the following command:tar cvaf must-gather.tar.gz must-gather.local.5421342344627712289/
$ tar cvaf must-gather.tar.gz must-gather.local.5421342344627712289/1 Copy to Clipboard Copied! Toggle word wrap Toggle overflow - 1
- Make sure to replace
must-gather-local.5421342344627712289/with the actual directory name.
- Attach the compressed file to your support case on the Red Hat Customer Portal.
Legal Notice
링크 복사링크가 클립보드에 복사되었습니다!
Copyright © 2025 Red Hat
OpenShift documentation is licensed under the Apache License 2.0 (https://www.apache.org/licenses/LICENSE-2.0).
Modified versions must remove all Red Hat trademarks.
Portions adapted from https://github.com/kubernetes-incubator/service-catalog/ with modifications by Red Hat.
Red Hat, Red Hat Enterprise Linux, the Red Hat logo, the Shadowman logo, JBoss, OpenShift, Fedora, the Infinity logo, and RHCE are trademarks of Red Hat, Inc., registered in the United States and other countries.
Linux® is the registered trademark of Linus Torvalds in the United States and other countries.
Java® is a registered trademark of Oracle and/or its affiliates.
XFS® is a trademark of Silicon Graphics International Corp. or its subsidiaries in the United States and/or other countries.
MySQL® is a registered trademark of MySQL AB in the United States, the European Union and other countries.
Node.js® is an official trademark of Joyent. Red Hat Software Collections is not formally related to or endorsed by the official Joyent Node.js open source or commercial project.
The OpenStack® Word Mark and OpenStack logo are either registered trademarks/service marks or trademarks/service marks of the OpenStack Foundation, in the United States and other countries and are used with the OpenStack Foundation’s permission. We are not affiliated with, endorsed or sponsored by the OpenStack Foundation, or the OpenStack community.
All other trademarks are the property of their respective owners.