Release Notes
Red Hat Virtualization 4.1
Release notes for Red Hat Virtualization 4.1
Abstract
The Release Notes provide high-level coverage of the improvements and additions that have been implemented in Red Hat Virtualization 4.1.
Chapter 1. Introduction Copy linkLink copied to clipboard!
Copy linkLink copied to clipboard!
1.1. Introduction to Red Hat Virtualization Copy linkLink copied to clipboard!
Copy linkLink copied to clipboard!
Red Hat Virtualization is an enterprise-grade server and desktop virtualization platform built on Red Hat Enterprise Linux. There are three core components to a Red Hat Virtualization environment:
- Red Hat Virtualization Manager
- A server that manages and provides access to the resources in the environment. This server can be a physical machine or a virtual machine. Moreover, the Manager can be set up as a virtual machine on a host in an environment that the Manager itself manages. This setup is known as a self-hosted engine setup.
- Storage
- Storage is used to store the data associated with virtual machines in the environment such as the virtual machines themselves and ISO files that can be used to install operating systems or other software on those virtual machines. Storage can be implemented on several storage types, including NFS, iSCSI, GlusterFS, and Fibre Channel.
- Hosts
- Hosts are servers that provide the processing capabilities and memory resources used to run virtual machines. In Red Hat Virtualization, servers running Red Hat Enterprise Linux 7 or a minimal operating system specifically designed to operate as a host in Red Hat Virtualization environments, known as the Red Hat Virtualization Host, can be used as hosts.
In addition to these core components, Red Hat Virtualization allows you to set up supporting components such as a scheduler and remote servers for hosting the databases used in the environment, and to take advantage of resources in Red Hat Enterprise Linux OpenStack Platform and VMware environments.
1.2. Subscriptions Copy linkLink copied to clipboard!
Copy linkLink copied to clipboard!
To install the Red Hat Virtualization Manager and hosts, your systems must be registered with the Content Delivery Network using Red Hat Subscription Management. This section outlines the entitlements and repositories required to set up a Red Hat Virtualization environment.
Important
Red Hat is transitioning the RHN-hosted interface into the Red Hat Subscription Management (RHSM) interfaces by July 31, 2017. If your current systems are registered to RHN Classic, see Migrating from RHN Classic to Red Hat Subscription Management (RHSM) for Red Hat Virtualization for instructions on how to migrate your systems to RHSM.
1.2.1. Required Entitlements and Repositories Copy linkLink copied to clipboard!
Copy linkLink copied to clipboard!
The packages provided in the following repositories are required to install and configure a functioning Red Hat Virtualization environment. When one of these repositories is required to install a package, the steps required to enable the repository are provided in the appropriate location in the Installation Guide or Self-Hosted Engine Guide.
|
Subscription Pool
|
Repository Name
|
Repository Label
|
Details
|
|---|---|---|---|
Red Hat Enterprise Linux Server
| Red Hat Enterprise Linux Server
| rhel-7-server-rpms
|
Provides the Red Hat Enterprise Linux 7 Server.
|
Red Hat Enterprise Linux Server
| RHEL Server Supplementary
| rhel-7-server-supplementary-rpms
|
Provides the virtio-win package, which provides the Windows VirtIO drivers for use in virtual machines.
|
Red Hat Virtualization
| Red Hat Virtualization
| rhel-7-server-rhv-4.1-rpms |
Provides the Red Hat Virtualization Manager.
|
Red Hat Virtualization
| Red Hat Virtualization Tools
| rhel-7-server-rhv-4-tools-rpms
|
Provides dependencies for the the Red Hat Virtualization Manager that are common to all Red Hat Virtualization 4 releases.
|
Red Hat Virtualization
| Red Hat JBoss Enterprise Application Platform
| jb-eap-7.1-for-rhel-7-server-rpms
| Provides the supported release of Red Hat JBoss Enterprise Application Platform on which the Manager runs. |
Red Hat Ansible Engine
| Red Hat Ansible Engine
| rhel-7-server-ansible-2-rpms
| Provides Red Hat Ansible Engine. |
|
Subscription Pool
|
Repository Name
|
Repository Label
|
Details
|
|---|---|---|---|
Red Hat Virtualization
| Red Hat Virtualization Host
| rhel-7-server-rhvh-4-rpms
|
Provides the rhev-hypervisor7-ng-image-update package, which allows you to update the image installed on the host.
|
|
Subscription Pool
|
Repository Name
|
Repository Label
|
Details
|
|---|---|---|---|
Red Hat Enterprise Linux Server
| Red Hat Enterprise Linux Server
| rhel-7-server-rpms
|
Provides the Red Hat Enterprise Linux 7 Server.
|
Red Hat Virtualization
| Red Hat Virtualization Management Agents (RPMs)
| rhel-7-server-rhv-4-mgmt-agent-rpms
|
Provides the QEMU and KVM packages required for using Red Hat Enterprise Linux 7 servers as virtualization hosts.
|
1.2.2. Optional Entitlements and Repositories Copy linkLink copied to clipboard!
Copy linkLink copied to clipboard!
The packages provided in the following repositories are not required to install and configure a functioning Red Hat Virtualization environment. However, they are required to install packages that provide supporting functionality on virtual machines and client systems such as virtual machine resource monitoring. When one of these repositories is required to install a package, the steps required to enable the repository are provided in the appropriate location in the Installation Guide or Self-Hosted Engine Guide.
|
Subscription Pool
|
Repository Name
|
Repository Label
|
Details
|
|---|---|---|---|
Red Hat Enterprise Linux Server
| Red Hat Enterprise Linux 7 Server - RH Common (v.7 Server for x86_64)
| rhel-7-server-rh-common-rpms
| Provides the rhevm-guest-agent-common package for Red Hat Enterprise Linux 7, which allows you to monitor virtual machine resources on Red Hat Enterprise Linux 7 clients. |
Red Hat Enterprise Linux Server
| Red Hat Enterprise Virt Agent (v.6 Server for x86_64)
| rhel-6-server-rhv-4-agent-rpms
| Provides the rhevm-guest-agent-common package for Red Hat Enterprise Linux 6, which allows you to monitor virtual machine resources on Red Hat Enterprise Linux 6 clients. |
Red Hat Enterprise Linux Server
| Red Hat Enterprise Virt Agent (v.5 Server for x86_64)
| rhel-5-server-rhv-4-agent-rpms
| Provides the rhev-guest-agent package for Red Hat Enterprise Linux 5, which allows you to monitor virtual machine resources on Red Hat Enterprise Linux 5 clients. |
Red Hat Virtualization
| Red Hat Virtualization Host Build
| rhel-7-server-rhvh-4-build-rpms
|
Provides packages used to build your own version of the Red Hat Virtualization Host image.
|
Chapter 2. RHV for IBM Power Copy linkLink copied to clipboard!
Copy linkLink copied to clipboard!
This release supports Red Hat Enterprise Linux 7 hosts on IBM POWER8 little endian hardware, and Red Hat Enterprise Linux 7 virtual machines on emulated IBM POWER8 hardware.
Important
Previous releases of RHV for IBM Power required Red Hat Enterprise Linux hosts on POWER8 hardware to be installed from an ISO image. These hosts cannot be updated for use with this release. You must reinstall Red Hat Enterprise Linux 7 hosts using the repositories outlined below.
The packages provided in the following repositories are required to install and configure aspects of a Red Hat Virtualization environment on POWER8 hardware.
|
Component
|
Subscription Pool
|
Repository Name
|
Repository Label
|
Details
|
|---|---|---|---|---|
|
Red Hat Virtualization Manager
| Red Hat Virtualization for IBM Power
| Red Hat Virtualization for IBM Power
| rhel-7-server-rhv-4-power-rpms
|
Provides the Red Hat Virtualization Manager for use with IBM POWER8 hosts. The Manager itself must be installed on x86_64 architecture.
|
|
Red Hat Enterprise Linux 7 hosts, little endian
| Red Hat Enterprise Linux for Power, little endian
| RHV Management Agent for IBM Power, little endian
| rhel-7-server-rhv-4-mgmt-agent-for-power-le-rpms |
Provides the QEMU and KVM packages required for using Red Hat Enterprise Linux 7 servers on IBM Power (little endian) hardware as virtualization hosts.
|
| | Red Hat Enterprise Linux for Power, little endian
| Red Hat Enterprise Linux for IBM Power, little endian
| rhel-7-for-power-le-rpms
|
Provides additional packages required for using Red Hat Enterprise Linux 7 servers on IBM Power (little endian) hardware as virtualization hosts.
|
|
Red Hat Enterprise Linux 7 virtual machines, big endian
| Red Hat Enterprise Linux for Power, big endian
| RHV Tools for IBM Power
| rhel-7-server-rhv-4-tools-for-power-le-rpms |
Provides the rhevm-guest-agent-common package for Red Hat Enterprise Linux 7 virtual machines on emulated IBM Power (big endian) hardware. The guest agents allow you to monitor virtual machine resources on Red Hat Enterprise Linux 7 clients.
|
|
Red Hat Enterprise Linux 7 virtual machines, little endian
| Red Hat Enterprise Linux for Power, little endian
| RHV Tools for IBM Power, little endian
| rhel-7-server-rhv-4-tools-for-power-le-rpms |
Provides the rhevm-guest-agent-common package for Red Hat Enterprise Linux 7 virtual machines on emulated IBM Power (little endian) hardware. The guest agents allow you to monitor virtual machine resources on Red Hat Enterprise Linux 7 clients.
|
Unsupported Features
The following Red Hat Virtualization features are not supported:
- Hot-plug CPU
- SPICE display
- SmartCard
- Sound device
- Guest SSO
- Integration with OpenStack Networking (Neutron), OpenStack Image (Glance), and OpenStack Volume (Cinder)
- Self-hosted engine
- Red Hat Virtualization Host (RHVH)
- Disk Block Alignment
For a full list of bugs that affect the RHV for IBM Power release, see https://bugzilla.redhat.com/show_bug.cgi?id=1359843.
Chapter 3. Release Information Copy linkLink copied to clipboard!
Copy linkLink copied to clipboard!
These release notes highlight technology preview items, recommended practices, known issues, and deprecated functionality to be taken into consideration when deploying this release of Red Hat Virtualization.
Notes for updates released during the support lifecycle of this Red Hat Virtualization release will appear in the advisory text associated with each update or the Red Hat Virtualization Technical Notes. This document is available from the following page:
3.1. Red Hat Virtualization 4.1 GA Copy linkLink copied to clipboard!
Copy linkLink copied to clipboard!
3.1.1. Enhancements Copy linkLink copied to clipboard!
Copy linkLink copied to clipboard!
This release of Red Hat Virtualization features the following enhancements:
- BZ#734120
With this update, the ability to sparsify a thin-provisioned disk has been added to Red Hat Virtualization. When a virtual machine is shut down, the user can sparsify the disk to convert the free space within the disk image back to free space on the host.
With this update, the ability to sparsify a thin-provisioned disk has been added to Red Hat Virtualization. When a virtual machine is shut down, the user can sparsify the disk to convert the free space within the disk image back to free space on the host.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#804272
Previously, if power management fencing was not available, automatic high availability of virtual machines did not work without manual confirmation that the host the virtual machine was running on was rebooted. In this release, a new option for virtual machine lease on storage domains enables automatic high availability failover of a virtual machine, when the host running the machine is down, without the availability of power management for the host.
Previously, if power management fencing was not available, automatic high availability of virtual machines did not work without manual confirmation that the host the virtual machine was running on was rebooted. In this release, a new option for virtual machine lease on storage domains enables automatic high availability failover of a virtual machine, when the host running the machine is down, without the availability of power management for the host.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#827529
This release introduces QCOW2 v3, which has a compatibility level of 1.1. This enables the QEMU to use this volume in a more efficient way, with its improved performance capabilities. In addition, it is fully backwards-compatible with the QCOW2 feature set, is easy to upgrade from QCOW2 v2 to QCOW2 v3, and supports extensibility.
This release introduces QCOW2 v3, which has a compatibility level of 1.1. This enables the QEMU to use this volume in a more efficient way, with its improved performance capabilities. In addition, it is fully backwards-compatible with the QCOW2 feature set, is easy to upgrade from QCOW2 v2 to QCOW2 v3, and supports extensibility.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#868811
This enhancement allows live migration of virtual machines that have SR-IOV NICs attached. Before the migration all SR-IOV NICs are hot unplugged, and after successful migration they are hot plugged back.
This enhancement allows live migration of virtual machines that have SR-IOV NICs attached. Before the migration all SR-IOV NICs are hot unplugged, and after successful migration they are hot plugged back.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1001181
This release provides a cleanup script for completely cleaning the host after a failed attempt to install a self-hosted engine.
This release provides a cleanup script for completely cleaning the host after a failed attempt to install a self-hosted engine.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1024063
Previously, it was impossible to reboot a host without enabling Power Management. In this release, it is now possible to shut down and reboot a host without using Power Management. From the Management menu, a new option called SSH Management is available, enabling administrators to select either Restart or Stop.
Previously, it was impossible to reboot a host without enabling Power Management. In this release, it is now possible to shut down and reboot a host without using Power Management. From the Management menu, a new option called SSH Management is available, enabling administrators to select either Restart or Stop.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1036221
With this update, if the web console, either noVnc or SPICE HTML5, is unable to connect to the websocket proxy server, a pop-up displays with troubleshooting suggestions. The pop-up contains a link to the default CA certificate.
With this update, if the web console, either noVnc or SPICE HTML5, is unable to connect to the websocket proxy server, a pop-up displays with troubleshooting suggestions. The pop-up contains a link to the default CA certificate.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1080379
Previously, almost all data path operations on the hosts could only be performed on the elected Storage Pool Manager. This could potentially cause bottlenecks. In this release, a new lightweight host jobs management infrastructure was introduced, which enables data path operations to run on any host. In addition, administrators can monitor the progress of Move Disk operations with the aid of a progress indicator.
Previously, almost all data path operations on the hosts could only be performed on the elected Storage Pool Manager. This could potentially cause bottlenecks. In this release, a new lightweight host jobs management infrastructure was introduced, which enables data path operations to run on any host. In addition, administrators can monitor the progress of Move Disk operations with the aid of a progress indicator.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1081536
With this release, when creating virtual machine pools using a template that is present in more than one storage domain, virtual machine disks can be distributed to multiple storage domains by selecting "Auto select target" in New Pool -> Resource Allocation -> Disk Allocation.
With this release, when creating virtual machine pools using a template that is present in more than one storage domain, virtual machine disks can be distributed to multiple storage domains by selecting "Auto select target" in New Pool -> Resource Allocation -> Disk Allocation.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1092907
Previously, when notification emails were successfully sent to a configured SMTP server, a success message did not appear in the notifier.log file. In this release, when a message is successfully sent to an SMTP server, the following message appears in the notifier.log file: E-mail subject='...' to='...' sent successfully
Previously, when notification emails were successfully sent to a configured SMTP server, a success message did not appear in the notifier.log file. In this release, when a message is successfully sent to an SMTP server, the following message appears in the notifier.log file: E-mail subject='...' to='...' sent successfullyCopy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1097589
This release adds support for CPU hot unplug to Red Hat Virtualization. Note that the guest operating system must also support the feature, and only previously hot plugged CPUs can be hot unplugged.
This release adds support for CPU hot unplug to Red Hat Virtualization. Note that the guest operating system must also support the feature, and only previously hot plugged CPUs can be hot unplugged.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1101554
With this update, the code interfacing with VDSM now uses the VDSM API directly instead of using vdsClient and xmlrpc.
With this update, the code interfacing with VDSM now uses the VDSM API directly instead of using vdsClient and xmlrpc.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1135976
Previously, CPU pinning information could be silently lost. Now, a pop-up warning, that CPU pinning information will be lost, displays when a user saves a virtual machine. This means that the user is aware that CPU pinning information will be lost and has the choice to cancel the operation.
Previously, CPU pinning information could be silently lost. Now, a pop-up warning, that CPU pinning information will be lost, displays when a user saves a virtual machine. This means that the user is aware that CPU pinning information will be lost and has the choice to cancel the operation.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1141422
With this update, the VDSM thread name is now included in the system monitoring tools. This makes it easier to track the resource usage of the threads.
With this update, the VDSM thread name is now included in the system monitoring tools. This makes it easier to track the resource usage of the threads.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1145240
Power Management alerts are now disabled when fencing is disabled in a cluster.
Power Management alerts are now disabled when fencing is disabled in a cluster.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1161625
With this update, virtual machines can now be searched for by the user who created them. Using the REST API, the search query is ".../api/vms?search=created_by_user_id%3D[USER_ID]". The required User ID can be retrieved by using ".../api/users". In addition, the Administration Portal shows the creator's name in the virtual machine general sub-tab. However, it is possible for the user to be removed from the system after the virtual machine is created.
With this update, virtual machines can now be searched for by the user who created them. Using the REST API, the search query is ".../api/vms?search=created_by_user_id%3D[USER_ID]". The required User ID can be retrieved by using ".../api/users". In addition, the Administration Portal shows the creator's name in the virtual machine general sub-tab. However, it is possible for the user to be removed from the system after the virtual machine is created.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1177229
With this update, image disks are now identifiable from within the guest by engine id, for example, by looking under /dev/disk/by-id. The disk id is now passed to the guest as the disk serial.
With this update, image disks are now identifiable from within the guest by engine id, for example, by looking under /dev/disk/by-id. The disk id is now passed to the guest as the disk serial.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1177782
This update provides a link to the gluster volume when creating a gluster storage domain, and enables a single unified flow. This enables the backup volfile server mount options to be auto-populated, and paves the way for integration features like Disaster Recovery setup using gluster geo-replication.
This update provides a link to the gluster volume when creating a gluster storage domain, and enables a single unified flow. This enables the backup volfile server mount options to be auto-populated, and paves the way for integration features like Disaster Recovery setup using gluster geo-replication.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1182369
Previously, in a hyper-converged cluster environment containing gluster and virt nodes, it was possible to create a replica set containing bricks from the same server. A warning appeared but the action was enabled even though there was a risk of losing data or service. In this release, it will no longer be possible to create a replica set containing multiple bricks from the same server in a hyper-converged environment.
Previously, in a hyper-converged cluster environment containing gluster and virt nodes, it was possible to create a replica set containing bricks from the same server. A warning appeared but the action was enabled even though there was a risk of losing data or service. In this release, it will no longer be possible to create a replica set containing multiple bricks from the same server in a hyper-converged environment.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1195140
The vioscsi.sys file is now compatible with Microsoft Cluster Services, which enables the Windows virtio-scsi driver to support Windows Server Failover Clustering (WSFC) using shared storage. As a result, vioscsi.sys can pass all tests in the "Validate a Configuration" process.
The vioscsi.sys file is now compatible with Microsoft Cluster Services, which enables the Windows virtio-scsi driver to support Windows Server Failover Clustering (WSFC) using shared storage. As a result, vioscsi.sys can pass all tests in the "Validate a Configuration" process.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1196433
Previously, in GlusterFS, if a node went down and then returned, GlusterFS would automatically initiate a self-heal operation. During this operation, which could be time-consuming, a subsequent maintenance mode action within the same GlusterFS replica set could result in a split-brain scenario. In this release, if a Gluster host is performing a self-heal activity, administrators will not be able to move it into maintenance mode. In extreme cases, administrators can use the force option to forcefully move a host into maintenance mode.
Previously, in GlusterFS, if a node went down and then returned, GlusterFS would automatically initiate a self-heal operation. During this operation, which could be time-consuming, a subsequent maintenance mode action within the same GlusterFS replica set could result in a split-brain scenario. In this release, if a Gluster host is performing a self-heal activity, administrators will not be able to move it into maintenance mode. In extreme cases, administrators can use the force option to forcefully move a host into maintenance mode.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1201020
Previously, the system performed automatic migrations, when required, without displaying the reason for doing so in the Event log or in the Administrator portal. In this release, after performing an automatic migration, the reason for doing so will be displayed.
Previously, the system performed automatic migrations, when required, without displaying the reason for doing so in the Event log or in the Administrator portal. In this release, after performing an automatic migration, the reason for doing so will be displayed.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1213291
This update introduced a check in the host maintenance flow to ensure glusterFS quorum can be maintained for all glusterFS volumes that have the 'cluster.quorum-type' option set. Similarly, there is a new check to ensure that the host moving to maintenance does not have a glusterFS brick that is a source of volume self-healing. These checks will be performed by default when moving the host to maintenance. There is an option in the Manager to skip these checks, but this can result in bringing your system to halt. This option should only be used in extreme cases.
This update introduced a check in the host maintenance flow to ensure glusterFS quorum can be maintained for all glusterFS volumes that have the 'cluster.quorum-type' option set. Similarly, there is a new check to ensure that the host moving to maintenance does not have a glusterFS brick that is a source of volume self-healing. These checks will be performed by default when moving the host to maintenance. There is an option in the Manager to skip these checks, but this can result in bringing your system to halt. This option should only be used in extreme cases.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1226206
Previously, when importing a virtual machine from a data storage domain, if the virtual machine had a “bad” MAC address, a MAC collision could occur in the target LAN. A “bad” MAC address is an address that is already in use or an address that is out of the range in the target cluster. In this release, it is possible to assign a new MAC address when importing the virtual machine from a data storage domain.
Previously, when importing a virtual machine from a data storage domain, if the virtual machine had a “bad” MAC address, a MAC collision could occur in the target LAN. A “bad” MAC address is an address that is already in use or an address that is out of the range in the target cluster. In this release, it is possible to assign a new MAC address when importing the virtual machine from a data storage domain.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1235200
Previously, when restoring a backup of a hosted engine on a different environment, for disaster recovery purposes, administrators were sometimes required to remove the previous hosts from the engine. This was accomplished from within the engine's database, which is a risk-prone procedure. In this release, a new CLI option can be used during the restore procedure to enable administrators to remove the previous host directly from the engine backup.
Previously, when restoring a backup of a hosted engine on a different environment, for disaster recovery purposes, administrators were sometimes required to remove the previous hosts from the engine. This was accomplished from within the engine's database, which is a risk-prone procedure. In this release, a new CLI option can be used during the restore procedure to enable administrators to remove the previous host directly from the engine backup.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1240466
Previously, when restoring a backup of a self-hosted engine on a different environment, for disaster recovery purposes, administrators were sometimes required to remove the previous self-hosted engine's storage domain and virtual machine. This was accomplished from within the engine's database, which was a risk-prone procedure. With this update, a new CLI option enables administrators to remove the previous self-hosted engine's storage domain and virtual machine directly from the backup of the engine during the restore procedure.Previously, when restoring a backup of a self-hosted engine on a different environment, for disaster recovery purposes, administrators were sometimes required to remove the previous self-hosted engine's storage domain and virtual machine. This was accomplished from within the engine's database, which was a risk-prone procedure. With this update, a new CLI option enables administrators to remove the previous self-hosted engine's storage domain and virtual machine directly from the backup of the engine during the restore procedure.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1241106
Previously, discard commands (UNMAP SCSI commands) that were sent from the guest were ignored by QEMU and were not passed on to the underlying storage. This meant that storage that was no longer in use could not be freed up. In this release, it is now possible to pass on discard commands to the underlying storage. A new property called Pass Discard was added to the Virtual Disk window. When selected, and if all the restrictions are met, discard commands that are sent from the guest will not be ignored by QEMU and will be passed on to the underlying storage. The reported unused blocks in the thinly provisioned LUNs of the underlying storage will be marked as free, and the reported consumed space will be reduced.
Previously, discard commands (UNMAP SCSI commands) that were sent from the guest were ignored by QEMU and were not passed on to the underlying storage. This meant that storage that was no longer in use could not be freed up. In this release, it is now possible to pass on discard commands to the underlying storage. A new property called Pass Discard was added to the Virtual Disk window. When selected, and if all the restrictions are met, discard commands that are sent from the guest will not be ignored by QEMU and will be passed on to the underlying storage. The reported unused blocks in the thinly provisioned LUNs of the underlying storage will be marked as free, and the reported consumed space will be reduced.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1246114
Previously, when the Virtual Machine was powered down, deleting a snapshot could potentially be a very long process. This was due to the need to copy the data from the base snapshot to the top snapshot, where the base snapshot is usually larger than the top snapshot. Now, when deleting a snapshot when the Virtual Machine is powered down, data is copied from the top snapshot to the base snapshot, which significantly reduces the time required to delete the snapshot.
Previously, when the Virtual Machine was powered down, deleting a snapshot could potentially be a very long process. This was due to the need to copy the data from the base snapshot to the top snapshot, where the base snapshot is usually larger than the top snapshot. Now, when deleting a snapshot when the Virtual Machine is powered down, data is copied from the top snapshot to the base snapshot, which significantly reduces the time required to delete the snapshot.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1254073
With this update, support for Gluster arbiter volume creation has been added to Red Hat Virtualization. Arbiter volumes are recommended in place of regular three-way replicated volumes to save storage space.
With this update, support for Gluster arbiter volume creation has been added to Red Hat Virtualization. Arbiter volumes are recommended in place of regular three-way replicated volumes to save storage space.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1256500
Previously, deleting a snapshot while the virtual machine was down took a long time. With this release, snapshot deletion while a virtual machine is down uses 'qemu-img commit' instead of 'qemu-img rebase', improving the performance of the operation.
Previously, deleting a snapshot while the virtual machine was down took a long time. With this release, snapshot deletion while a virtual machine is down uses 'qemu-img commit' instead of 'qemu-img rebase', improving the performance of the operation.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1264085
With this update, the user can configure the number of memory slots reserved for spare self-hosted engine hosts if the current host crashes. Previously, there was a chance that the self-hosted engine virtual machine would not have a place to start on a loaded cluster and this compromised the high availability feature. Now, the self-hosted engine will have a place to start a backup host so that it is ready to accept the virtual machine if the current host crashes.
With this update, the user can configure the number of memory slots reserved for spare self-hosted engine hosts if the current host crashes. Previously, there was a chance that the self-hosted engine virtual machine would not have a place to start on a loaded cluster and this compromised the high availability feature. Now, the self-hosted engine will have a place to start a backup host so that it is ready to accept the virtual machine if the current host crashes.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1270719
With this update, the option '--accept-defaults' has been added to the engine-setup command. This option causes engine-setup to no longer prompt for answers that have a default. This option saves time for the user as they no longer need to answer the prompts individually, if they are planning to accept the defaults, and also allows other tools to run engine-setup unattended. If the engine-setup command is run using this option and a weak password is provided, the user will be prompted for a stronger password because the default answer to 'Use weak password?' is No. To work around this, add the answer to an answer file.
With this update, the option '--accept-defaults' has been added to the engine-setup command. This option causes engine-setup to no longer prompt for answers that have a default. This option saves time for the user as they no longer need to answer the prompts individually, if they are planning to accept the defaults, and also allows other tools to run engine-setup unattended. If the engine-setup command is run using this option and a weak password is provided, the user will be prompted for a stronger password because the default answer to 'Use weak password?' is No. To work around this, add the answer to an answer file.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1277675
This feature allows you to map external VNIC profiles that are defined on an imported VM to the ones that are present in the cluster the VM is going to be imported to. The previous solution exchanged all external VNIC profiles that were not present in the target cluster with an empty profile, which removed the imported VM's network functionality. Now, after importing a VM from a data domain, the VM is configured properly according to the VNIC profiles that are defined in the target cluster.
This feature allows you to map external VNIC profiles that are defined on an imported VM to the ones that are present in the cluster the VM is going to be imported to. The previous solution exchanged all external VNIC profiles that were not present in the target cluster with an empty profile, which removed the imported VM's network functionality. Now, after importing a VM from a data domain, the VM is configured properly according to the VNIC profiles that are defined in the target cluster.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1279378
With this update, a new option to check for upgrades has been added when installing a host. In the Administration Portal this is available in the host installation menu and it can be triggered by using the hosts upgradecheck endpoint in the REST API.
With this update, a new option to check for upgrades has been added when installing a host. In the Administration Portal this is available in the host installation menu and it can be triggered by using the hosts upgradecheck endpoint in the REST API.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1285499
Previously, the column control menu in the Administration Portal, which is accessed by right-clicking on a column header in any of the tables, contained arrows for controlling the order of the columns. In this release, the order of the columns in this menu can be defined by dragging and dropping the column to the required position within the menu.
Previously, the column control menu in the Administration Portal, which is accessed by right-clicking on a column header in any of the tables, contained arrows for controlling the order of the columns. In this release, the order of the columns in this menu can be defined by dragging and dropping the column to the required position within the menu.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1286632
In this release, a link has been added to the Edit Fence Agent window, which opens the online help and displays information about the parameters that can be set for fence agents.
In this release, a link has been added to the Edit Fence Agent window, which opens the online help and displays information about the parameters that can be set for fence agents.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1289155
With this update, the limit of virtual CPUs has been increased to reflect the capabilities of Red Hat Enterprise Linux 7.3 hosts. It is now possible to configure a virtual machine with up to 288 vCPUs when hosts are in a 4.1 cluster.
With this update, the limit of virtual CPUs has been increased to reflect the capabilities of Red Hat Enterprise Linux 7.3 hosts. It is now possible to configure a virtual machine with up to 288 vCPUs when hosts are in a 4.1 cluster.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1289290
Previously, if a live migration was performed with extreme memory write intensive workloads, the migration would never be able to complete because QEMU could not transfer the memory changes fast enough. In this case, the migration could not reach the non-live finishing phase. In this release and in these situations, RHV will restrict the amount of CPU given to the guest to reduce the rate at which memory is changed and allow the migration to complete.
Previously, if a live migration was performed with extreme memory write intensive workloads, the migration would never be able to complete because QEMU could not transfer the memory changes fast enough. In this case, the migration could not reach the non-live finishing phase. In this release and in these situations, RHV will restrict the amount of CPU given to the guest to reduce the rate at which memory is changed and allow the migration to complete.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1294629
With this update, the loading performance of external virtual machines from an external server has been improved for VMware, KVM, and Xen. Previously, when displaying a list of virtual machines, libvirt was asked for the full information for each virtual machine when only the virtual machine names were required. Now, libvirt is only asked for the virtual machine names at the first import dialog and only imports the full virtual machine data list after the user has selected the required virtual machines.
With this update, the loading performance of external virtual machines from an external server has been improved for VMware, KVM, and Xen. Previously, when displaying a list of virtual machines, libvirt was asked for the full information for each virtual machine when only the virtual machine names were required. Now, libvirt is only asked for the virtual machine names at the first import dialog and only imports the full virtual machine data list after the user has selected the required virtual machines.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1299677
With this update, CPU hotplug is now supported on POWER for selected guest operating systems. CPU hotplug is now supported on both x86_64 architecture and on selected guest operating systems for ppc64le.
With this update, CPU hotplug is now supported on POWER for selected guest operating systems. CPU hotplug is now supported on both x86_64 architecture and on selected guest operating systems for ppc64le.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1300591
With this update, the user can now customize the virtual machine disk size when using the engine-appliance.
With this update, the user can now customize the virtual machine disk size when using the engine-appliance.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1300959
The latest virtio-win release, which includes Windows 10 drivers, is now required by the Manager to ensure that suitable drivers can be supplied to virtual machines during installation of Windows 10.
The latest virtio-win release, which includes Windows 10 drivers, is now required by the Manager to ensure that suitable drivers can be supplied to virtual machines during installation of Windows 10.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1301681
This feature allows you to edit the configuration stored in the self-hosted engine's shared storage.
This feature allows you to edit the configuration stored in the self-hosted engine's shared storage.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1302185
Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1306263
The weighting for virtual machine scheduling has been updated. The best host for the virtual machine is now selected using a weighted rank algorithm instead of the pure sum of weights. A rank is calculated for the policy unit and host, and the weight multiplier is then used to multiply the ranks for the given policy unit. The host with the highest number is selected. The reason for the change is that current weight policy units do not use a common result value range. Each unit reports numbers as needed, and this causes issues with user configured preferences. For example, memory (which has high numbers) always wins over CPU (0-100). This update ensures that the impact of the policy unit multiplier for the scheduling policy configuration is more predictable. However, users who are using it should check their configuration for sanity when upgrading.
The weighting for virtual machine scheduling has been updated. The best host for the virtual machine is now selected using a weighted rank algorithm instead of the pure sum of weights. A rank is calculated for the policy unit and host, and the weight multiplier is then used to multiply the ranks for the given policy unit. The host with the highest number is selected. The reason for the change is that current weight policy units do not use a common result value range. Each unit reports numbers as needed, and this causes issues with user configured preferences. For example, memory (which has high numbers) always wins over CPU (0-100). This update ensures that the impact of the policy unit multiplier for the scheduling policy configuration is more predictable. However, users who are using it should check their configuration for sanity when upgrading.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1309930
This update ensures that the Manager signs certificates using the SHA-256 algorithm instead of SHA-1 because SHA-256 is more secure and is expected to have a longer life expectancy. Only the default for new certificates was changed. To change certificates for existing hosts, they need to be reinstalled, or to have their certificates enrolled. Other certificates require a completely new setup, using engine-cleanup and engine-setup, including the one for httpd.
This update ensures that the Manager signs certificates using the SHA-256 algorithm instead of SHA-1 because SHA-256 is more secure and is expected to have a longer life expectancy. Only the default for new certificates was changed. To change certificates for existing hosts, they need to be reinstalled, or to have their certificates enrolled. Other certificates require a completely new setup, using engine-cleanup and engine-setup, including the one for httpd.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1314387
This update adds the ability to download or upload Red Hat Virtualization images (for example, virtual machine images) using the Red Hat Virtualization API.
This update adds the ability to download or upload Red Hat Virtualization images (for example, virtual machine images) using the Red Hat Virtualization API.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1316692
Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1317429
This update adds the ability to acquire a lease per virtual machine on shared storage, without attaching the lease to a disk. This adds the capability to avoid split-brain, and to avoid starting a virtual machine on another host if the original host becomes non-responsive, thereby improving virtual machine high availability.
This update adds the ability to acquire a lease per virtual machine on shared storage, without attaching the lease to a disk. This adds the capability to avoid split-brain, and to avoid starting a virtual machine on another host if the original host becomes non-responsive, thereby improving virtual machine high availability.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1317436
MAC address pools are now bound to the cluster instead of the data center because certain environments require MAC address pools on the cluster level. On the REST layer, the mac_pool attribute was added to the cluster, and can be set or queried. The StoragePool resource (represents the data center) was also altered. When updating its mac_pool_id, all clusters in a given data center will be updated to use this MAC address pool. When StoragePool is queried using the REST GET method, the ID of the MAC address pool will be reported only when all clusters in the given data center are using the same MAC address pool. Otherwise, the user needs to use the Cluster resource to get the MAC address pool of each individual cluster.
MAC address pools are now bound to the cluster instead of the data center because certain environments require MAC address pools on the cluster level. On the REST layer, the mac_pool attribute was added to the cluster, and can be set or queried. The StoragePool resource (represents the data center) was also altered. When updating its mac_pool_id, all clusters in a given data center will be updated to use this MAC address pool. When StoragePool is queried using the REST GET method, the ID of the MAC address pool will be reported only when all clusters in the given data center are using the same MAC address pool. Otherwise, the user needs to use the Cluster resource to get the MAC address pool of each individual cluster.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1317447
This feature allows you to request a new MAC address when importing a virtual machine from a data storage domain. This allows you to avoid importing a virtual machine with a bad MAC address, which might cause a MAC address collision in the target LAN. A MAC address would be considered "bad" if it is already in use in the target cluster or is out of the range of the MAC pool of the target cluster.
This feature allows you to request a new MAC address when importing a virtual machine from a data storage domain. This allows you to avoid importing a virtual machine with a bad MAC address, which might cause a MAC address collision in the target LAN. A MAC address would be considered "bad" if it is already in use in the target cluster or is out of the range of the MAC pool of the target cluster.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1318350
The NTP configuration is automatically set when deploying the self-hosted engine.
The NTP configuration is automatically set when deploying the self-hosted engine.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1325864
This update adds support for deploying Gluster storage during the self-hosted engine deployment through the Cockpit UI. Previously, the user needed to first deploy the Gluster storage using gdeploy, then deploy the self-hosted engine using the Cockpit UI, and configuration files had to be manually updated.
This update adds support for deploying Gluster storage during the self-hosted engine deployment through the Cockpit UI. Previously, the user needed to first deploy the Gluster storage using gdeploy, then deploy the self-hosted engine using the Cockpit UI, and configuration files had to be manually updated.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1330138
This update ensures that a self-hosted engine deployment works without needing to disable NetworkManager.
This update ensures that a self-hosted engine deployment works without needing to disable NetworkManager.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1331858
This update allows you to enable SSH access for the Manager virtual machine when deploying the self-hosted engine. You can choose between yes, no, and without-password. You can also pass a public SSH key for the root user during deployment.
This update allows you to enable SSH access for the Manager virtual machine when deploying the self-hosted engine. You can choose between yes, no, and without-password. You can also pass a public SSH key for the root user during deployment.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1333045
A new 'original_template' property has been introduced for the 'vm' REST API resource. This enables the user to get information about the template the virtual machine was based on before cloning.
A new 'original_template' property has been introduced for the 'vm' REST API resource. This enables the user to get information about the template the virtual machine was based on before cloning.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1333436
Previously, support for Legacy USB was deprecated and the UI displayed three options: Native, Legacy (Deprecated), and Disabled. In this release, the Legacy option has been completely removed and the UI now displays two options: Enabled and Disabled.
Previously, support for Legacy USB was deprecated and the UI displayed three options: Native, Legacy (Deprecated), and Disabled. In this release, the Legacy option has been completely removed and the UI now displays two options: Enabled and Disabled.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1337101
With this release, /dev/random is now the default random number generator in clusters with a cluster compatibility level of 4.0 and earlier, and /dev/urandom is now the default random number generator in clusters with a cluster compatibility level of 4.1 and later. Because these random number generators are enabled by default, the option to enable them has now been removed from the New Cluster and Edit Cluster windows. However, you can select the random number generator source for individual virtual machines from the New Virtual Machine and Edit Virtual Machine windows.
With this release, /dev/random is now the default random number generator in clusters with a cluster compatibility level of 4.0 and earlier, and /dev/urandom is now the default random number generator in clusters with a cluster compatibility level of 4.1 and later. Because these random number generators are enabled by default, the option to enable them has now been removed from the New Cluster and Edit Cluster windows. However, you can select the random number generator source for individual virtual machines from the New Virtual Machine and Edit Virtual Machine windows.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1341153
Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1342919
With this update, it is now possible to configure discard after delete per block storage domain. Previously, a user could get similar functionality by configuring the discard_enable parameter in the VDSM configuration file. This caused each logical volume (disk or snapshot) that was about to be removed by this specific host to be discarded first. Now, discard after delete can be enabled for a block storage domain, rather than a host. This means that if discard after delete is enabled, it no longer matters which host removes the logical volume, as each logical volume under this domain will be discarded before it is removed.
With this update, it is now possible to configure discard after delete per block storage domain. Previously, a user could get similar functionality by configuring the discard_enable parameter in the VDSM configuration file. This caused each logical volume (disk or snapshot) that was about to be removed by this specific host to be discarded first. Now, discard after delete can be enabled for a block storage domain, rather than a host. This means that if discard after delete is enabled, it no longer matters which host removes the logical volume, as each logical volume under this domain will be discarded before it is removed.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1343562
This update ensures that only hosts that have the status Up or NonOperational are checked for upgrades. Previously hosts with the status Maintenance were also checked, but often they were not reachable, which caused errors in Events.
This update ensures that only hosts that have the status Up or NonOperational are checked for upgrades. Previously hosts with the status Maintenance were also checked, but often they were not reachable, which caused errors in Events.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1344521
Previously, if the guest agent was not running or was out of date, the hover text message that appeared next to the explanation mark for the problematic virtual machine informed the user that the operating system did not match or that the timezone configuration was incorrect. In this release, the hover text will correctly display a message informing the user that the guest agent needs to be installed and running in the guest.
Previously, if the guest agent was not running or was out of date, the hover text message that appeared next to the explanation mark for the problematic virtual machine informed the user that the operating system did not match or that the timezone configuration was incorrect. In this release, the hover text will correctly display a message informing the user that the guest agent needs to be installed and running in the guest.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1347631
With this update, a log file located in /var/log/httpd/ovirt-requests-log now logs all requests made to the Red Hat Virtualization Manager via HTTPS, including how long each request took. There is a 'Correlation-Id' header included to allow for easier comparison of requests with the engine.log. CorrelationIds are now generated for every request automatically and can be passed to the REST API per Correlation-Id header or per correlation_id query parameter.
With this update, a log file located in /var/log/httpd/ovirt-requests-log now logs all requests made to the Red Hat Virtualization Manager via HTTPS, including how long each request took. There is a 'Correlation-Id' header included to allow for easier comparison of requests with the engine.log. CorrelationIds are now generated for every request automatically and can be passed to the REST API per Correlation-Id header or per correlation_id query parameter.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1348107
With this update, a user can now save a provider for external libvirt connection in the Providers tree section. When a user tries to import a virtual machine from libvirt+kvm to the Red Hat Virtualization environment the saved provider is available instead of having to re-enter the address.
With this update, a user can now save a provider for external libvirt connection in the Providers tree section. When a user tries to import a virtual machine from libvirt+kvm to the Red Hat Virtualization environment the saved provider is available instead of having to re-enter the address.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1349301
The self-hosted engine only supports deployment using the RHV-M Appliance. With this release, the deployment script allows you to download and install the Appliance RPM directly, instead of having to install it before deployment.
The self-hosted engine only supports deployment using the RHV-M Appliance. With this release, the deployment script allows you to download and install the Appliance RPM directly, instead of having to install it before deployment.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1349321
When users import virtual machines from Xen on RHEL to Red Hat Virtualization it will access the saved provider address instead of the user having to re-enter the address.
When users import virtual machines from Xen on RHEL to Red Hat Virtualization it will access the saved provider address instead of the user having to re-enter the address.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1349907
Previously, after_hibernation hooks were never executed. With this release, before_hibernation and after_hibernation hooks are always executed on the guest operating system (with the guest agent installed) when suspending and resuming a virtual machine.
Previously, after_hibernation hooks were never executed. With this release, before_hibernation and after_hibernation hooks are always executed on the guest operating system (with the guest agent installed) when suspending and resuming a virtual machine.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1350465
Previously, when importing a virtual machine, if the import failed, the output of the virt-v2v tool was not available for investigating the reason for the failure, and the import had to be reproduced manually. In this release, the output of virt-v2v is now stored in the /var/log/vdsm/import directory. All logs older than 30 days are automatically removed.
Previously, when importing a virtual machine, if the import failed, the output of the virt-v2v tool was not available for investigating the reason for the failure, and the import had to be reproduced manually. In this release, the output of virt-v2v is now stored in the /var/log/vdsm/import directory. All logs older than 30 days are automatically removed.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1353556
Previously, a Dashboard tab was introduced to the Administration Portal. However, when loading the Administration Portal the user landed at the Virtual Machines tab followed by an immediate switch to the Dashboard tab. Now, the UI plugin has been improved to allow pre-loading of UI plugins, such as ovirt-engine-dashboard. This means that the user lands directly at the Dashboard tab.
Previously, a Dashboard tab was introduced to the Administration Portal. However, when loading the Administration Portal the user landed at the Virtual Machines tab followed by an immediate switch to the Dashboard tab. Now, the UI plugin has been improved to allow pre-loading of UI plugins, such as ovirt-engine-dashboard. This means that the user lands directly at the Dashboard tab.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1353750
With this update, the debug logging for ovirt-engine-extension-aaa-ldap has been updated. When ovirt-engine-extension-aaa-ldap is enabled the following messages will show in the logs. The LDAP server that authenticated a user is shown as "User 'myuser1' is performing bind request to: ldap.example.com" and the LDAP server that performed a search request is shown as "Performing SearchRequest '...' request on server ldap.example.com."
With this update, the debug logging for ovirt-engine-extension-aaa-ldap has been updated. When ovirt-engine-extension-aaa-ldap is enabled the following messages will show in the logs. The LDAP server that authenticated a user is shown as "User 'myuser1' is performing bind request to: ldap.example.com" and the LDAP server that performed a search request is shown as "Performing SearchRequest '...' request on server ldap.example.com."Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1354343
This update includes the Post-copy migration policy, which is available as a Technology Preview feature. The policy is similar to the Minimal Downtime policy, and enables the virtual machine to start running on the destination host as soon as possible. During the final phase of the migration (post-copy phase), the missing parts of the memory content is transferred between the hosts on demand. This guarantees that the migration will eventually converge with very little downtime. The disadvantage of this policy is that in the post-copy phase, the virtual machine may slow down significantly as the missing parts of memory are transferred between the hosts. If anything goes wrong during the post-copy phase, such as a network failure between the hosts, the running virtual machine instance will be lost. It is therefore not possible to abort a migration during the post-copy phase.
This update includes the Post-copy migration policy, which is available as a Technology Preview feature. The policy is similar to the Minimal Downtime policy, and enables the virtual machine to start running on the destination host as soon as possible. During the final phase of the migration (post-copy phase), the missing parts of the memory content is transferred between the hosts on demand. This guarantees that the migration will eventually converge with very little downtime. The disadvantage of this policy is that in the post-copy phase, the virtual machine may slow down significantly as the missing parts of memory are transferred between the hosts. If anything goes wrong during the post-copy phase, such as a network failure between the hosts, the running virtual machine instance will be lost. It is therefore not possible to abort a migration during the post-copy phase.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1356161
With this release, if you do not specify any NUMA mapping, Red Hat Virtualization defaults to a NUMA node that contains the host device's memory-mapped I/O (MMIO). This configuration is only preferred, rather than strictly required.
With this release, if you do not specify any NUMA mapping, Red Hat Virtualization defaults to a NUMA node that contains the host device's memory-mapped I/O (MMIO). This configuration is only preferred, rather than strictly required.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1358716
The self-hosted engine setup wizard now warns users if the host is already registered to Red Hat Virtualization Manager. Previously, a host that was registered to the Manager but not running a self-hosted engine would present the option to set up a self-hosted engine, which ran the risk of unregistering the host. Now, hosts that are registered to the Manager present a "Redeploy" button in the Hosted Engine wizard in Cockpit, which must be selected in order to continue.
The self-hosted engine setup wizard now warns users if the host is already registered to Red Hat Virtualization Manager. Previously, a host that was registered to the Manager but not running a self-hosted engine would present the option to set up a self-hosted engine, which ran the risk of unregistering the host. Now, hosts that are registered to the Manager present a "Redeploy" button in the Hosted Engine wizard in Cockpit, which must be selected in order to continue.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1361115
This update adds Gluster-related fencing policies for hyper-converged clusters. Previous fencing policies ignored Gluster processes. But in Hyper-converged mode, fencing policies are required to ensure that a host is not fenced if there is a brick process running, or to ensure no loss of quorum when shutting down the host with an active brick. The following fencing policies have been added to Hyper-converged clusters: - SkipFencingIfGlusterBricksUp: Fencing will be skipped if bricks are running and can be reached from other peers. - SkipFencingIfGlusterQuorumNotMet: Fencing will be skipped if bricks are running and shutting down the host will cause loss of quorum
This update adds Gluster-related fencing policies for hyper-converged clusters. Previous fencing policies ignored Gluster processes. But in Hyper-converged mode, fencing policies are required to ensure that a host is not fenced if there is a brick process running, or to ensure no loss of quorum when shutting down the host with an active brick. The following fencing policies have been added to Hyper-converged clusters: - SkipFencingIfGlusterBricksUp: Fencing will be skipped if bricks are running and can be reached from other peers. - SkipFencingIfGlusterQuorumNotMet: Fencing will be skipped if bricks are running and shutting down the host will cause loss of quorumCopy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1361230
Red Hat Virtualization Host (RHVH) 4.0 allows users to install RPMs. However, installed RPMs are lost after upgrading RHVH. RHVH 4.1 now includes a yum plugin that saves and reinstalls RPM packages after upgrading, to ensure that installed RPMs are no longer lost after upgrading. This will not work when upgrading from RHVH 4.0 to RHVH 4.1.
Red Hat Virtualization Host (RHVH) 4.0 allows users to install RPMs. However, installed RPMs are lost after upgrading RHVH. RHVH 4.1 now includes a yum plugin that saves and reinstalls RPM packages after upgrading, to ensure that installed RPMs are no longer lost after upgrading. This will not work when upgrading from RHVH 4.0 to RHVH 4.1.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1362604
The rng-tools package has been added to oVirt Node NG / RHV-H. This tool is required for the TPM module to be able to work with the Random Number Generator.
The rng-tools package has been added to oVirt Node NG / RHV-H. This tool is required for the TPM module to be able to work with the Random Number Generator.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1365021
This enhancement is a rebase on the jsonrpc Dispatcher APIs to provide better performance and make the code more robust.
This enhancement is a rebase on the jsonrpc Dispatcher APIs to provide better performance and make the code more robust.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1365937
A mobile client for Red Hat Enterprise Virtualization, which is compatible with Red Hat Enterprise Virtualization 3.5 onwards, is available for Android devices.
A mobile client for Red Hat Enterprise Virtualization, which is compatible with Red Hat Enterprise Virtualization 3.5 onwards, is available for Android devices.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1366118
oVirt release now provides repository configuration files for enabling GlusterFS 3.8 repositories on Red Hat Enterprise Linux, CentOS Linux, and similar operating systems.
oVirt release now provides repository configuration files for enabling GlusterFS 3.8 repositories on Red Hat Enterprise Linux, CentOS Linux, and similar operating systems.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1366183
Since Red Hat Virtualization now has the capability to deploy additional self-hosted engine hosts from the Manager with host-deploy, the capability to deploy additional self-hosted engine hosts from hosted-engine setup is not required anymore. It has now been removed. Similarly, the RHV-M Appliance has proved to be the easiest flow to have a working self-hosted engine environment; all other bootstrap flows have now been removed.
Since Red Hat Virtualization now has the capability to deploy additional self-hosted engine hosts from the Manager with host-deploy, the capability to deploy additional self-hosted engine hosts from hosted-engine setup is not required anymore. It has now been removed. Similarly, the RHV-M Appliance has proved to be the easiest flow to have a working self-hosted engine environment; all other bootstrap flows have now been removed.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1366899
This release adds support for overlay networks using Open Virtual Network (OVN) as a Technology Preview. This feature allows you to add OVN as an external network provider, and import or create networks from it in the Red Hat Virtualization Manager. You can then provision virtual machines with network interfaces connected using these logical overlays (OVN networks).
This release adds support for overlay networks using Open Virtual Network (OVN) as a Technology Preview. This feature allows you to add OVN as an external network provider, and import or create networks from it in the Red Hat Virtualization Manager. You can then provision virtual machines with network interfaces connected using these logical overlays (OVN networks).Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1367826
Previously, Python SDK was configured to communicate with the server using uncompressed responses. This caused long response times. In this release, the default configuration is to send compressed responses.
Previously, Python SDK was configured to communicate with the server using uncompressed responses. This caused long response times. In this release, the default configuration is to send compressed responses.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1367924
Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1369175
The "Enable USB Auto-Share" option in the "Console options" dialog is now only available if "USB Support" is enabled on the virtual machine.
The "Enable USB Auto-Share" option in the "Console options" dialog is now only available if "USB Support" is enabled on the virtual machine.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1370485
Previously, Java SDK was configured to communicate with the server using uncompressed responses. This caused long response times. In this release, the default configuration is to send compressed responses.
Previously, Java SDK was configured to communicate with the server using uncompressed responses. This caused long response times. In this release, the default configuration is to send compressed responses.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1371530
In this release, when installing or reinstalling hosts, the collectd and fluentd packages are now installed, including the relevant plugins. These can be used to send statistics to a central metrics store.
In this release, when installing or reinstalling hosts, the collectd and fluentd packages are now installed, including the relevant plugins. These can be used to send statistics to a central metrics store.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1373223
Previously, if SPICE USB redirection was disabled, libvirt created a default USB controller. With this update, if SPICE USB redirection is disabled then the virtual machine has a new USB controller, which is configurable per guest operating system and cluster version. This is defined in the osinfo-defaults.properties configuration file.
Previously, if SPICE USB redirection was disabled, libvirt created a default USB controller. With this update, if SPICE USB redirection is disabled then the virtual machine has a new USB controller, which is configurable per guest operating system and cluster version. This is defined in the osinfo-defaults.properties configuration file.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1375568
The tcpdump package is now included with Red Hat Virtualization Host.
The tcpdump package is now included with Red Hat Virtualization Host.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1376454
A reinstalling and restoring workflow was tested and confirmed for moving from version 3.6 Red Hat Enterprise Virtualization Hypervisor hosts to the new implementation, Red Hat Virtualization Host, in 4.0 or 4.1.
A reinstalling and restoring workflow was tested and confirmed for moving from version 3.6 Red Hat Enterprise Virtualization Hypervisor hosts to the new implementation, Red Hat Virtualization Host, in 4.0 or 4.1.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1379000
With this update, IBM Security (Tivoli) Directory Server has been added to supported LDAP servers in ovirt-engine-extension-aaa-ldap. This allows customers to attach Red Hat Virtualization 4.1 to their IBM Security (Tivoli) Directory Server setup and to use users and groups from this setup in Red Hat Virtualization.
With this update, IBM Security (Tivoli) Directory Server has been added to supported LDAP servers in ovirt-engine-extension-aaa-ldap. This allows customers to attach Red Hat Virtualization 4.1 to their IBM Security (Tivoli) Directory Server setup and to use users and groups from this setup in Red Hat Virtualization.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1379074
Previously, the ExportVmCommand appeared in the Engine log without the ID of the virtual machine being exported. This information has now been added to the log. Note: After this change, users must have export permissions for the virtual machine and its disks to export a virtual machine. Previously, permissions to export virtual machine disks were sufficient.
Previously, the ExportVmCommand appeared in the Engine log without the ID of the virtual machine being exported. This information has now been added to the log. Note: After this change, users must have export permissions for the virtual machine and its disks to export a virtual machine. Previously, permissions to export virtual machine disks were sufficient.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1379154
Previously, users who wanted to use Cockpit for system configuration needed to log in to the system and retrieve IP address information manually. Now, Red Hat Virtualization Host provides a message on login informing users of the URL to the Cockpit user interface.
Previously, users who wanted to use Cockpit for system configuration needed to log in to the system and retrieve IP address information manually. Now, Red Hat Virtualization Host provides a message on login informing users of the URL to the Cockpit user interface.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1379763
The "screen" package is now available as part of the base RHVH image.
The "screen" package is now available as part of the base RHVH image.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1379771
This release introduces a 'force' flag, which can be used to update a storage server connection regardless of the associated storage domain status (allowing updates even when the storage domain is not in Maintenance mode). For example: PUT /ovirt-engine/api/storageconnections/123;force=true
This release introduces a 'force' flag, which can be used to update a storage server connection regardless of the associated storage domain status (allowing updates even when the storage domain is not in Maintenance mode). For example: PUT /ovirt-engine/api/storageconnections/123;force=trueCopy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1380365
Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1381184
Red Hat Virtualization now supports headless virtual machines that run without a graphical console and display device. Headless mode is also supported for templates, pools, and instance types. This feature supports running a headless virtual machine from start, or after the initial setup (after "Run Once"). Headless mode can be enabled or disabled for a new or existing virtual machine at any time.
Red Hat Virtualization now supports headless virtual machines that run without a graphical console and display device. Headless mode is also supported for templates, pools, and instance types. This feature supports running a headless virtual machine from start, or after the initial setup (after "Run Once"). Headless mode can be enabled or disabled for a new or existing virtual machine at any time.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1382843
With this update, the Red Hat Virtualization Host includes sysstat as part of the base image.
With this update, the Red Hat Virtualization Host includes sysstat as part of the base image.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1383342
This feature allows you to request a console ticket for a specific graphics device by means of the REST API. The existing endpoint, /api/vms/{vmId}/ticket, defaulted to SPICE in scenarios when SPICE+VNC was configured as the graphics protocol, making it impossible to request a VNC ticket. Now, a ticket action has been added to the /api/vms/{vmId}/graphicsconsoles/{consoleId} resource, making it possible to request a ticket for a specific console. This specific endpoint is now preferred, and the pre-existing per-VM endpoint is considered deprecated.This feature allows you to request a console ticket for a specific graphics device by means of the REST API. The existing endpoint, /api/vms/{vmId}/ticket, defaulted to SPICE in scenarios when SPICE+VNC was configured as the graphics protocol, making it impossible to request a VNC ticket. Now, a ticket action has been added to the /api/vms/{vmId}/graphicsconsoles/{consoleId} resource, making it possible to request a ticket for a specific console. This specific endpoint is now preferred, and the pre-existing per-VM endpoint is considered deprecated.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1387254
Previously when integrating the Manager with an LDAP server using the ovirt-engine-extension-aaa-ldap-setup tool, the root of the LDAP tree (base DN) was selected automatically based on the LDAP server defaults. However, sometimes the defaults are incorrect for Manager integrations, and administrators are required to edit configuration files manually after the setup job completes. Now the ovirt-engine-extension-aaa-ldap-setup tool offers to override the default base DN retrieved from LDAP server, so manual changes are no longer necessary.
Previously when integrating the Manager with an LDAP server using the ovirt-engine-extension-aaa-ldap-setup tool, the root of the LDAP tree (base DN) was selected automatically based on the LDAP server defaults. However, sometimes the defaults are incorrect for Manager integrations, and administrators are required to edit configuration files manually after the setup job completes. Now the ovirt-engine-extension-aaa-ldap-setup tool offers to override the default base DN retrieved from LDAP server, so manual changes are no longer necessary.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1388245
This release adds the ability to specify a Maximum Memory value in all VM-like dialogs (Virtual Machine, Template, Pool, and Instance Type). It is accessible in the '{vm, template, instance_type}/memory_policy/max' tag in the REST API. The value defines the upper limit to which memory hot plug can be performed. The default value is 4x memory size.This release adds the ability to specify a Maximum Memory value in all VM-like dialogs (Virtual Machine, Template, Pool, and Instance Type). It is accessible in the '{vm, template, instance_type}/memory_policy/max' tag in the REST API. The value defines the upper limit to which memory hot plug can be performed. The default value is 4x memory size.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1388430
This release adds a maintenance tool to run vacuum actions on the engine database (or specific tables). This tool optimizes table stats and compacts the internals of tables, resulting in less disk space usage, more efficient future maintenance work, and updated table stats for better query planning. Also provided is an engine-setup dialog that offers to perform vacuum during upgrades. This can be automated by the answer file.
This release adds a maintenance tool to run vacuum actions on the engine database (or specific tables). This tool optimizes table stats and compacts the internals of tables, resulting in less disk space usage, more efficient future maintenance work, and updated table stats for better query planning. Also provided is an engine-setup dialog that offers to perform vacuum during upgrades. This can be automated by the answer file.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1388724
Previously, it was not possible to install Windows Server 2016 on a virtual machine. In this release, it is now possible to install Windows Server 2016 on a virtual machine. When adding a virtual machine, Windows Server 2016 appears in the list of available operating systems.
Previously, it was not possible to install Windows Server 2016 on a virtual machine. In this release, it is now possible to install Windows Server 2016 on a virtual machine. When adding a virtual machine, Windows Server 2016 appears in the list of available operating systems.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1389324
Previously, the Networking tab in Cockpit was disabled in Red Hat Virtualization Host (RHVH) images. This is now enabled, meaning that system networking can be configured through Cockpit in RHVH.
Previously, the Networking tab in Cockpit was disabled in Red Hat Virtualization Host (RHVH) images. This is now enabled, meaning that system networking can be configured through Cockpit in RHVH.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1392393
Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1392418
The user experience for HA global maintenance has been improved in the UI by moving the options to a more logical location, and providing a visual indication about the current state of HA global maintenance for a given host. The "Enable HA Global Maintenance" and "Disable HA Global Maintenance" buttons are now displayed on the right-click menu for hosts instead of virtual machines, and reflect the global maintenance state of the host by disabling the button matching the host's current HA global maintenance state. The previous method of displaying the options for virtual machines was unintuitive. Additionally, both the enable and disable options remained available regardless of whether or not the host was in HA global maintenance mode.
The user experience for HA global maintenance has been improved in the UI by moving the options to a more logical location, and providing a visual indication about the current state of HA global maintenance for a given host. The "Enable HA Global Maintenance" and "Disable HA Global Maintenance" buttons are now displayed on the right-click menu for hosts instead of virtual machines, and reflect the global maintenance state of the host by disabling the button matching the host's current HA global maintenance state. The previous method of displaying the options for virtual machines was unintuitive. Additionally, both the enable and disable options remained available regardless of whether or not the host was in HA global maintenance mode.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1392872
With this release, Intel Skylake family CPUs are now supported.
With this release, Intel Skylake family CPUs are now supported.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1393257
Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1393407
During the authorization stage of the login flow, the user's group memberships, including nested groups, are retrieved. Nested group memberships are resolved using recursive LDAP searches, which could take significant amount of time. This update uses a special Active Directory feature called LDAP_MATCHING_RULE_IN_CHAIN, which allows you to fetch complete group memberships, including nested groups, in one LDAP search.
During the authorization stage of the login flow, the user's group memberships, including nested groups, are retrieved. Nested group memberships are resolved using recursive LDAP searches, which could take significant amount of time. This update uses a special Active Directory feature called LDAP_MATCHING_RULE_IN_CHAIN, which allows you to fetch complete group memberships, including nested groups, in one LDAP search.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1393918
With this update, some ancillary self-hosted engine commands that were still based on xmlrpc have been moved to jsonrpc.
With this update, some ancillary self-hosted engine commands that were still based on xmlrpc have been moved to jsonrpc.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1396672
Since Red Hat Enterprise Virtualization 3.6, ovirt-ha-agent read its configuration and the Manager virtual machine specification from shared storage. Previously, they were local files replicated on each involved host. This enhancement modifies the output of hosted-engine --vm-status to show whether the configuration and the Manager virtual machine specification have been, on each reported host, correctly read from the shared storage.
Since Red Hat Enterprise Virtualization 3.6, ovirt-ha-agent read its configuration and the Manager virtual machine specification from shared storage. Previously, they were local files replicated on each involved host. This enhancement modifies the output of hosted-engine --vm-status to show whether the configuration and the Manager virtual machine specification have been, on each reported host, correctly read from the shared storage.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1397833
Previously, the Java heap size for Data Warehouse was not explicitly set. This resulted in the Java virtual machine using the default size, which could have been as large as a quarter of the machine's total memory. With this release, Data Warehouse's configuration was updated to allocate 1 GB of RAM, with the addition of two new parameters: DWH_HEAP_MIN=1g DWH_HEAP_MAX=1g The size can be set to a higher value for larger environments using these new parameters.
Previously, the Java heap size for Data Warehouse was not explicitly set. This resulted in the Java virtual machine using the default size, which could have been as large as a quarter of the machine's total memory. With this release, Data Warehouse's configuration was updated to allocate 1 GB of RAM, with the addition of two new parameters: DWH_HEAP_MIN=1g DWH_HEAP_MAX=1g The size can be set to a higher value for larger environments using these new parameters.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1398593
This feature integrates the setup for data sync to a remote location using geo-replication for Gluster-based storage domains, to improve disaster recovery. A user is able to schedule data sync to a remote location from the Red Hat Virtualization UI.
This feature integrates the setup for data sync to a remote location using geo-replication for Gluster-based storage domains, to improve disaster recovery. A user is able to schedule data sync to a remote location from the Red Hat Virtualization UI.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1399142
This release changes the default disk interface type from virtio-blk to virtio-scsi. virtio-blk is still supported, but users are encouraged to use the more modern virtio-scsi. When creating or attaching a disk to a virtual machine, the virtio-scsi interface type will now be selected by default.
This release changes the default disk interface type from virtio-blk to virtio-scsi. virtio-blk is still supported, but users are encouraged to use the more modern virtio-scsi. When creating or attaching a disk to a virtual machine, the virtio-scsi interface type will now be selected by default.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1400366
This update allows you to change the default network used by the host from the management network (ovirtmgmt) to a non-management network.
This update allows you to change the default network used by the host from the management network (ovirtmgmt) to a non-management network.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1402435
The self-hosted engine's machine type has now been upgraded for Red Hat Enterprise Linux 7 compatibility.
The self-hosted engine's machine type has now been upgraded for Red Hat Enterprise Linux 7 compatibility.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1403839
With this update, the ability to remove LUNs from a block data domain has been added. This means that LUNs can be removed from a block data domain provided that there is enough free space on the other domain devices to contain the data stored on the LUNs being removed.
With this update, the ability to remove LUNs from a block data domain has been added. This means that LUNs can be removed from a block data domain provided that there is enough free space on the other domain devices to contain the data stored on the LUNs being removed.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1404232
Previously, when a USB hub containing a redirected device was unplugged, spice-usbdk-win failed to clean up the redirected device. When the USB hub and its attached device were replugged, the device could not be redirected. In this release, the issue has been fixed. spice-usbdk-win will now clean up the redirected device as required. When the USB hub and the USB device are re-plugged, the device can be redirected to the guest.
Previously, when a USB hub containing a redirected device was unplugged, spice-usbdk-win failed to clean up the redirected device. When the USB hub and its attached device were replugged, the device could not be redirected. In this release, the issue has been fixed. spice-usbdk-win will now clean up the redirected device as required. When the USB hub and the USB device are re-plugged, the device can be redirected to the guest.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1404660
This feature adds rule enforcement support for VM to host affinity. VM to host affinity groups require the affinity rule enforcer to handle them in addition to the existing enforcement of VM to VM affinity. The rule enforcer will now be able to find VM to host affinity violations and choose a VM to migrate according to these violations.
This feature adds rule enforcement support for VM to host affinity. VM to host affinity groups require the affinity rule enforcer to handle them in addition to the existing enforcement of VM to VM affinity. The rule enforcer will now be able to find VM to host affinity violations and choose a VM to migrate according to these violations.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1405032
This release adds the VirtIO-RNG driver installer to the guest tools ISO for supported Windows versions.
This release adds the VirtIO-RNG driver installer to the guest tools ISO for supported Windows versions.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1405813
A script is now supplied to configure collectd and fluentd on hosts to send statistics to a central store.
A script is now supplied to configure collectd and fluentd on hosts to send statistics to a central store.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1405912
Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1405975
With this update, support for Red Hat Virtualization and oVirt has been added to Ansible. For more information about oVirt Ansible modules see http://docs.ansible.com/ansible/list_of_cloud_modules.html#ovirt
With this update, support for Red Hat Virtualization and oVirt has been added to Ansible. For more information about oVirt Ansible modules see http://docs.ansible.com/ansible/list_of_cloud_modules.html#ovirtCopy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1406814
This fix allows administrators to set the engine-config option "HostPackagesUpdateTimeInHours" to 0, which disables automatic periodical checks for host upgrades. Automatic periodical checks are not always needed, for example when managing hosts using Satellite.
This fix allows administrators to set the engine-config option "HostPackagesUpdateTimeInHours" to 0, which disables automatic periodical checks for host upgrades. Automatic periodical checks are not always needed, for example when managing hosts using Satellite.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1408193
Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1408876
This release enables Virtual Machines to lease areas on the storage domain. If a Virtual Machine has a lease on a storage domain, it will not be possible to move this storage domain into Maintenance mode. If the user attempts to do so, an error message will appear explaining that a virtual machine currently has a lease on this storage.
This release enables Virtual Machines to lease areas on the storage domain. If a Virtual Machine has a lease on a storage domain, it will not be possible to move this storage domain into Maintenance mode. If the user attempts to do so, an error message will appear explaining that a virtual machine currently has a lease on this storage.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1411654
Previously, rhvm-appliance was not available via subscriptions on the RHV-H repositories. In this release, rhvm-appliance is now used as the preferred deployment mechanism by ovirt-hosted-engine-setup, and is now available in the RHV-H repositories.
Previously, rhvm-appliance was not available via subscriptions on the RHV-H repositories. In this release, rhvm-appliance is now used as the preferred deployment mechanism by ovirt-hosted-engine-setup, and is now available in the RHV-H repositories.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1412095
Previously, the Networking tab was not available in the Cockpit in Red Hat Virtualization Host, even though the NetworkManager was enabled. With this release, the Networking tab is now available in the Cockpit, and administrators can use it to configure the network.
Previously, the Networking tab was not available in the Cockpit in Red Hat Virtualization Host, even though the NetworkManager was enabled. With this release, the Networking tab is now available in the Cockpit, and administrators can use it to configure the network.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1412547
Previously, when the Manager attempted to connect to VDSM it tried to negotiate the highest available version of TLS but due to previous issues there was a limitation to try TLSv1.0 as the highest version and to not try any higher version. Now, the limit has been removed so that TLSv1.1 and TLSv1.2 can be negotiated if they are available on the VDSM side. Removing this limit will allow TLSv1.0 to be dropped from future versions of VDSM.
Previously, when the Manager attempted to connect to VDSM it tried to negotiate the highest available version of TLS but due to previous issues there was a limitation to try TLSv1.0 as the highest version and to not try any higher version. Now, the limit has been removed so that TLSv1.1 and TLSv1.2 can be negotiated if they are available on the VDSM side. Removing this limit will allow TLSv1.0 to be dropped from future versions of VDSM.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1413150
The Red Hat Virtualization Manager now provides warnings for all data centers and clusters that have not been upgraded to latest installed version. The compatibility version of all data centers is checked once a week and on Manager startup. If it is not the latest version, an alert is raised and stored in the audit log. The Data Centers and Clusters main tabs now also show an exclamation mark icon for each data center or cluster that is not at the latest version. Hovering over this icon displays a recommendation to upgrade the compatibility version.
The Red Hat Virtualization Manager now provides warnings for all data centers and clusters that have not been upgraded to latest installed version. The compatibility version of all data centers is checked once a week and on Manager startup. If it is not the latest version, an alert is raised and stored in the audit log. The Data Centers and Clusters main tabs now also show an exclamation mark icon for each data center or cluster that is not at the latest version. Hovering over this icon displays a recommendation to upgrade the compatibility version.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1413181
In RHV 4.1 a new tools repository containing packages required by the Red Hat Virtualization Manager has been added. See the Release Notes or Installation Guide for repository details.
In RHV 4.1 a new tools repository containing packages required by the Red Hat Virtualization Manager has been added. See the Release Notes or Installation Guide for repository details.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1418145
Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1422024
With this release, the net-snmp package is part of the Red Hat Virtualization Host image by default.
With this release, the net-snmp package is part of the Red Hat Virtualization Host image by default.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1424821
It is now possible to create NFS storage domains with NFS version 4.2 via the REST API.
It is now possible to create NFS storage domains with NFS version 4.2 via the REST API.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1427987
Copy to Clipboard Copied! Toggle word wrap Toggle overflow
3.1.2. Technology Preview Copy linkLink copied to clipboard!
Copy linkLink copied to clipboard!
The items listed in this section are provided as Technology Previews. For further information on the scope of Technology Preview status, and the associated support implications, refer to https://access.redhat.com/support/offerings/techpreview/.
- BZ#1400569
In this release, a new user interface for the User Portal has been introduced as a Technology Preview. The new user interface offers improved performance. The new User Portal can be accessed from the following link: https://[ENGINE_HOST]/ovirt-engine/web-ui
In this release, a new user interface for the User Portal has been introduced as a Technology Preview. The new user interface offers improved performance. The new User Portal can be accessed from the following link: https://[ENGINE_HOST]/ovirt-engine/web-uiCopy to Clipboard Copied! Toggle word wrap Toggle overflow
3.1.3. Release Notes Copy linkLink copied to clipboard!
Copy linkLink copied to clipboard!
This section outlines important details about the release, including recommended practices and notable changes to Red Hat Virtualization. You must take this information into account to ensure the best possible outcomes for your deployment.
- BZ#1364978
Previously, Red Hat Virtualization Host (RHVH) was shipped without an End User License Agreement (EULA). In this release, this bug has been fixed and RHVH now includes an EULA.
Previously, Red Hat Virtualization Host (RHVH) was shipped without an End User License Agreement (EULA). In this release, this bug has been fixed and RHVH now includes an EULA.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1375563
Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1394831
RHEV Agent has been renamed to oVirt Agent on Windows.
RHEV Agent has been renamed to oVirt Agent on Windows.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1405612
rhevm-spice-client packages were renamed to spice-client-msi.
rhevm-spice-client packages were renamed to spice-client-msi.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1414085
Previously the name of the “rhevm-appliance” RPM contained only a timestamp, without versioning information. In this release, the Red Hat Virtualization release will now be included in the name of the “rhevm-appliance” RPM and will be visible from the node channel.
Previously the name of the “rhevm-appliance” RPM contained only a timestamp, without versioning information. In this release, the Red Hat Virtualization release will now be included in the name of the “rhevm-appliance” RPM and will be visible from the node channel.Copy to Clipboard Copied! Toggle word wrap Toggle overflow
3.1.4. Known Issues Copy linkLink copied to clipboard!
Copy linkLink copied to clipboard!
These known issues exist in Red Hat Virtualization at this time:
- BZ#1326709
When hosted-engine --deploy is run on additional hosts that have multiple FQDNs associated with it, the script will pick the host address that is returned by default. Ensure that a host's hostname resolves to the required FQDN/em1 before deploying the self-hosted engine.
When hosted-engine --deploy is run on additional hosts that have multiple FQDNs associated with it, the script will pick the host address that is returned by default. Ensure that a host's hostname resolves to the required FQDN/em1 before deploying the self-hosted engine.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1416153
Previously, after deleting a snapshot in a data center, the original volume's allocation policy and size differed from the pre-snapshot state. In this release, if a snapshot is created from a preallocated volume, when the snapshot is deleted, qemu-img is called to copy data from the top volume to the base volume. As a result, the original volume's allocation policy and size are identical.
Previously, after deleting a snapshot in a data center, the original volume's allocation policy and size differed from the pre-snapshot state. In this release, if a snapshot is created from a preallocated volume, when the snapshot is deleted, qemu-img is called to copy data from the top volume to the base volume. As a result, the original volume's allocation policy and size are identical.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1422442
Due to an unstable slave order in NetworkManager, DHCP over a bond created by NetworkManager may receive a different IP address after adding it to Red Hat Virtualization (RHV) or after rebooting. The workaround is to avoid using DHCP on a NetworkManager-controlled bond. NetworkManager may also remove a DHCP-provided host name after a host is added to RHV. To avoid this, persist the host name explicitly via Cockpit or hostnamectl.
Due to an unstable slave order in NetworkManager, DHCP over a bond created by NetworkManager may receive a different IP address after adding it to Red Hat Virtualization (RHV) or after rebooting. The workaround is to avoid using DHCP on a NetworkManager-controlled bond. NetworkManager may also remove a DHCP-provided host name after a host is added to RHV. To avoid this, persist the host name explicitly via Cockpit or hostnamectl.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1433434
In Red Hat Virtualization 4.1, when the Manager deploys a host, collectd is always installed; however, host deployment will fail if you are attempting to deploy a new or reinstalled version 3.y host (in a cluster with 3.6 compatibility level), because collectd is not shipped in the 3.y repositories. To avoid this, ensure that you install and deploy any version 3.y hosts prior to upgrading the Manager to 4.1. Note that after the Manager upgrade, these hosts will continue to work, but you will not be able to reinstall them without first upgrading them to version 4.1.
In Red Hat Virtualization 4.1, when the Manager deploys a host, collectd is always installed; however, host deployment will fail if you are attempting to deploy a new or reinstalled version 3.y host (in a cluster with 3.6 compatibility level), because collectd is not shipped in the 3.y repositories. To avoid this, ensure that you install and deploy any version 3.y hosts prior to upgrading the Manager to 4.1. Note that after the Manager upgrade, these hosts will continue to work, but you will not be able to reinstall them without first upgrading them to version 4.1.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1441711
Cockpit is currently available only for x86_64 architecture. As a result, in Red Hat Virtualization, Cockpit is supported only for x86_64 hosts, and is not supported for ppc64le (IBM POWER8) hosts.
Cockpit is currently available only for x86_64 architecture. As a result, in Red Hat Virtualization, Cockpit is supported only for x86_64 hosts, and is not supported for ppc64le (IBM POWER8) hosts.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1442254
Copy to Clipboard Copied! Toggle word wrap Toggle overflow
3.1.5. Deprecated Functionality Copy linkLink copied to clipboard!
Copy linkLink copied to clipboard!
The items in this section are either no longer supported or will no longer be supported in a future release
- BZ#1219545
This release removes the ability to export Gluster volume profile statistics as a PDF file (a feature that was not widely used) as part of removing the dependency on the avalon-framework package.
This release removes the ability to export Gluster volume profile statistics as a PDF file (a feature that was not widely used) as part of removing the dependency on the avalon-framework package.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1350334
The rhevm-guest-agent packages for Red Hat Enterprise Linux have now been renamed to ovirt-guest-agent, to align with upstream.
The rhevm-guest-agent packages for Red Hat Enterprise Linux have now been renamed to ovirt-guest-agent, to align with upstream.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1372237
This release removes a no-longer-needed workaround for the vdsm-jsonrpc deprecation warning.
This release removes a no-longer-needed workaround for the vdsm-jsonrpc deprecation warning.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1429809
IFCFG persistence mode has been declared deprecated. The Unified persistence mode has been the default from version 3.5 and should now be used in all systems.
IFCFG persistence mode has been declared deprecated. The Unified persistence mode has been the default from version 3.5 and should now be used in all systems.Copy to Clipboard Copied! Toggle word wrap Toggle overflow
3.2. Red Hat Virtualization Batch Update 1 (ovirt-4.1.2) Copy linkLink copied to clipboard!
Copy linkLink copied to clipboard!
These release notes highlight Technology Preview items, recommended practices, known issues, and deprecated functionality to be taken into consideration when deploying this release of Red Hat Enterprise Virtualization.
Notes for updates released during the support lifecycle of this Red Hat Enterprise Virtualization release will appear in the advisory text associated with each update or the Red Hat Enterprise Virtualization Technical Notes. This document is available from the following page:
3.2.1. Enhancements Copy linkLink copied to clipboard!
Copy linkLink copied to clipboard!
This release of Red Hat Enterprise Virtualization features the following enhancements:
- BZ#1360983
With this update, the host name is automatically set to the virtual machine name in the RunOnce dialog. If required, this name can be changed.
With this update, the host name is automatically set to the virtual machine name in the RunOnce dialog. If required, this name can be changed.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1388433
Previously, the Manager's frequently updated PostgreSQL tables would fill up with obsolete data, creating the risk of disk flooding and transaction ID wraparound issues. This release has introduced a more aggressive vacuum daemon configuration for collecting obsolete rows. This enables the Manager's tables to remain healthy, and for disk space usage to be better correlated with the actual amount of kept data.
Previously, the Manager's frequently updated PostgreSQL tables would fill up with obsolete data, creating the risk of disk flooding and transaction ID wraparound issues. This release has introduced a more aggressive vacuum daemon configuration for collecting obsolete rows. This enables the Manager's tables to remain healthy, and for disk space usage to be better correlated with the actual amount of kept data.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1406398
Red Hat Virtualization Manager now supports NFS version 4.2 connections (when also supported by the storage).
Red Hat Virtualization Manager now supports NFS version 4.2 connections (when also supported by the storage).Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1421533
In this release, the vdsm-client has been added as a dependency, and it replaces the now deprecated vdsClient.
In this release, the vdsm-client has been added as a dependency, and it replaces the now deprecated vdsClient.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1427790
Previously, it was possible to limit the highest SSL/TLS protocol version that was negotiated when establishing a connection between the Manager and VDSM. This was required for older clients. In this release, this option has been removed from engine-config as it was verified that it is no longer required for VDSM 3.6 and later. VDSM 3.6 can successfully negotiate the highest available version.
Previously, it was possible to limit the highest SSL/TLS protocol version that was negotiated when establishing a connection between the Manager and VDSM. This was required for older clients. In this release, this option has been removed from engine-config as it was verified that it is no longer required for VDSM 3.6 and later. VDSM 3.6 can successfully negotiate the highest available version.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1443508
Previously, ovirt-host-deploy failed if the collectd/fluentd packages and their plug-ins were not available. In this release, if these packages are missing, ovirt-host-deploy emits a warning but does not fail. This enables a 3.6 host to be added to a 4.1 Manager, even though the 3.6 repositories do not include these packages.
Previously, ovirt-host-deploy failed if the collectd/fluentd packages and their plug-ins were not available. In this release, if these packages are missing, ovirt-host-deploy emits a warning but does not fail. This enables a 3.6 host to be added to a 4.1 Manager, even though the 3.6 repositories do not include these packages.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1446056
Previously, after running engine-upgrade-check, the user was not informed that the system may not be up-to-date if engine-setup was not run after running yum update, for example. This is despite engine-upgrade-check stating that no upgrade is available. This has now been fixed so that a warning message is displayed to the user if engine-setup was not run.
Previously, after running engine-upgrade-check, the user was not informed that the system may not be up-to-date if engine-setup was not run after running yum update, for example. This is despite engine-upgrade-check stating that no upgrade is available. This has now been fixed so that a warning message is displayed to the user if engine-setup was not run.Copy to Clipboard Copied! Toggle word wrap Toggle overflow
3.2.2. Release Notes Copy linkLink copied to clipboard!
Copy linkLink copied to clipboard!
This section outlines important details about the release, including recommended practices and notable changes to Red Hat Enterprise Virtualization. You must take this information into account to ensure the best possible outcomes for your deployment.
- BZ#1367107
Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1433961
Previously, it was possible to configure memory overcommit without setting memory ballooning or KSM control. This configuration affected the scheduling, but the memory was not freed. This has now been fixed by disabling ballooning and KSM by default, and setting memory optimization to “None” (100%).
Previously, it was possible to configure memory overcommit without setting memory ballooning or KSM control. This configuration affected the scheduling, but the memory was not freed. This has now been fixed by disabling ballooning and KSM by default, and setting memory optimization to “None” (100%).Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1436161
The recommended number of logical volumes per storage domain was increased from 300 to 1000. The alert that is issued when exceeding this number was changed to reflect that. Note that the limit will only change to 1000 if it was set to 300. If the customer manually configured the limit, the original configuration will not be overridden.
The recommended number of logical volumes per storage domain was increased from 300 to 1000. The alert that is issued when exceeding this number was changed to reflect that. Note that the limit will only change to 1000 if it was set to 300. If the customer manually configured the limit, the original configuration will not be overridden.Copy to Clipboard Copied! Toggle word wrap Toggle overflow
3.3. Red Hat Virtualization Batch Update 2 (ovirt-4.1.3) Copy linkLink copied to clipboard!
Copy linkLink copied to clipboard!
3.3.1. Enhancements Copy linkLink copied to clipboard!
Copy linkLink copied to clipboard!
This release of Red Hat Virtualization features the following enhancements:
- BZ#1408825
With this update, the ability to unplug a lease from and plug another lease into a running virtual machine has been added using new APIs. This provides the ability to move a virtual machine lease from one storage domain to another so that the original storage domain can be placed into maintenance.
With this update, the ability to unplug a lease from and plug another lease into a running virtual machine has been added using new APIs. This provides the ability to move a virtual machine lease from one storage domain to another so that the original storage domain can be placed into maintenance.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1418659
With this update, engine.log has been added to the logs collection. The records are now parsed and sent to the central metrics store. This allows the administrator to analyse the logs in a more simple and comfortable way. This means that the engine.log records are available for analysis in Kibana.
With this update, engine.log has been added to the logs collection. The records are now parsed and sent to the central metrics store. This allows the administrator to analyse the logs in a more simple and comfortable way. This means that the engine.log records are available for analysis in Kibana.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1421204
Previous versions of Red Hat Virtualization were able to negotiate encryption protocol up to TLSv1 for external provider communication. This update adds the the ability to negotiate encryption protocol up to TLSv1.2. Note: The exact version used for communication depends on highest version available on the external provider target.
Previous versions of Red Hat Virtualization were able to negotiate encryption protocol up to TLSv1 for external provider communication. This update adds the the ability to negotiate encryption protocol up to TLSv1.2. Note: The exact version used for communication depends on highest version available on the external provider target.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1429861
With this update, ovirt-engine-metrics can now configure collectd to connect to the Red Hat Virtualization Manager's PostgreSQL database and get information from it. Previously, making collectd connect to the PostgreSQL database did not work and was disabled by BZ#1436001.
With this update, ovirt-engine-metrics can now configure collectd to connect to the Red Hat Virtualization Manager's PostgreSQL database and get information from it. Previously, making collectd connect to the PostgreSQL database did not work and was disabled by BZ#1436001.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1431545
Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1436981
The Python SDK now supports asynchronous requests and HTTP pipe-lining. Users can send requests asynchronously and wait for the response later in code. This makes it possible to send multiple requests using multiple connections or pipelined connections, and wait for the response later, which improves performance when fetching multiple objects from the API.
The Python SDK now supports asynchronous requests and HTTP pipe-lining. Users can send requests asynchronously and wait for the response later in code. This makes it possible to send multiple requests using multiple connections or pipelined connections, and wait for the response later, which improves performance when fetching multiple objects from the API.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1439281
The self-hosted engine Manager virtual machine upgrade procedure from Red Hat Enterprise Virtualization 3.6 on Red Hat Enterprise Linux 6 to Red Hat Virtualization 4.0 on Red Hat Enterprise Linux 7 now always tries to connect to the Manager API as admin@internal, making it more flexible.
The self-hosted engine Manager virtual machine upgrade procedure from Red Hat Enterprise Virtualization 3.6 on Red Hat Enterprise Linux 6 to Red Hat Virtualization 4.0 on Red Hat Enterprise Linux 7 now always tries to connect to the Manager API as admin@internal, making it more flexible.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1444992
Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1445631
With this update, the user can now control the location of the temporary directory created by the engine-backup command, using the command line option --tmpdir=DIR and the variable $TMPDIR. Previously, the engine-backup command always used /tmp. If /tmp was full, or close to being full, it was not possible to tell the engine-backup command to use a different directory.
With this update, the user can now control the location of the temporary directory created by the engine-backup command, using the command line option --tmpdir=DIR and the variable $TMPDIR. Previously, the engine-backup command always used /tmp. If /tmp was full, or close to being full, it was not possible to tell the engine-backup command to use a different directory.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1447887
Red Hat Virtualization Host (RHVH) now supports NIST 800-53 partitioning requirements to improve security. RHVH uses a NIST 800-53 partition layout by default, and existing configurations will be changed on update.
Red Hat Virtualization Host (RHVH) now supports NIST 800-53 partitioning requirements to improve security. RHVH uses a NIST 800-53 partition layout by default, and existing configurations will be changed on update.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1450646
This update makes it easier to debug a disk snapshot's live deletion failures because the initial state of the volume chain is logged.
This update makes it easier to debug a disk snapshot's live deletion failures because the initial state of the volume chain is logged.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1455771
When collecting SOS reports from hosts, chrony and systemd SOS plugins can collect information about time synchronization. In addition, a new option --time-only has been added to ovirt-log-collector allowing information about time differences to be gathered from the hosts without gathering full SOS reports, saving a considerable amount of time for the operation.
When collecting SOS reports from hosts, chrony and systemd SOS plugins can collect information about time synchronization. In addition, a new option --time-only has been added to ovirt-log-collector allowing information about time differences to be gathered from the hosts without gathering full SOS reports, saving a considerable amount of time for the operation.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1456877
Previously, the self-hosted engine setup failed if the system clock was set to a time zone that was behind UTC during the installation. This is due to the fact that the Red Hat Virtualization Host generates VDSM certificates at the first boot and if the clock is incorrect, the chronyd or ntpd processes resynchronized the clock. This lead to an invalid certificate if the time zone was behind UTC. Now, Red Hat Virtualization Host generates the certificates after the chronyd or ntpd processes and waits two seconds for the clock to synchronize. Note that if the Red Hat Virtualization Host is configured after the installation, or if the NTP server is too slow, the self-hosted engine setup may fail.
Previously, the self-hosted engine setup failed if the system clock was set to a time zone that was behind UTC during the installation. This is due to the fact that the Red Hat Virtualization Host generates VDSM certificates at the first boot and if the clock is incorrect, the chronyd or ntpd processes resynchronized the clock. This lead to an invalid certificate if the time zone was behind UTC. Now, Red Hat Virtualization Host generates the certificates after the chronyd or ntpd processes and waits two seconds for the clock to synchronize. Note that if the Red Hat Virtualization Host is configured after the installation, or if the NTP server is too slow, the self-hosted engine setup may fail.Copy to Clipboard Copied! Toggle word wrap Toggle overflow
3.3.2. Technology Preview Copy linkLink copied to clipboard!
Copy linkLink copied to clipboard!
The items listed in this section are provided as Technology Previews. For further information on the scope of Technology Preview status, and the associated support implications, refer to https://access.redhat.com/support/offerings/techpreview/.
- BZ#1456568
This release introduces a new VM Portal as a Technology Preview. A "VM Portal" link is now available on the Red Hat Virtualization Welcome Page. The VM Portal provides the same functionality that is available in the Basic tab of the current User Portal (now deprecated).
This release introduces a new VM Portal as a Technology Preview. A "VM Portal" link is now available on the Red Hat Virtualization Welcome Page. The VM Portal provides the same functionality that is available in the Basic tab of the current User Portal (now deprecated).Copy to Clipboard Copied! Toggle word wrap Toggle overflow
3.3.3. Known Issues Copy linkLink copied to clipboard!
Copy linkLink copied to clipboard!
These known issues exist in Red Hat Virtualization at this time:
- BZ#1455441
Previously, the self-hosted engine setup failed if the system clock was set to a time zone that was behind UTC during the installation. This is due to the fact that the Red Hat Virtualization Host generates VDSM certificates at the first boot and if the clock is incorrect, the chronyd or ntpd processes resynchronized the clock. This lead to an invalid certificate if the time zone was behind UTC. Now, Red Hat Virtualization Host generates the certificates after the chronyd or ntpd processes and waits two seconds for the clock to synchronize. Note that if the Red Hat Virtualization Host is configured after the installation, or if the NTP server is too slow, the self-hosted engine setup may fail.
Previously, the self-hosted engine setup failed if the system clock was set to a time zone that was behind UTC during the installation. This is due to the fact that the Red Hat Virtualization Host generates VDSM certificates at the first boot and if the clock is incorrect, the chronyd or ntpd processes resynchronized the clock. This lead to an invalid certificate if the time zone was behind UTC. Now, Red Hat Virtualization Host generates the certificates after the chronyd or ntpd processes and waits two seconds for the clock to synchronize. Note that if the Red Hat Virtualization Host is configured after the installation, or if the NTP server is too slow, the self-hosted engine setup may fail.Copy to Clipboard Copied! Toggle word wrap Toggle overflow
3.3.4. Deprecated Functionality Copy linkLink copied to clipboard!
Copy linkLink copied to clipboard!
The items in this section are either no longer supported or will no longer be supported in a future release
- BZ#1441632
With this update, the migration of a virtual machine to a different cluster can no longer be invoked from the UI. Regular migration within a cluster remains unchanged.
With this update, the migration of a virtual machine to a different cluster can no longer be invoked from the UI. Regular migration within a cluster remains unchanged.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1464113
The User Portal is now deprecated. From Red Hat Virtualization 4.2, it will no longer be available. A new VM Portal will be added, but certain features may not be present at the time of release. A technology preview of the VM Portal is available with the 4.1.3 release. Customers utilizing the Red Hat CloudForms Self Service user interface to manage virtual machines will not be affected.
The User Portal is now deprecated. From Red Hat Virtualization 4.2, it will no longer be available. A new VM Portal will be added, but certain features may not be present at the time of release. A technology preview of the VM Portal is available with the 4.1.3 release. Customers utilizing the Red Hat CloudForms Self Service user interface to manage virtual machines will not be affected.Copy to Clipboard Copied! Toggle word wrap Toggle overflow
3.4. Red Hat Virtualization Batch Update 3 (ovirt-4.1.4) Copy linkLink copied to clipboard!
Copy linkLink copied to clipboard!
3.4.1. Enhancements Copy linkLink copied to clipboard!
Copy linkLink copied to clipboard!
This release of Red Hat Virtualization features the following enhancements:
- BZ#917062
ABRT is now integrated with Red Hat Virtualization when initializing hypervisors. ABRT is configured by VDSM and saves meaningful debug information, instead of enormous core-dump reports.
ABRT is now integrated with Red Hat Virtualization when initializing hypervisors. ABRT is configured by VDSM and saves meaningful debug information, instead of enormous core-dump reports.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1438408
The ClusterInMaintenance scheduling policy was added to limit activity in a cluster during maintenance tasks. When set, no new VMs may be started, except highly available VMs. Highly available VMs are still restarted properly upon host failure and any VM can be migrated. Users can create new highly available VMs and start them manually.
The ClusterInMaintenance scheduling policy was added to limit activity in a cluster during maintenance tasks. When set, no new VMs may be started, except highly available VMs. Highly available VMs are still restarted properly upon host failure and any VM can be migrated. Users can create new highly available VMs and start them manually.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1450293
Newer browsers require the subjectAltName extension in certificates for HTTPS. Previously (version 4.1.2 and later), engine-setup created certificates with subjectAltName for new setups, but did not update existing certificates during upgrade. Now engine-setup can add subjectAltName to existing internal certificates, so that newer browsers will accept them.
Newer browsers require the subjectAltName extension in certificates for HTTPS. Previously (version 4.1.2 and later), engine-setup created certificates with subjectAltName for new setups, but did not update existing certificates during upgrade. Now engine-setup can add subjectAltName to existing internal certificates, so that newer browsers will accept them.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1452668
Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1461813
RHV-H now supports vGPU for nVidia GPUs, by including vdsm-hook-vfio-mdev.
RHV-H now supports vGPU for nVidia GPUs, by including vdsm-hook-vfio-mdev.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1465862
Previously, users could use the REST API to add/edit affinity labels for virtual machines and hosts, but they could not view this information in the Administration Portal. In this release, affinity label information is now available in the Administration Portal. It can be seen in the Add/Edit windows for virtual machines and hosts. The window displays all available affinity labels in the system, and for entities with assigned labels, those labels are selected.
Previously, users could use the REST API to add/edit affinity labels for virtual machines and hosts, but they could not view this information in the Administration Portal. In this release, affinity label information is now available in the Administration Portal. It can be seen in the Add/Edit windows for virtual machines and hosts. The window displays all available affinity labels in the system, and for entities with assigned labels, those labels are selected.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1467058
Previously, when creating a new VLAN-tagged network, the Manager did not allow the same VLAN ID to be duplicated across various entities, such as networks and data centers. In this release, it is now possible to use the same VLAN ID across multiple entities, as long as they are not used on the same interface on the hardware node.
Previously, when creating a new VLAN-tagged network, the Manager did not allow the same VLAN ID to be duplicated across various entities, such as networks and data centers. In this release, it is now possible to use the same VLAN ID across multiple entities, as long as they are not used on the same interface on the hardware node.Copy to Clipboard Copied! Toggle word wrap Toggle overflow
3.4.2. Release Notes Copy linkLink copied to clipboard!
Copy linkLink copied to clipboard!
This section outlines important details about the release, including recommended practices and notable changes to Red Hat Virtualization. You must take this information into account to ensure the best possible outcomes for your deployment.
- BZ#1430285
In order to run VMs on hosts with limited memory resources, the cluster scheduling policy should be set to "evenly_distributed", with "maxFreeMemoryForOverUtilized = 99". This configuration enables the engine to schedule VMs on hosts with more free memory, for better distribution.
In order to run VMs on hosts with limited memory resources, the cluster scheduling policy should be set to "evenly_distributed", with "maxFreeMemoryForOverUtilized = 99". This configuration enables the engine to schedule VMs on hosts with more free memory, for better distribution.Copy to Clipboard Copied! Toggle word wrap Toggle overflow
3.5. Red Hat Virtualization Batch Update 4 (ovirt-4.1.5) Copy linkLink copied to clipboard!
Copy linkLink copied to clipboard!
3.5.1. Enhancements Copy linkLink copied to clipboard!
Copy linkLink copied to clipboard!
This release of Red Hat Virtualization features the following enhancements:
- BZ#1022961
This release adds libgfapi support to the Manager and VDSM. libgfapi provides virtual machines with faster access to their images, stored on a Gluster volume, compared to a fuse interface. With the 'LibgfApi' data center feature enabled, or 'lubgfapi_supported' cluster-level feature enabled, virtual machines will access their images, stored on Gluster volumes, directly via libgfapi.
This release adds libgfapi support to the Manager and VDSM. libgfapi provides virtual machines with faster access to their images, stored on a Gluster volume, compared to a fuse interface. With the 'LibgfApi' data center feature enabled, or 'lubgfapi_supported' cluster-level feature enabled, virtual machines will access their images, stored on Gluster volumes, directly via libgfapi.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1364083
With this update, oVirt Host related packages are now also available for ppc64le architecture.
With this update, oVirt Host related packages are now also available for ppc64le architecture.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1412552
VDSM now supports TLSv1.1 and TLSv1.2, enabling all clients to leverage these new encryption protocols, depending on their capabilities.
VDSM now supports TLSv1.1 and TLSv1.2, enabling all clients to leverage these new encryption protocols, depending on their capabilities.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1451490
This release adds the list of required collectd and fluentd packages to the metrics script, and installs them if missing. Previously, there were cases where the packages were available in the repository and required for the metrics setup script, but were not installed, causing the configuration to fail. Now, the missing packages are installed.
This release adds the list of required collectd and fluentd packages to the metrics script, and installs them if missing. Previously, there were cases where the packages were available in the repository and required for the metrics setup script, but were not installed, causing the configuration to fail. Now, the missing packages are installed.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1454827
Previously, Red Hat Virtualization supported TLS version 1.0. From this release, all components can use TLS version 1.2 for encrypted communication, providing that it is used by all components.
Previously, Red Hat Virtualization supported TLS version 1.0. From this release, all components can use TLS version 1.2 for encrypted communication, providing that it is used by all components.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1457357
Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1462500
This release adds checks to the ovirt_env_name, to make sure it is a valid OpenShift namespace identifier. If it is not, the metrics script will now fail. The name must be valid in order for the records to be kept by the Elasticsearch running in OpenShift.
This release adds checks to the ovirt_env_name, to make sure it is a valid OpenShift namespace identifier. If it is not, the metrics script will now fail. The name must be valid in order for the records to be kept by the Elasticsearch running in OpenShift.Copy to Clipboard Copied! Toggle word wrap Toggle overflow
3.5.2. Release Notes Copy linkLink copied to clipboard!
Copy linkLink copied to clipboard!
This section outlines important details about the release, including recommended practices and notable changes to Red Hat Virtualization. You must take this information into account to ensure the best possible outcomes for your deployment.
- BZ#1472047
The batik package is no longer required by the Red Hat Virtualization Manager. Red Hat recommends removing it after upgrading to Red Hat Virtualization 4.1.
The batik package is no longer required by the Red Hat Virtualization Manager. Red Hat recommends removing it after upgrading to Red Hat Virtualization 4.1.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1472048
The fop package is no longer required by the Red Hat Virtualization Manager. Red Hat recommends removing it after upgrading to Red Hat Virtualization 4.1.
The fop package is no longer required by the Red Hat Virtualization Manager. Red Hat recommends removing it after upgrading to Red Hat Virtualization 4.1.Copy to Clipboard Copied! Toggle word wrap Toggle overflow
3.5.3. Known Issues Copy linkLink copied to clipboard!
Copy linkLink copied to clipboard!
These known issues exist in Red Hat Virtualization at this time:
- BZ#1351211
When removing a Gluster host after moving it to maintenance mode, the host is not removed from the peer list of other hosts. To work around this issue, do not stop the Gluster services while moving the host to maintenance mode if the host is going to be removed from cluster.
When removing a Gluster host after moving it to maintenance mode, the host is not removed from the peer list of other hosts. To work around this issue, do not stop the Gluster services while moving the host to maintenance mode if the host is going to be removed from cluster.Copy to Clipboard Copied! Toggle word wrap Toggle overflow
3.5.4. Deprecated Functionality Copy linkLink copied to clipboard!
Copy linkLink copied to clipboard!
The items in this section are either no longer supported or will no longer be supported in a future release
- BZ#1473179
The ovirt-engine-setup-plugin-dockerc package is now deprecated, and will be removed in version 4.2.
The ovirt-engine-setup-plugin-dockerc package is now deprecated, and will be removed in version 4.2.Copy to Clipboard Copied! Toggle word wrap Toggle overflow
3.6. Red Hat Virtualization Batch Update 5 (ovirt-4.1.6) Copy linkLink copied to clipboard!
Copy linkLink copied to clipboard!
3.6.1. Enhancements Copy linkLink copied to clipboard!
Copy linkLink copied to clipboard!
This release of Red Hat Virtualization features the following enhancements:
- BZ#1472254
Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1473535
The ovirt-ansible-roles package contains Ansible roles, which can help administrators with common Red Hat Virtualization administration tasks. All roles can be executed from the command line using Ansible, but some of those roles can also be executed directly from the Red Hat Virtualization Manager. More details about the roles can be found in README.md included in the package (/usr/share/doc/ovirt-ansible-roles/README.md), or directly in the source code repository (https://github.com/ovirt/ovirt-ansible).
The ovirt-ansible-roles package contains Ansible roles, which can help administrators with common Red Hat Virtualization administration tasks. All roles can be executed from the command line using Ansible, but some of those roles can also be executed directly from the Red Hat Virtualization Manager. More details about the roles can be found in README.md included in the package (/usr/share/doc/ovirt-ansible-roles/README.md), or directly in the source code repository (https://github.com/ovirt/ovirt-ansible).Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1477053
This update lists the additional features available for a cluster level, and allows you to enable or disable additional cluster features using an API.
This update lists the additional features available for a cluster level, and allows you to enable or disable additional cluster features using an API.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1479677
Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1481095
The user now has the option of specifying a path to the appliance OVF, as an alternative to installing the appliance RPM.
The user now has the option of specifying a path to the appliance OVF, as an alternative to installing the appliance RPM.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1482104
The ovirt-ansible-roles package contains Ansible roles, which can help administrators with common Red Hat Virtualization administration tasks. All roles can be executed from the command line using Ansible, but some of those roles can also be executed directly from the Red Hat Virtualization Manager. More details about the roles can be found in README.md included in the package (/usr/share/doc/ovirt-ansible-roles/README.md), or directly in the source code repository (https://github.com/ovirt/ovirt-ansible). Please note that currently Ansible has the following bug: https://bugzilla.redhat.com/show_bug.cgi?id=1487113 Because of this bug you should avoid calling multiple oVirt Ansible roles in a single playbook, or calling oVirt Ansible modules prior to calling oVirt Ansible roles in a single playbook (the working solution is to call the modules from the pre_tasks).
The ovirt-ansible-roles package contains Ansible roles, which can help administrators with common Red Hat Virtualization administration tasks. All roles can be executed from the command line using Ansible, but some of those roles can also be executed directly from the Red Hat Virtualization Manager. More details about the roles can be found in README.md included in the package (/usr/share/doc/ovirt-ansible-roles/README.md), or directly in the source code repository (https://github.com/ovirt/ovirt-ansible). Please note that currently Ansible has the following bug: https://bugzilla.redhat.com/show_bug.cgi?id=1487113 Because of this bug you should avoid calling multiple oVirt Ansible roles in a single playbook, or calling oVirt Ansible modules prior to calling oVirt Ansible roles in a single playbook (the working solution is to call the modules from the pre_tasks).Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1486573
This Ansible role can be used to automate upgrading the whole cluster to a higher version. More information can be found at https://github.com/oVirt/ovirt-ansible/blob/ovirt-ansible-1.0/roles/ovirt-cluster-upgrade/README.md Please note that currently Ansible has the following bug: https://bugzilla.redhat.com/show_bug.cgi?id=1487113 Because of this bug you should avoid calling multiple oVirt Ansible roles in a single playbook, or calling oVirt Ansible modules prior to calling oVirt Ansible roles in a single playbook (the working solution is to call the modules from the pre_tasks).
This Ansible role can be used to automate upgrading the whole cluster to a higher version. More information can be found at https://github.com/oVirt/ovirt-ansible/blob/ovirt-ansible-1.0/roles/ovirt-cluster-upgrade/README.md Please note that currently Ansible has the following bug: https://bugzilla.redhat.com/show_bug.cgi?id=1487113 Because of this bug you should avoid calling multiple oVirt Ansible roles in a single playbook, or calling oVirt Ansible modules prior to calling oVirt Ansible roles in a single playbook (the working solution is to call the modules from the pre_tasks).Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1486574
This Ansible role can be used to automate data center setup. More information can be found at https://github.com/oVirt/ovirt-ansible/blob/ovirt-ansible-1.0/roles/ovirt-infra/README.md Please note that currently Ansible has the following bug: https://bugzilla.redhat.com/show_bug.cgi?id=1487113 Because of this bug you should avoid calling multiple oVirt Ansible roles in a single playbook, or calling oVirt Ansible modules prior to calling oVirt Ansible roles in a single playbook (the working solution is to call the modules from the pre_tasks).
This Ansible role can be used to automate data center setup. More information can be found at https://github.com/oVirt/ovirt-ansible/blob/ovirt-ansible-1.0/roles/ovirt-infra/README.md Please note that currently Ansible has the following bug: https://bugzilla.redhat.com/show_bug.cgi?id=1487113 Because of this bug you should avoid calling multiple oVirt Ansible roles in a single playbook, or calling oVirt Ansible modules prior to calling oVirt Ansible roles in a single playbook (the working solution is to call the modules from the pre_tasks).Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1486575
This Ansible role can be used to automate creating a VM template from an external image. More information can be found at https://github.com/oVirt/ovirt-ansible/blob/ovirt-ansible-1.0/roles/ovirt-image-template/README.md Please note that currently Ansible has the following bug: https://bugzilla.redhat.com/show_bug.cgi?id=1487113 Because of this bug you should avoid calling multiple oVirt Ansible roles in a single playbook, or calling oVirt Ansible modules prior to calling oVirt Ansible roles in a single playbook (the working solution is to call the modules from the pre_tasks).
This Ansible role can be used to automate creating a VM template from an external image. More information can be found at https://github.com/oVirt/ovirt-ansible/blob/ovirt-ansible-1.0/roles/ovirt-image-template/README.md Please note that currently Ansible has the following bug: https://bugzilla.redhat.com/show_bug.cgi?id=1487113 Because of this bug you should avoid calling multiple oVirt Ansible roles in a single playbook, or calling oVirt Ansible modules prior to calling oVirt Ansible roles in a single playbook (the working solution is to call the modules from the pre_tasks).Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1486576
This Ansible role can be used to download a ManageIQ qcow2 image, create a VM from it, and register a provider within the ManageIQ installation. More information can be found at https://github.com/oVirt/ovirt-ansible/blob/ovirt-ansible-1.0/roles/ovirt-manageiq/README.md Please note that currently Ansible has the following bug: https://bugzilla.redhat.com/show_bug.cgi?id=1487113 Because of this bug you should avoid calling multiple oVirt Ansible roles in a single playbook, or calling oVirt Ansible modules prior to calling oVirt Ansible roles in a single playbook (the working solution is to call the modules from the pre_tasks).
This Ansible role can be used to download a ManageIQ qcow2 image, create a VM from it, and register a provider within the ManageIQ installation. More information can be found at https://github.com/oVirt/ovirt-ansible/blob/ovirt-ansible-1.0/roles/ovirt-manageiq/README.md Please note that currently Ansible has the following bug: https://bugzilla.redhat.com/show_bug.cgi?id=1487113 Because of this bug you should avoid calling multiple oVirt Ansible roles in a single playbook, or calling oVirt Ansible modules prior to calling oVirt Ansible roles in a single playbook (the working solution is to call the modules from the pre_tasks).Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1486577
This Ansible role can be used to automate creating a virtual machine. More information can be found at https://github.com/oVirt/ovirt-ansible/blob/ovirt-ansible-1.0/roles/ovirt-vm-infra/README.md Please note that currently Ansible has the following bug: https://bugzilla.redhat.com/show_bug.cgi?id=1487113 Because of this bug you should avoid calling multiple oVirt Ansible roles in a single playbook, or calling oVirt Ansible modules prior to calling oVirt Ansible roles in a single playbook (the working solution is to call the modules from the pre_tasks).
This Ansible role can be used to automate creating a virtual machine. More information can be found at https://github.com/oVirt/ovirt-ansible/blob/ovirt-ansible-1.0/roles/ovirt-vm-infra/README.md Please note that currently Ansible has the following bug: https://bugzilla.redhat.com/show_bug.cgi?id=1487113 Because of this bug you should avoid calling multiple oVirt Ansible roles in a single playbook, or calling oVirt Ansible modules prior to calling oVirt Ansible roles in a single playbook (the working solution is to call the modules from the pre_tasks).Copy to Clipboard Copied! Toggle word wrap Toggle overflow
3.6.2. Release Notes Copy linkLink copied to clipboard!
Copy linkLink copied to clipboard!
This section outlines important details about the release, including recommended practices and notable changes to Red Hat Virtualization. You must take this information into account to ensure the best possible outcomes for your deployment.
- BZ#1471615
This release includes a new Spice QXL WDDM DOD driver (Display-only driver for QXL virtual devices) for Windows 10 guests.
This release includes a new Spice QXL WDDM DOD driver (Display-only driver for QXL virtual devices) for Windows 10 guests.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1472047
The batik package is no longer required by the Red Hat Virtualization Manager. Red Hat recommends removing it after upgrading to Red Hat Virtualization 4.1.
The batik package is no longer required by the Red Hat Virtualization Manager. Red Hat recommends removing it after upgrading to Red Hat Virtualization 4.1.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1472048
The fop package is no longer required by the Red Hat Virtualization Manager. Red Hat recommends removing it after upgrading to Red Hat Virtualization 4.1.
The fop package is no longer required by the Red Hat Virtualization Manager. Red Hat recommends removing it after upgrading to Red Hat Virtualization 4.1.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1481150
A new VDSM hook that configures the nested virtualization feature in Red Hat Enterprise Linux 7 is now available as a tech preview. Nested virtualization allows a virtual machine to serve as a hypervisor. VDSM hooks are a means to insert code, commands, or scripts into a point in the lifecycle of a virtual machine, as well as the lifecycle of the VDSM daemon. Packages and features released under tech preview are not fully supported under Red Hat Subscription Level Agreements, and may not be functionally complete, and are not intended for production use.
A new VDSM hook that configures the nested virtualization feature in Red Hat Enterprise Linux 7 is now available as a tech preview. Nested virtualization allows a virtual machine to serve as a hypervisor. VDSM hooks are a means to insert code, commands, or scripts into a point in the lifecycle of a virtual machine, as well as the lifecycle of the VDSM daemon. Packages and features released under tech preview are not fully supported under Red Hat Subscription Level Agreements, and may not be functionally complete, and are not intended for production use.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1481212
Copy to Clipboard Copied! Toggle word wrap Toggle overflow
3.6.3. Known Issues Copy linkLink copied to clipboard!
Copy linkLink copied to clipboard!
These known issues exist in Red Hat Virtualization at this time:
- BZ#1478865
Copy to Clipboard Copied! Toggle word wrap Toggle overflow
3.7. Red Hat Virtualization Batch Update 6 (ovirt-4.1.7) Copy linkLink copied to clipboard!
Copy linkLink copied to clipboard!
These release notes highlight technology preview items, recommended practices, known issues, and deprecated functionality to be taken into consideration when deploying this release of Red Hat Enterprise Virtualization.
Notes for updates released during the support lifecycle of this Red Hat Enterprise Virtualization release will appear in the advisory text associated with each update or the Red Hat Enterprise Virtualization Technical Notes. This document is available from the following page:
3.7.1. Enhancements Copy linkLink copied to clipboard!
Copy linkLink copied to clipboard!
This release of Red Hat Enterprise Virtualization features the following enhancements:
- BZ#1490810
Previously, while VDSM performed periodic monitoring and maintenance tasks, operations would occasionally become too slow or even become blocked. In this case, the internal queue became full and the periodic operations were no longer performed. A "TooManyTasks" warning appeared in the log files at a maximum rate of once every 10 seconds. In this release, if VDSM cannot perform its periodic operations, in addition to issuing a warning in the log files, VDSM also dumps the contents of the queue into the logs.
Previously, while VDSM performed periodic monitoring and maintenance tasks, operations would occasionally become too slow or even become blocked. In this case, the internal queue became full and the periodic operations were no longer performed. A "TooManyTasks" warning appeared in the log files at a maximum rate of once every 10 seconds. In this release, if VDSM cannot perform its periodic operations, in addition to issuing a warning in the log files, VDSM also dumps the contents of the queue into the logs.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1502510
The precision of rx_rate, tx_rate, rx_drop, and tx_drop of virtual and host network interfaces have been increased. Network traffic 100 times smaller can now be detected on network interface statistics. If traffic on the network interface is below the precision of the network interface statistics, it is not reflected in the statistics.
The precision of rx_rate, tx_rate, rx_drop, and tx_drop of virtual and host network interfaces have been increased. Network traffic 100 times smaller can now be detected on network interface statistics. If traffic on the network interface is below the precision of the network interface statistics, it is not reflected in the statistics.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1502653
This release adds caching of the OVF storage location. As the OVF storage location rarely changes, it does not need to be searched for on every monitoring loop iteration. Instead it can be saved and reused, and expired only in the case of an error. As a result, the monitoring loop execution time is decreased significantly.
This release adds caching of the OVF storage location. As the OVF storage location rarely changes, it does not need to be searched for on every monitoring loop iteration. Instead it can be saved and reused, and expired only in the case of an error. As a result, the monitoring loop execution time is decreased significantly.Copy to Clipboard Copied! Toggle word wrap Toggle overflow
3.7.2. Known Issues Copy linkLink copied to clipboard!
Copy linkLink copied to clipboard!
These known issues exist in Red Hat Enterprise Virtualization at this time:
- BZ#1399766
If no self-hosted engine host is available other than the host that is currently running the self-hosted engine, you will not be able to move the host that runs the Manager virtual machine to maintenance mode. Furthermore, even after another host moves to a status of "Up", it could take a few minutes for it to receive a score that will enable it to run the Manager virtual machine. The host will remain in a "preparing for maintenance" state until the Manager virtual machine can migrate to another host.
If no self-hosted engine host is available other than the host that is currently running the self-hosted engine, you will not be able to move the host that runs the Manager virtual machine to maintenance mode. Furthermore, even after another host moves to a status of "Up", it could take a few minutes for it to receive a score that will enable it to run the Manager virtual machine. The host will remain in a "preparing for maintenance" state until the Manager virtual machine can migrate to another host.Copy to Clipboard Copied! Toggle word wrap Toggle overflow
3.8. Red Hat Virtualization Batch Update 7 (ovirt-4.1.8) Copy linkLink copied to clipboard!
Copy linkLink copied to clipboard!
3.8.1. Enhancements Copy linkLink copied to clipboard!
Copy linkLink copied to clipboard!
This release of Red Hat Virtualization features the following enhancements:
- BZ#1501793
There are several cluster and host settings which require reinstallation of the host if changed. The requirement to reinstall was always mentioned in documentation and a WARNING event was raised. With this release, the Administration Portal now also shows an exclamation mark icon for each host that needs to be reinstalled. When an exclamation mark icon is shown, you can find the details about it in the Action Items section of the host's details view.
There are several cluster and host settings which require reinstallation of the host if changed. The requirement to reinstall was always mentioned in documentation and a WARNING event was raised. With this release, the Administration Portal now also shows an exclamation mark icon for each host that needs to be reinstalled. When an exclamation mark icon is shown, you can find the details about it in the Action Items section of the host's details view.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1504691
In this release, a refresh button has been added to the Dashboard tab in the Administration Portal to enable users to view up-to-date system summary information.
In this release, a refresh button has been added to the Dashboard tab in the Administration Portal to enable users to view up-to-date system summary information.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1516194
The ipa-client package is now installed on hosts, and is included in the Red Hat Virtualization Host image. This enables Cockpit certificate signing and SSO with Red Hat IdM, and adding the host to an IdM realm.
The ipa-client package is now installed on hosts, and is included in the Red Hat Virtualization Host image. This enables Cockpit certificate signing and SSO with Red Hat IdM, and adding the host to an IdM realm.Copy to Clipboard Copied! Toggle word wrap Toggle overflow
3.8.2. Release Notes Copy linkLink copied to clipboard!
Copy linkLink copied to clipboard!
This section outlines important details about the release, including recommended practices and notable changes to Red Hat Virtualization. You must take this information into account to ensure the best possible outcomes for your deployment.
- BZ#1513886
This update ensures that TLSv12 support is enabled by default and no manual configuration is required.
This update ensures that TLSv12 support is enabled by default and no manual configuration is required.Copy to Clipboard Copied! Toggle word wrap Toggle overflow
3.8.3. Known Issues Copy linkLink copied to clipboard!
Copy linkLink copied to clipboard!
These known issues exist in Red Hat Virtualization at this time:
- BZ#1455441
Previously, the self-hosted engine setup failed if the system clock was set to a time zone that was behind UTC during the installation. This is due to the fact that the Red Hat Virtualization Host generates VDSM certificates at the first boot and if the clock is incorrect, the chronyd or ntpd processes resynchronized the clock. This lead to an invalid certificate if the time zone was behind UTC. Now, Red Hat Virtualization Host generates the certificates after the chronyd or ntpd processes, and waits two seconds for the clock to synchronize. Note that if the Red Hat Virtualization Host is configured after the installation, or if the NTP server is too slow, the self-hosted engine setup may fail.
Previously, the self-hosted engine setup failed if the system clock was set to a time zone that was behind UTC during the installation. This is due to the fact that the Red Hat Virtualization Host generates VDSM certificates at the first boot and if the clock is incorrect, the chronyd or ntpd processes resynchronized the clock. This lead to an invalid certificate if the time zone was behind UTC. Now, Red Hat Virtualization Host generates the certificates after the chronyd or ntpd processes, and waits two seconds for the clock to synchronize. Note that if the Red Hat Virtualization Host is configured after the installation, or if the NTP server is too slow, the self-hosted engine setup may fail.Copy to Clipboard Copied! Toggle word wrap Toggle overflow
3.9. Red Hat Virtualization Batch Update 8 (ovirt-4.1.9) Copy linkLink copied to clipboard!
Copy linkLink copied to clipboard!
3.9.1. Enhancements Copy linkLink copied to clipboard!
Copy linkLink copied to clipboard!
This release of Red Hat Virtualization features the following enhancements:
- BZ#1484404
You can now use key=key=value format (where the value of a fence option key is itself in a key=value format) when setting fence agent options, in order to support old drac5 devices that require "ssh_options=-oCiphers=+3des-cbc" in the Options field to enable no-longer-enabled-insecure ciphers for SSH connection.
You can now use key=key=value format (where the value of a fence option key is itself in a key=value format) when setting fence agent options, in order to support old drac5 devices that require "ssh_options=-oCiphers=+3des-cbc" in the Options field to enable no-longer-enabled-insecure ciphers for SSH connection.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1518545
otopi now optionally logs the list of network connections on the machine after failures. Enable this by installing the optional otopi-debug-plugins package. This can help debug failures to start network services due to "Address already in use" errors.
otopi now optionally logs the list of network connections on the machine after failures. Enable this by installing the optional otopi-debug-plugins package. This can help debug failures to start network services due to "Address already in use" errors.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1525473
Proton is a high performance, lightweight messaging library. It can be used in the widest range of messaging applications including brokers, client libraries, routers, bridges, proxies, and more. Proton is based on the AMQP 1.0 messaging standard. Proton enables integration with the AMQP 1.0 ecosystem from any platform, environment, or language.
Proton is a high performance, lightweight messaging library. It can be used in the widest range of messaging applications including brokers, client libraries, routers, bridges, proxies, and more. Proton is based on the AMQP 1.0 messaging standard. Proton enables integration with the AMQP 1.0 ecosystem from any platform, environment, or language.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1525933
Katello Agent is included in RHV and RHVH, and installed on the hosts during deployment. Katello Agent provides a list of the installed RPMs, for better integration with Satellite.
Katello Agent is included in RHV and RHVH, and installed on the hosts during deployment. Katello Agent provides a list of the installed RPMs, for better integration with Satellite.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1528290
otopi's yum plugin is now more verbose, also logging all RPM scriptlets' output. This can help debug failures caused by errors from such scriptlets.
otopi's yum plugin is now more verbose, also logging all RPM scriptlets' output. This can help debug failures caused by errors from such scriptlets.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1531016
This update adds IBRS-enabled CPU Models to clusters, which can be used to mitigate Spectre vulnerability in guests. See https://access.redhat.com/solutions/3307851 for more information.
This update adds IBRS-enabled CPU Models to clusters, which can be used to mitigate Spectre vulnerability in guests. See https://access.redhat.com/solutions/3307851 for more information.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1532180
The Software subtab in the Host General pane now contains a new "Kernel Features" field showing the PTI, IBPB, and IBRS kernel features of the host system, if the information is available. The Virtual Machine General pane now contains a new "Guest CPU Type" field showing the guest machine CPU type. These can be used to easily recognize if meltdown fixes are applied. For more info on meltdown patches detection, see https://access.redhat.com/solutions/3317841.
The Software subtab in the Host General pane now contains a new "Kernel Features" field showing the PTI, IBPB, and IBRS kernel features of the host system, if the information is available. The Virtual Machine General pane now contains a new "Guest CPU Type" field showing the guest machine CPU type. These can be used to easily recognize if meltdown fixes are applied. For more info on meltdown patches detection, see https://access.redhat.com/solutions/3317841.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1533061
When live or cold merge fails, snapshot disks may be left in an illegal state. If virtual machines with illegal snapshot disks are shut down, they will not re-start. Virtual machines with illegal snapshot disks are now marked with an exclamation mark and a warning message not to shut them down.
When live or cold merge fails, snapshot disks may be left in an illegal state. If virtual machines with illegal snapshot disks are shut down, they will not re-start. Virtual machines with illegal snapshot disks are now marked with an exclamation mark and a warning message not to shut them down.Copy to Clipboard Copied! Toggle word wrap Toggle overflow
3.9.2. Known Issues Copy linkLink copied to clipboard!
Copy linkLink copied to clipboard!
These known issues exist in Red Hat Virtualization at this time:
- BZ#1516494
Previously, having an empty VM directory in the export domain caused an error when trying to read the VMs stored in it, resulting in endless SPM re-elections. Now, this issue no longer causes SPM re-election. To fix the VM list loading error, you must remove the empty directory manually.
Previously, having an empty VM directory in the export domain caused an error when trying to read the VMs stored in it, resulting in endless SPM re-elections. Now, this issue no longer causes SPM re-election. To fix the VM list loading error, you must remove the empty directory manually.Copy to Clipboard Copied! Toggle word wrap Toggle overflow
3.9.3. Deprecated Functionality Copy linkLink copied to clipboard!
Copy linkLink copied to clipboard!
The items in this section are either no longer supported or will no longer be supported in a future release
- BZ#1529305
The kernel no longer supports the /sys/block/dm-X/queue/discard_zeroes_data file in sysfs. It is therefore no longer possible to determine whether discarded blocks from a block device returns zeros or the actual data. Therefore, the virtual machine disk properties "Wipe After Delete" and "Enable Discard" are no longer supported at the same time.
The kernel no longer supports the /sys/block/dm-X/queue/discard_zeroes_data file in sysfs. It is therefore no longer possible to determine whether discarded blocks from a block device returns zeros or the actual data. Therefore, the virtual machine disk properties "Wipe After Delete" and "Enable Discard" are no longer supported at the same time.Copy to Clipboard Copied! Toggle word wrap Toggle overflow
3.10. Red Hat Virtualization Batch Update 9 (ovirt-4.1.10) Copy linkLink copied to clipboard!
Copy linkLink copied to clipboard!
3.10.1. Enhancements Copy linkLink copied to clipboard!
Copy linkLink copied to clipboard!
This release of Red Hat Virtualization features the following enhancements:
- BZ#1534421
Previously, the self-hosted engine did not support IBRS-compatible CPUs. In this release, IBRS-compatible CPUs are now supported by the self-hosted engine.
Previously, the self-hosted engine did not support IBRS-compatible CPUs. In this release, IBRS-compatible CPUs are now supported by the self-hosted engine.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1547031
In this release, the ENGINE_JBOSS_BLOCKING_TIMEOUT option, which changes the default value of jboss.as.management.blocking.timeout, is exposed. To change the default value, create "/etc/ovirt-engine/engine.conf.d/99-jboss-blocking-timeout.conf" with the line, "ENGINE_JBOSS_BLOCKING_TIMEOUT=NNN", where "NNN" is the number of seconds for the JBoss timeout.
In this release, the ENGINE_JBOSS_BLOCKING_TIMEOUT option, which changes the default value of jboss.as.management.blocking.timeout, is exposed. To change the default value, create "/etc/ovirt-engine/engine.conf.d/99-jboss-blocking-timeout.conf" with the line, "ENGINE_JBOSS_BLOCKING_TIMEOUT=NNN", where "NNN" is the number of seconds for the JBoss timeout.Copy to Clipboard Copied! Toggle word wrap Toggle overflow
3.11. Red Hat Virtualization Batch Update 10 (ovirt-4.1.11) Copy linkLink copied to clipboard!
Copy linkLink copied to clipboard!
3.11.1. Release Notes Copy linkLink copied to clipboard!
Copy linkLink copied to clipboard!
This section outlines important details about the release, including recommended practices and notable changes to Red Hat Virtualization. You must take this information into account to ensure the best possible outcomes for your deployment.
- BZ#1472047
The batik package is no longer required by the Red Hat Virtualization Manager. Red Hat recommends removing it after upgrading to Red Hat Virtualization 4.1.
The batik package is no longer required by the Red Hat Virtualization Manager. Red Hat recommends removing it after upgrading to Red Hat Virtualization 4.1.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1472048
The fop package is no longer required by the Red Hat Virtualization Manager. Red Hat recommends removing it after upgrading to Red Hat Virtualization 4.1.
The fop package is no longer required by the Red Hat Virtualization Manager. Red Hat recommends removing it after upgrading to Red Hat Virtualization 4.1.Copy to Clipboard Copied! Toggle word wrap Toggle overflow - BZ#1562074
Do not upgrade to RHEL 7.5 if you have virtual machines using Direct LUN disks with the "Use Privileged SCSI I/O" option checked, to avoid unexpected behavior.
Do not upgrade to RHEL 7.5 if you have virtual machines using Direct LUN disks with the "Use Privileged SCSI I/O" option checked, to avoid unexpected behavior.Copy to Clipboard Copied! Toggle word wrap Toggle overflow
Legal Notice Copy linkLink copied to clipboard!
Copy linkLink copied to clipboard!
Copyright © 2018 Red Hat, Inc.
This document is licensed by Red Hat under the Creative Commons Attribution-ShareAlike 3.0 Unported License. If you distribute this document, or a modified version of it, you must provide attribution to Red Hat, Inc. and provide a link to the original. If the document is modified, all Red Hat trademarks must be removed.
Red Hat, as the licensor of this document, waives the right to enforce, and agrees not to assert, Section 4d of CC-BY-SA to the fullest extent permitted by applicable law.
Red Hat, Red Hat Enterprise Linux, the Shadowman logo, JBoss, OpenShift, Fedora, the Infinity logo, and RHCE are trademarks of Red Hat, Inc., registered in the United States and other countries.
Linux® is the registered trademark of Linus Torvalds in the United States and other countries.
Java® is a registered trademark of Oracle and/or its affiliates.
XFS® is a trademark of Silicon Graphics International Corp. or its subsidiaries in the United States and/or other countries.
MySQL® is a registered trademark of MySQL AB in the United States, the European Union and other countries.
Node.js® is an official trademark of Joyent. Red Hat Software Collections is not formally related to or endorsed by the official Joyent Node.js open source or commercial project.
The OpenStack® Word Mark and OpenStack logo are either registered trademarks/service marks or trademarks/service marks of the OpenStack Foundation, in the United States and other countries and are used with the OpenStack Foundation's permission. We are not affiliated with, endorsed or sponsored by the OpenStack Foundation, or the OpenStack community.
All other trademarks are the property of their respective owners.