Chapter 5. Technology Preview, Deprecated, and Removed Features

5.1. Technology Preview Features

Important

Technology Preview features are not supported with Red Hat production service-level agreements (SLAs) and might not be functionally complete, and Red Hat does not recommend using them for production. These features provide early access to upcoming product features, enabling customers to test functionality and provide feedback during the development process. For more information see Red Hat Technology Preview Features Support Scope.

The following table describes features available as Technology Previews in Red Hat Virtualization.

Table 5.1. Technology Preview Features

Technology Preview FeatureDetails

IPv6

Static IPv6 assignment is fully supported in Red Hat Virtualization 4.3 and 4.4, but Dynamic IPv6 assignment is available as a Technology Preview.

Note

All hosts in the cluster must use IPv4 or IPv6 for RHV networks, not simultaneous IPv4 and IPv6, because dual stack is not supported.

For details about IPv6 support, see IPv6 Networking Support in the Administration Guide.

NoVNC console option

Option for opening a virtual machine console in the browser using HTML5.

Websocket proxy

Allows users to connect to virtual machines through a noVNC console.

VDSM hook for nested virtualization

Allows a virtual machine to serve as a host. For details, see Enabling nested virtualization for all virtual machines in the Administration Guide.

Import Debian and Ubuntu virtual machines from VMware and RHEL 5 Xen

Allows virt-v2v to convert Debian and Ubuntu virtual machines from VMware or RHEL 5 Xen to KVM.

Known Issues:

  • virt-v2v cannot change the default kernel in the GRUB2 configuration. The kernel configured on the guest operating system is not changed during the conversion, even if a more optimal version is available.
  • After converting a Debian or Ubuntu virtual machine from VMware to KVM, the name of the virtual machine’s network interface may change, and will need to be configured manually

NVDIMM host devices

Support for attaching an emulated NVDIMM to virtual machines that are backed by NVDIMM on the host machine. For details, see NVDIMM Host Devices.

Open vSwitch (OVS) cluster type support

Adds Open vSwitch networking capabilities.

Shared and local storage in the same data center

Allows the creation of single-brick Gluster volumes to enable local storage to be used as a storage domain in shared data centers.

Cinderlib integration

Leverages CinderLib library to use Cinder-supported storage drivers in Red Hat Virtualization without a Full Cinder-OpenStack deployment. Adds support for Ceph storage along with Fibre Channel and iSCSI storage. The Cinder volume has multipath support on the Red Hat Virtualization Host.

SSO with OpenID Connect

Adds support for external OpenID Connect authentication using Keycloak in both the user interface and with the REST API.

oVirt Engine Backup

Adds support to back up and restore Red Hat Virtualization Manager with the Ansible ovirt-engine-backup role.

Failover vNIC profile

Allows users to migrate a virtual machine connected via SR-IOV with minimal downtime by using a failover network that is activated during migration.

Dedicated CPU pinning policy

Guest vCPUs will be exclusively pinned to a set of host pCPUs (similar to static CPU pinning). The set of pCPUs will be chosen to match the required guest CPU topology. If the host has an SMT architecture, thread siblings are preferred.