Chapter 10. MTV performance recommendations


Review recommendations for network and storage performance, cold and warm migrations, and multiple migrations or single migrations.

10.1. MTV performance recommendations

The purpose of this section is to share recommendations for efficient and effective migration of virtual machines (VMs) using Migration Toolkit for Virtualization (MTV), based on findings observed through testing.

The data provided here was collected from testing in Red Hat labs and is provided for reference only. 

Overall, these numbers should be considered to show the best-case scenarios.

The observed performance of migration can differ from these results and depends on several factors.

10.1.1. Ensure fast storage and network speeds

Ensure fast storage and network speeds, both for VMware and Red Hat OpenShift (OCP) environments.

  • To perform fast migrations, VMware must have fast read access to datastores. Networking between VMware ESXi hosts should be fast, ensure a 10 GiB network connection, and avoid network bottlenecks.

    • Extend the VMware network to the OCP Workers Interface network environment.
    • It is important to ensure that the VMware network offers high throughput (10 Gigabit Ethernet) and rapid networking to guarantee that the reception rates align with the read rate of the ESXi datastore.
    • Be aware that the migration process uses significant network bandwidth and that the migration network is utilized. If other services use that network, it might have an impact on those services and their migration rates.
    • For example, 200 to 325 MiB/s was the average network transfer rate from the vmnic for each ESXi host associated with transferring data to the OCP interface.

Datastores read rates impact the total transfer times, so it is essential to ensure fast reads are possible from the ESXi datastore to the ESXi host.  

Example in numbers: 200 to 300 MiB/s was the average read rate for both vSphere and ESXi endpoints for a single ESXi server. When multiple ESXi servers are used, higher datastore read rates are possible.

10.1.3. Endpoint types 

MTV 2.6 allows for the following vSphere provider options:

  • ESXi endpoint (inventory and disk transfers from ESXi), introduced in MTV 2.6
  • vCenter Server endpoint; no networks for the ESXi host (inventory and disk transfers from vCenter)
  • vCenter endpoint and ESXi networks are available (inventory from vCenter, disk transfers from ESXi).

When transferring many VMs that are registered to multiple ESXi hosts, using the vCenter endpoint and ESXi network is suggested.

Note

As of vSphere 7.0, ESXi hosts can label which network to use for Network Block Device (NBD) transport. This is accomplished by tagging the desired virtual network interface card (NIC) with the appropriate vSphereBackupNFC label. When this is done, MTV is able to use the ESXi interface for network transfer to OpenShift provided that the worker and ESXi host interfaces are reachable. This is especially useful when migration users might not have access to the ESXi credentials yet want to be able to control which ESXi interface is used for migration. 

For more details, see: (MTV-1230)

You can use the following ESXi command, which designates interface vmk2 for NBD backup:

esxcli network ip interface tag add -t vSphereBackupNFC -i vmk2
Copy to Clipboard Toggle word wrap

Where possible, ensure that hosts used to perform migrations are set with BIOS profiles related to maximum performance. Hosts which use Host Power Management controlled within vSphere should check that High Performance is set.

Testing showed that when transferring more than 10 VMs with both BIOS and host power management set accordingly, migrations had an increase of 15 MiB in the average datastore read rate.

You can reduce the network load on VMware networks by selecting the migration network when using the ESXi endpoint.

By incorporating a virtualization provider, MTV enables the selection of a specific network, which is accessible on the ESXi hosts, for the purpose of migrating virtual machines to OpenShift. Selecting this migration network from the ESXi host in the MTV UI ensures that the transfer is performed using the selected network as an ESXi endpoint.

It is imperative to ensure that the network selected has connectivity to the OCP interface, has adequate bandwidth for migrations, and that the network interface is not saturated.

In environments with fast networks, such as 10GbE networks, migration network impacts can be expected to match the rate of ESXi datastore reads.

Set the MAX_VM_INFLIGHT MTV variable to control the maximum number of concurrent VMs transfers allowed for the ESXi host. 

MTV allows for concurrency to be controlled using this variable; by default, it is set to 20.

When setting MAX_VM_INFLIGHT, consider the number of maximum concurrent VMs transfers are required for ESXi hosts. It is important to consider the type of migration to be transferred concurrently. Warm migrations, which are defined by migrations of a running VM that will be migrated over a scheduled time.

Warm migrations use snapshots to compare and migrate only the differences between previous snapshots of the disk. The migration of the differences between snapshots happens over specific intervals before a final cut-over of the running VM to OpenShift occurs. 

In MTV 2.6, MAX_VM_INFLIGHT reserves one transfer slot per VM, regardless of current migration activity for a specific snapshot or the number of disks that belong to a single VM. The total set by MAX_VM_INFLIGHT is used to indicate how many concurrent VM tranfers per ESXi host is allowed.

Example

  • MAX_VM_INFLIGHT = 20 and 2 ESXi hosts defined in the provider mean each host can transfer 20 VMs.

When multiple VMs from a specific ESXi host are to be migrated, starting concurrent migrations for multiple VMs leads to faster migration times. 

Testing demonstrated that migrating 10 VMs (each containing 35 GiB of data, with a total size of 50 GiB) from a single host is significantly faster than migrating the same number of VMs sequentially, one after another. 

It is possible to increase concurrent migration to more than 10 virtual machines from a single host, but it does not show a significant improvement. 

Examples

  • 1 single disk VMs took 6 minutes, with migration rate of 100 MiB/s
  • 10 single disk VMs took 22 minutes, with migration rate of 272 MiB/s
  • 20 single disk VMs took 42 minutes, with migration rate of 284 MiB/s
Note

From the aforementioned examples, it is evident that the migration of 10 virtual machines simultaneously is three times faster than the migration of identical virtual machines in a sequential manner.

The migration rate was almost the same when moving 10 or 20 virtual machines simultaneously.

Using multiple hosts with registered VMs equally distributed among the ESXi hosts used for migrations leads to faster migration times.

Testing showed that when transferring more than 10 single disk VMS, each containing 35 GiB of data out of a total of 50G total, using an additional host can reduce migration time.

Examples

  • 80 single disk VMs, containing 35 GiB of data each, using a single host took 2 hours and 43 minutes, with a migration rate of 294 MiB/s.
  • 80 single disk VMs, containing 35 GiB of data each, using 8 ESXi hosts took 41 minutes, with a migration rate of 1,173 MiB/s.
Note

From the aforementioned examples, it is evident that migrating 80 VMs from 8 ESXi hosts, 10 from each host, concurrently is four times faster than running the same VMs from a single ESXi host. 

Migrating a larger number of VMs from more than 8 ESXi hosts concurrently could potentially show increased performance. However, it was not tested and therefore not recommended.

The maximum number of disks that can be referenced by a single migration plan is 500. For more details, see (MTV-1203)

When attempting to migrate many VMs in a single migration plan, it can take some time for all migrations to start. By breaking up one migration plan into several migration plans, it is possible to start them at the same time.

Comparing migrations of:

  • 500 VMs using 8 ESXi hosts in 1 plan, max_vm_inflight=100, took 5 hours and 10 minutes.
  • 800 VMs using 8 ESXi hosts with 8 plans, max_vm_inflight=100, took 57 minutes.

Testing showed that by breaking one single large plan into multiple moderately sized plans, for example, by using 100 VMs per plan, the total migration time can be reduced.

10.1.10. Maximum values tested for cold migrations

  • Maximum number of ESXi hosts tested: 8
  • Maximum number of VMs in a single migration plan: 500
  • Maximum number of VMs migrated in a single test: 5000
  • Maximum number of migration plans performed concurrently: 40
  • Maximum single disk size migrated: 6 TB disk, which contained 3 TB of data
  • Maximum number of disks on a single VM migrated: 50
  • Highest observed single datastore read rate from a single ESXi server:  312 MiB/second
  • Highest observed multi-datastore read rate using eight ESXi servers and two datastores: 1,242 MiB/second
  • Highest observed virtual NIC transfer rate to an OpenShift worker: 327 MiB/second
  • Maximum migration transfer rate of a single disk: 162 MiB/second (rate observed when transferring nonconcurrent migration of 1.5 TB utilized data)
  • Maximum cold migration transfer rate of the multiple VMs (single disk) from a single ESXi host: 294 MiB/s (concurrent migration of 30 VMs, 35/50 GiB used, from Single ESXi)
  • Maximum cold migration transfer rate of the multiple VMs (single disk) from multiple ESXi hosts: 1173MB/s (concurrent migration of 80 VMs, 35/50 GiB used, from 8 ESXi servers, 10 VMs from each ESXi)

10.1.11. Warm migration recommendations

The following recommendations are specific to warm migrations:

  • Migrate up to 400 disks in parallel

Testing involved migrating 200 VMs in parallel, with 2 disks each using 8 ESXi hosts, for a total of 400 disks. No tests were run on migration plans migrating over 400 disks in parallel, so it is not recommended to migrate over this number of disks in parallel.

  • Migrate up to 200 disks in parallel for the fastest rate

Testing was successfully performed on parallel disk migrations with 200, 300, and 400 disks. There was a decrease in the precopy migration rate, approximately 25%, between the tests migrating 200 disks and those migrating 300 and 400 disks.

Therefore, it is recommended to perform parallel disk migrations in groups of 200 or fewer, instead of 300 to 400 disks, unless a decline of 25% in precopy speed does not affect your cutover planning.

  • When possible, set cutover time to be immediately after a migration plan starts

To reduce the overall time of warm migrations, it is recommended to set the cutover to occur immediately after the migration plan is started. This causes MTV to run only one precopy per VM. This recommendation is valid, no matter how many VMs are in the migration plan.

  • Increase precopy intervals between snapshots

If you are creating many migration plans with a single VM and have enough time between the migration start and the cutover, increase the value of the controller_precopy_interval parameter to between 120 and 240 minutes, inclusive. The longer setting will reduce the total number of snapshots and disk transfers per VM before the cutover.

10.1.12. Maximum values tested for warm migrations

  • Maximum number of ESXi hosts tested: 8
  • Maximum number of worker nodes: 12
  • Maximum number of VMs in a single migration plan: 200
  • Maximum number of total parallel disk transfers: 400, with 200 VMs, 6 ESXis, and a transfer rate of 667 MB/s
  • Maximum single disk size migrated: 6 TB disk, which contained 3 TB of data
  • Maximum number of disks on a single VM migrated: 3
  • Maximum number of parallel disk transfers per ESXi host: 68
  • Maximum transfer rate observed of a single disk with no concurrent migrations: 76.5 MB/s
  • Maximum transfer rate observed of multiple disks from a single ESXi host: 253 MB/s (concurrent migration of 10 VMs, 1 disk each, 35/50 GiB used per disk)
  • Total transfer rate observed of multiple disks (210) from 8 ESXi hosts: 802 MB/s (concurrent migration of 70 VMs, 3 disks each, 35/50 GiB used per disk)

The following recommendations are suggested for VMs with data on disk totaling to 1 TB or greater for each individual disk:

  • Schedule appropriate maintenance windows for migrating large disk virtual machines (VMs). Such migrations are sensitive operations and might require careful planning of maintenance windows and downtime, especially during periods of lower storage and network activity.
  • Check that no other migration activities or other heavy network or storage activities are run during those large virtual machine (VM) migrations. You should treat these large virtual machine migrations as a special case. During those migrations, prioritize MTV activities. Plan to migrate those VMs to a time when there are fewer activities on those VMs and related datastore.
  • For large VMs with a high churn rate, which means data is frequently changed in amounts of 100 GB or more between snapshots, consider reducing the warm migration controller_precopy_interval from the default, which is 60 minutes. It is important to ensure that this process is started at least 24 hours before the scheduled cutover to allow for multiple successful precopy snapshots to complete. When scheduling the cutover, ensure that the maintenance window allows for enough time for the last snapshot of changes to be copied over and that the cutover process begins at the beginning of that maintenance window.
  • In cases of particularly large single-disk VMs, where some downtime is possible, select cold migrations rather than warm migrations, especially in the case of large VM snapshots.
  • Consider splitting data on particularly large disks to multiple disks, which enables parallel disk migration with MTV when warm migration is used.
  • If you have large database disks with continuous writes of large amounts of data, where downtime and VM snapshots are not possible, it might be necessary to consider database vendor-specific replication options of the database data to target these specific migrations outside MTV. Consult the vendor-specific options of your database if this case applies.

You can change Network Block Device (NBD) transport network file copy (NFC) parameters to increase migration performance when you use Asynchronous Input/Output (AIO) buffering with the Migration Toolkit for Virtualization (MTV).

Warning

Using AIO buffering is only suitable for cold migration use cases.

Disable AIO settings before initializing warm migrations. For more details, see Disabling AIO Buffer Configuration.

10.1.14.1. Key findings

  • The best migration performance was achieved by migrating multiple (10) virtual machines (VMs) on a single ESXi host with the following values:

    • VixDiskLib.nfcAio.Session.BufSizeIn64KB=16
    • vixDiskLib.nfcAio.Session.BufCount=4
  • The following improvements were noted when using AIO buffer settings (asynchronous buffer counts):

    • Migration time was reduced by 31.1%, from 0:24:32 to 0:16:54.
    • Read rate was increased from 347.83 MB/s to 504.93 MB/s.
  • There was no significant improvement observed when using AIO buffer settings with a single VM.
  • There was no significant improvement observed when using AIO buffer settings with multiple VMs from multiple hosts.

Support is based upon tests performed using the following versions:

  • vSphere 7.0.3
  • VDDK 7.0.3

10.1.15. Enabling and configuring AIO buffering

You can enable and configure Asynchronous Input/Output (AIO) buffering for use with the Migration Toolkit for Virtualization (MTV).

Procedure

  1. Ensure that the forklift-controller pod in the openshift-mtv namespace supports the AIO buffer values. Since the pod name prefix is dynamic, check the pod name by running the following command:

    oc get pods -n openshift-mtv | grep forklift-controller | awk '{print $1}'
    Copy to Clipboard Toggle word wrap

    For example, the output if the pod name prefix is "forklift-controller-667f57c8f8-qllnx" would be:

    forklift-controller-667f57c8f8-qllnx
    Copy to Clipboard Toggle word wrap
  2. Check the environment variables of the pod by running the following command:

    oc get pod forklift-controller-667f57c8f8-qllnx -n openshift-mtv -o yaml
    Copy to Clipboard Toggle word wrap
  3. Check for the following lines in the output:

    ...
    \- name: VIRT\_V2V\_EXTRA\_ARGS
    \- name: VIRT\_V2V\_EXTRA\_CONF\_CONFIG\_MAP
    ...
    Copy to Clipboard Toggle word wrap
  4. In the openshift-mtv namespace, edit the ForkliftController custom resource (CR) by performing the following steps:

    1. Access the ForkliftController CR for editing by running the following command:

      oc edit forkliftcontroller -n openshift-mtv
      Copy to Clipboard Toggle word wrap
    2. Add the following lines to the spec section of the ForkliftController CR:

      virt_v2v_extra_args: "--vddk-config /mnt/extra-v2v-conf/input.conf"
      virt_v2v_extra_conf_config_map: "perf"
      Copy to Clipboard Toggle word wrap
  5. Create the required config map perf by running the following command:

    oc -n openshift-mtv create cm perf
    Copy to Clipboard Toggle word wrap
  6. Convert the desired buffer configuration values to Base64. For example, for 16/4, run the following command:

    echo -e "VixDiskLib.nfcAio.Session.BufSizeIn64KB=16\nvixDiskLib.nfcAio.Session.BufCount=4" | base64
    Copy to Clipboard Toggle word wrap

    The output will be similar to the following:

    Vml4RGlza0xpYi5uZmNBaW8uU2Vzc2lvbi5CdWZTaXplSW42NEtCPTE2CnZpeERpc2tMaWIubmZjQWlvLlNlc3Npb24uQnVmQ291bnQ9NAo=
    Copy to Clipboard Toggle word wrap
  7. In the config map perf, enter the Base64 string in the binaryData section, for example:

    apiVersion: v1
    kind: ConfigMap
    binaryData:
      input.conf: Vml4RGlza0xpYi5uZmNBaW8uU2Vzc2lvbi5CdWZTaXplSW42NEtCPTE2CnZpeERpc2tMaWIubmZjQWlvLlNlc3Npb24uQnVmQ291bnQ9NAo=
    metadata:
      name: perf
      namespace: openshift-mtv
    Copy to Clipboard Toggle word wrap
  8. Restart the forklift-controller pod to apply the new configuration.
  9. Ensure the VIRT_V2V_EXTRA_ARGS environment variable reflects the updated settings.
  10. Run a migration plan and check the logs of the migration pod. Confirm that the AIO buffer settings are passed as parameters, particularly the --vddk-config value.

    For example, if you run the following command:

    exec: /usr/bin/virt-v2v … --vddk-config /mnt/extra-v2v-conf/input.conf
    Copy to Clipboard Toggle word wrap

    The logs include a section similar to the following, if debug_level = 4:

    Buffer size calc for 16 value:
    (16 * 64 * 1024 = 1048576)
    nbdkit: vddk[1]: debug: [NFC VERBOSE] NfcAio_OpenSession:
    Opening an AIO session.
    nbdkit: vddk[1]: debug: [NFC INFO] NfcAioInitSession:
    Disabling
    read-ahead buffer since the AIO buffer size of 1048576 is >=
    the read-ahead buffer size of 65536. Explicitly setting flag
    '`NFC_AIO_SESSION_NO_NET_READ_AHEAD`'
    nbdkit: vddk[1]: debug: [NFC VERBOSE] NfcAioInitSession: AIO Buffer Size is 1048576
    nbdkit: vddk[1]: debug: [NFC VERBOSE] NfcAioInitSession: AIO Buffer
    Count is 4
    Copy to Clipboard Toggle word wrap
  11. Verify that the correct config map values are in the migration pod. Do this by logging into the migration pod and running the following command:

    cat /mnt/extra-v2v-conf/input.conf
    Copy to Clipboard Toggle word wrap

    Example output is as follows:

    VixDiskLib.nfcAio.Session.BufSizeIn64KB=16
    vixDiskLib.nfcAio.Session.BufCount=4
    Copy to Clipboard Toggle word wrap
  12. Optional: Enable debug logs by running the following command. The command converts the configuration to Base64, including a high log level:

    echo -e
    "`VixDiskLib.nfcAio.Session.BufSizeIn64KB=16\nVixDiskLib.nfcAio.Session.BufCount=4\nVixDiskLib.nfc.LogLevel=4`"
    | base64
    Copy to Clipboard Toggle word wrap
    Note

    Adding a high log level reduces performance and is for debugging purposes only.

10.1.16. Disabling AIO buffering

You can disable AIO buffering for a cold migration using Migration Toolkit for Virtualization (MTV). You must disable AIO buffering for a warm migration using MTV.

Note

The procedure that follows assumes the AIO buffering was enabled and configured according to the procedure in Enabling and configuring AIO buffering.

Procedure

  1. In the openshift-mtv namespace, edit the ForkliftController custom resource (CR) by performing the following steps:

    1. Access the ForkliftController CR for editing by running the following command:

      oc edit forkliftcontroller -n openshift-mtv
      Copy to Clipboard Toggle word wrap
    2. Remove the following lines from the spec section of the ForkliftController CR:

      virt_v2v_extra_args: "`–vddk-config /mnt/extra-v2v-conf/input.conf`"
      virt_v2v_extra_conf_config_map: "`perf`"
      Copy to Clipboard Toggle word wrap
  2. Delete the config map named perf:

    oc delete cm perf -n openshift-mtv
    Copy to Clipboard Toggle word wrap
  3. Optional: Restart the forklift-controller pod to ensure that the changes took effect.

10.2. MTV performance addendum

The data provided here was collected from testing in Red Hat labs and is provided for reference only. 

Overall, these numbers should be considered to show the best-case scenarios.

The observed performance of migration can differ from these results and depends on several factors.

Back to top
Red Hat logoGithubredditYoutubeTwitter

Learn

Try, buy, & sell

Communities

About Red Hat Documentation

We help Red Hat users innovate and achieve their goals with our products and services with content they can trust. Explore our recent updates.

Making open source more inclusive

Red Hat is committed to replacing problematic language in our code, documentation, and web properties. For more details, see the Red Hat Blog.

About Red Hat

We deliver hardened solutions that make it easier for enterprises to work across platforms and environments, from the core datacenter to the network edge.

Theme

© 2025 Red Hat