Chapter 11. nova
The following chapter contains information about the configuration options in the nova
service.
11.1. nova.conf
This section contains options for the /etc/nova/nova.conf
file.
11.1.1. DEFAULT
The following table outlines the options available under the [DEFAULT]
group in the /etc/nova/nova.conf
file.
Configuration option = Default value | Type | Description |
---|---|---|
| boolean value | Allow destination machine to match source for resize. Useful when testing in single-host environments. By default it is not allowed to resize to the same host. Setting this option to true will add the same host to the destination options. Also set to true if you allow the ServerGroupAffinityFilter and need to resize. |
| boolean value | Determine whether to allow network traffic from same network. When set to true, hosts on the same subnet are not filtered and are allowed to pass all types of traffic between them. On a flat network, this allows all instances from all projects unfiltered communication. With VLAN networking, this allows access between instances within the same project.
This option only applies when using the Possible values:
Related options:
|
| boolean value | Autoassigning floating IP to VM When set to True, floating IP is auto allocated and associated to the VM upon creation. Related options:
|
| string value | Enable eventlet backdoor. Acceptable values are 0, <port>, and <start>:<end>, where 0 results in listening on a random tcp port number; <port> results in listening on the specified port number (and not enabling backdoor if that port is in use); and <start>:<end> results in listening on the smallest unused port number within the specified range of port numbers. The chosen port is displayed in the service’s log file. |
| string value | Enable eventlet backdoor, using the provided path as a unix socket that can receive connections. This option is mutually exclusive with backdoor_port in that only one should be provided. If both are provided then the existence of this option overrides the usage of that option. |
| integer value | Interval to pull network bandwidth usage info. Not supported on all hypervisors. If a hypervisor doesn’t support bandwidth usage, it will not get the info in the usage events. Possible values:
|
| string value | The directory where the Nova binaries are installed. This option is only relevant if the networking capabilities from Nova are used (see services below). Nova’s networking capabilities are targeted to be fully replaced by Neutron in the future. It is very unlikely that you need to change this option from its default value. Possible values:
|
| integer value | Number of times to retry block device allocation on failures. Starting with Liberty, Cinder can use image volume cache. This may help with block device allocation performance. Look at the cinder image_volume_cache_enabled configuration option. Possible values:
|
| integer value | Interval (in seconds) between block device allocation retries on failures. This option allows the user to specify the time interval between consecutive retries. block_device_allocate_retries option specifies the maximum number of retries. Possible values:
Related options:
|
| string value | Path to SSL certificate file. |
| integer value |
This option represents the number of IP addresses to reserve at the top of the address range for VPN clients. It also will be ignored if the configuration option for Possible values:
Related options:
|
| string value | Defines which driver to use for controlling virtualization. Possible values:
|
| list value | A comma-separated list of monitors that can be used for getting compute metrics. You can use the alias/name from the setuptools entry points for nova.compute.monitors.* namespaces. If no namespace is supplied, the "cpu." namespace is assumed for backwards-compatibility. Note Only one monitor per namespace (For example: cpu) can be loaded at a time. Possible values:
|
| string value | Configuration drive format Configuration drive format that will contain metadata attached to the instance when it boots. Possible values:
Related options:
|
| integer value | The pool size limit for connections expiration policy |
| integer value | The time-to-live in sec of idle connections in the pool |
| string value | Console proxy host to be used to connect to instances on this host. It is the publicly visible name for the console host. Possible values:
|
| string value | The default exchange under which topics are scoped. May be overridden by an exchange name specified in the transport_url option. |
| floating point value | This option helps you specify virtual CPU to physical CPU allocation ratio. From Ocata (15.0.0) this is used to influence the hosts selected by the Placement API. Note that when Placement is used, the CoreFilter is redundant, because the Placement API will have already filtered out hosts that would have failed the CoreFilter. This configuration specifies ratio for CoreFilter which can be set per compute node. For AggregateCoreFilter, it will fall back to this configuration value if no per-aggregate setting is found. Note This can be set per-compute, or if set to 0.0, the value set on the scheduler node(s) or compute node(s) will be used and defaulted to 16.0. Once set to a non-default value, it is not possible to "unset" the config to get back to the default behavior. If you want to reset back to the default, explicitly specify 16.0. Note As of the 16.0.0 Pike release, this configuration option is ignored for the ironic.IronicDriver compute driver and is hardcoded to 1.0. Possible values:
|
| integer value |
This option determines how many times nova-network will attempt to create a unique MAC address before giving up and raising a Possible values:
Related options:
|
| boolean value | Run as a background process. |
| string value | The driver to use for database access |
| boolean value | If set to true, the logging level will be set to DEBUG instead of the default INFO level. |
| string value | Name of the network to be used to set access IPs for instances. If there are multiple IPs to choose from, an arbitrary one will be chosen. Possible values:
|
| string value | Default availability zone for compute services. This option determines the default availability zone for nova-compute services, which will be used if the service(s) do not belong to aggregates with availability zone metadata. Possible values:
|
| string value | The default format an ephemeral_volume will be formatted with on creation. Possible values:
|
| string value | Default flavor to use for the EC2 API only. The Nova API does not support a default flavor. |
| string value | Default pool for floating IPs. This option specifies the default floating IP pool for allocating floating IPs. While allocating a floating ip, users can optionally pass in the name of the pool they want to allocate from, otherwise it will be pulled from the default pool. If this option is not set, then nova is used as default floating pool. Possible values:
|
| list value | List of package logging levels in logger=LEVEL pairs. This option is ignored if log_config_append is set. |
| string value | Default availability zone for instances. This option determines the default availability zone for instances, which will be used when a user does not specify one when creating an instance. The instance(s) will be bound to this availability zone for their lifetime. Possible values:
|
| boolean value | Whether to batch up the application of IPTables rules during a host restart and apply all at the end of the init phase. |
| string value | This option allows you to specify the domain for the DHCP server. Possible values:
Related options:
|
| integer value | The lifetime of a DHCP lease, in seconds. The default is 86400 (one day). Possible values:
|
| string value | The location of the binary nova-dhcpbridge. By default it is the binary named nova-dhcpbridge that is installed with all the other nova binaries. Possible values:
|
| multi valued | This option is a list of full paths to one or more configuration files for dhcpbridge. In most cases the default path of /etc/nova/nova-dhcpbridge.conf should be sufficient, but if you have special needs for configuring dhcpbridge, you can change or add to this list. Possible values
|
| floating point value | This option helps you specify virtual disk to physical disk allocation ratio. From Ocata (15.0.0) this is used to influence the hosts selected by the Placement API. Note that when Placement is used, the DiskFilter is redundant, because the Placement API will have already filtered out hosts that would have failed the DiskFilter. A ratio greater than 1.0 will result in over-subscription of the available physical disk, which can be useful for more efficiently packing instances created with images that do not use the entire virtual disk, such as sparse or compressed images. It can be set to a value between 0.0 and 1.0 in order to preserve a percentage of the disk for uses other than instances. Note This can be set per-compute, or if set to 0.0, the value set on the scheduler node(s) or compute node(s) will be used and defaulted to 1.0. Once set to a non-default value, it is not possible to "unset" the config to get back to the default behavior. If you want to reset back to the default, explicitly specify 1.0. Note As of the 16.0.0 Pike release, this configuration option is ignored for the ironic.IronicDriver compute driver and is hardcoded to 1.0. Possible values:
|
| list value | This option is a list of zero or more IP address ranges in your network’s DMZ that should be accepted. Possible values:
|
| multi valued | Despite the singular form of the name of this option, it is actually a list of zero or more server addresses that dnsmasq will use for DNS nameservers. If this is not empty, dnsmasq will not read /etc/resolv.conf, but will only use the servers specified in this option. If the option use_network_dns_servers is True, the dns1 and dns2 servers from the network will be appended to this list, and will be used as DNS servers, too. Possible values:
Related options:
|
| integer value | This option determines the time, in seconds, to wait between refreshing DNS entries for the network. Possible values:
Related options:
|
`dnsmasq_config_file = ` | string value | The path to the custom dnsmasq configuration file, if any. Possible values:
|
| integer value | This option determines the number of times to retry ebtables commands before giving up. The minimum number of retries is 1. Possible values:
Related options:
|
| floating point value | This option determines the time, in seconds, that the system will sleep in between ebtables retries. Note that each successive retry waits a multiple of this value, so for example, if this is set to the default of 1.0 seconds, and ebtables_exec_attempts is 4, after the first failure, the system will sleep for 1 * 1.0 seconds, after the second failure it will sleep 2 * 1.0 seconds, and after the third failure it will sleep 3 * 1.0 seconds. Possible values:
Related options:
|
| boolean value | This option is used to enable or disable quota checking for tenant networks. Related options:
|
| boolean value | Enable new nova-compute services on this host automatically. When a new nova-compute service starts up, it gets registered in the database as an enabled service. Sometimes it can be useful to register new compute services in disabled state and then enabled them at a later point in time. This option only sets this behavior for nova-compute services, it does not auto-disable other services like nova-conductor, nova-scheduler, nova-consoleauth, or nova-osapi_compute. Possible values:
|
| list value | List of APIs to be enabled by default. |
| list value | List of APIs with enabled SSL. Nova provides SSL support for the API servers. enabled_ssl_apis option allows configuring the SSL support. |
| integer value | Size of executor thread pool when executor is threading or eventlet. |
| boolean value | This option is used mainly in testing to avoid calls to the underlying network utilities. |
| boolean value | Enables or disables fatal status of deprecations. |
| string value |
Firewall driver to use with
This option only applies when using the Possible values:
Related options:
|
| integer value | This is the number of seconds to wait before disassociating a deallocated fixed IP address. This is only used with the nova-network service, and has no effect when using neutron for networking. Possible values:
Related options:
|
| string value | This option determines the fixed IPv6 address block when creating a network. Please note that this option is only used when using nova-network instead of Neutron in your deployment. Possible values:
Related options:
|
| boolean value | This option determines whether the network setup information is injected into the VM before it is booted. While it was originally designed to be used only by nova-network, it is also used by the vmware and xenapi virt drivers to control whether network information is injected into a VM. The libvirt virt driver also uses it when we use config_drive to configure network to control whether network information is injected into a VM. |
| string value | This option is the name of the virtual interface of the VM on which the bridge will be built. While it was originally designed to be used only by nova-network, it is also used by libvirt for the bridge interface name. Possible values:
|
| string value | This option determines the bridge used for simple network interfaces when no bridge is specified in the VM creation request. Please note that this option is only used when using nova-network instead of Neutron in your deployment. Possible values:
Related options:
|
| string value | This is the address of the DNS server for a simple network. If this option is not specified, the default of 8.8.4.4 is used. Please note that this option is only used when using nova-network instead of Neutron in your deployment. Possible values:
Related options:
|
| string value | Full class name for the DNS Manager for floating IPs. This option specifies the class of the driver that provides functionality to manage DNS entries associated with floating IPs. When a user adds a DNS entry for a specified domain to a floating IP, nova will add a DNS entry using the specified floating DNS driver. When a floating IP is deallocated, its DNS entry will automatically be deleted. Possible values:
Related options:
|
| boolean value | Force injection to take place on a config drive When this option is set to true configuration drive functionality will be forced enabled by default, otherwise user can still enable configuration drives via the REST API or image metadata properties. Possible values:
Related options:
|
| boolean value | When this option is True, a call is made to release the DHCP for the instance when that instance is terminated. Related options:
|
| boolean value | Force conversion of backing images to raw format. Possible values:
Related options:
|
| multi valued |
This is a list of zero or more IP ranges that traffic from the Possible values:
Related options:
|
| multi valued | One or more interfaces that bridges can forward traffic to. If any of the items in this list is the special keyword all, then all traffic will be forwarded. Possible values:
|
| string value | This is the default IPv4 gateway. It is used only in the testing suite. Please note that this option is only used when using nova-network instead of Neutron in your deployment. Possible values:
Related options:
|
| string value | This is the default IPv6 gateway. It is used only in the testing suite. Please note that this option is only used when using nova-network instead of Neutron in your deployment. Possible values:
Related options:
|
| integer value | Specify a timeout after which a gracefully shutdown server will exit. Zero value means endless wait. |
| integer value | Interval between instance network information cache updates. Number of seconds after which each compute node runs the task of querying Neutron for all of its instances networking information, then updates the Nova db with that information. Nova will never update it’s cache if this option is set to 0. If we don’t update the cache, the metadata service and nova-api endpoints will be proxying incorrect network data about the instance. So, it is not recommended to set this option to 0. Possible values:
|
| string value | Hostname, FQDN or IP address of this host. Used as:
Must be valid within AMQP key. Possible values:
|
| integer value | Number of seconds to wait between runs of the image cache manager. Possible values: * 0: run at the default rate. * -1: disable * Any other value |
| string value | Location of cached images. This is NOT the full path - just a folder name relative to $instances_path. For per-compute-host cached images, set to base$my_ip |
| string value | Path to /etc/network/interfaces template. The path to a template file for the /etc/network/interfaces-style file, which will be populated by nova and subsequently used by cloudinit. This provides a method to configure network connectivity in environments without a DHCP server.
The template will be rendered using Jinja2 template engine, and receive a top-level key called Refer to the cloudinit documentaion for more information: https://cloudinit.readthedocs.io/en/latest/topics/datasources.html Possible values:
Related options:
|
| integer value | Maximum time in seconds that an instance can take to build. If this timer expires, instance status will be changed to ERROR. Enabling this option will make sure an instance will not be stuck in BUILD state for a longer period. Possible values:
|
| integer value | Interval for retrying failed instance file deletes. This option depends on maximum_instance_delete_attempts. This option specifies how often to retry deletes whereas maximum_instance_delete_attempts specifies the maximum number of retry attempts that can be made. Possible values:
Related options:
|
`instance_dns_domain = ` | string value | If specified, Nova checks if the availability_zone of every instance matches what the database says the availability_zone should be for the specified dns_domain. Related options:
|
| string value | Full class name for the DNS Manager for instance IPs. This option specifies the class of the driver that provides functionality to manage DNS entries for instances. On instance creation, nova will add DNS entries for the instance name and id, using the specified instance DNS driver and domain. On instance deletion, nova will remove the DNS entries. Possible values:
Related options:
|
`instance_format = [instance: %(uuid)s] ` | string value | The format for an instance that is passed with the log message. |
| string value | Template string to be used to generate instance names.
This template controls the creation of the database name of an instance. This is not the display name you enter when creating an instance (via Horizon or CLI). For a new deployment it is advisable to change the default value (which uses the database autoincrement) to another value which makes use of the attributes of an instance, like Possible values:
Related options:
|
| boolean value | This option enables periodic compute.instance.exists notifications. Each compute node must be configured to generate system usage data. These notifications are consumed by OpenStack Telemetry service. |
| string value | Time period to generate instance usages for. It is possible to define optional offset to given period by appending @ character followed by a number defining offset. Possible values:
|
`instance_uuid_format = [instance: %(uuid)s] ` | string value | The format for an instance UUID that is passed with the log message. |
| string value | Specifies where instances are stored on the hypervisor’s disk. It can point to locally attached storage or a directory on NFS. Possible values:
Related options:
|
| string value | Availability zone for internal services. This option determines the availability zone for the various internal nova services, such as nova-scheduler, nova-conductor, etc. Possible values:
|
`iptables_bottom_regex = ` | string value | This expression, if defined, will select any matching iptables rules and place them at the bottom when applying metadata changes to the rules. Possible values:
Related options:
|
| string value | By default, packets that do not pass the firewall are DROPped. In many cases, though, an operator may find it more useful to change this from DROP to REJECT, so that the user issuing those packets may have a better idea as to what’s going on, or LOGDROP in order to record the blocked traffic before DROPping. Possible values:
|
`iptables_top_regex = ` | string value | This expression, if defined, will select any matching iptables rules and place them at the top when applying metadata changes to the rules. Possible values:
Related options:
|
| string value | Abstracts out IPv6 address generation to pluggable backends. nova-network can be put into dual-stack mode, so that it uses both IPv4 and IPv6 addresses. In dual-stack mode, by default, instances acquire IPv6 global unicast addresses with the help of stateless address auto-configuration mechanism. Related options:
|
| string value | SSL key file (if separate from cert). |
| string value | This option allows you to specify the L3 management library to be used. Possible values:
Related options:
|
| string value | Base distinguished name for the LDAP search query This option helps to decide where to look up the host in LDAP. |
| string value | Bind user’s password for LDAP server |
| multi valued | DNS Servers for LDAP DNS driver Possible values:
|
| integer value | Expiry interval (in seconds) for LDAP DNS driver Start of Authority Time interval, a secondary/slave DNS server holds the information before it is no longer considered authoritative. |
| string value | Hostmaster for LDAP DNS driver Statement of Authority Possible values:
|
| integer value | Minimum interval (in seconds) for LDAP DNS driver Start of Authority It is Minimum time-to-live applies for all resource records in the zone file. This value is supplied to other servers how long they should keep the data in cache. |
| integer value | Refresh interval (in seconds) for LDAP DNS driver Start of Authority Time interval, a secondary/slave DNS server waits before requesting for primary DNS server’s current SOA record. If the records are different, secondary DNS server will request a zone transfer from primary. Note Lower values would cause more traffic. |
| integer value | Retry interval (in seconds) for LDAP DNS driver Start of Authority Time interval, a secondary/slave DNS server should wait, if an attempt to transfer zone failed during the previous refresh interval. |
| uri value | URL for LDAP server which will store DNS entries Possible values:
|
| string value | Bind user for LDAP server |
| string value | This is the class used as the ethernet device driver for linuxnet bridge operations. The default value should be all you need for most cases, but if you wish to use a customized class, set this option to the full dot-separated import path for that class. Possible values:
|
| string value | The name of the Open vSwitch bridge that is used with linuxnet when connecting with Open vSwitch." Possible values:
|
| integer value | Maximum number of 1 second retries in live_migration. It specifies number of retries to iptables when it complains. It happens when an user continuously sends live-migration request to same host leading to concurrent request to iptables. Possible values:
|
| string value | The name of a logging configuration file. This file is appended to any existing logging configuration files. For details about logging configuration files, see the Python logging module documentation. Note that when logging configuration files are used then all logging configuration is set in the configuration file and other logging configuration options are ignored (for example, logging_context_format_string). |
| string value | Defines the format string for %%(asctime)s in log records. Default: %(default)s . This option is ignored if log_config_append is set. |
| string value | (Optional) The base directory used for relative log_file paths. This option is ignored if log_config_append is set. |
| string value | (Optional) Name of log file to send logging output to. If no default is set, logging will go to stderr as defined by use_stderr. This option is ignored if log_config_append is set. |
| boolean value | Enables or disables logging values of all registered options when starting a service (at DEBUG level). |
| string value | Format string to use for log messages with context. |
| string value | Additional data to append to log message when logging level for the message is DEBUG. |
| string value | Format string to use for log messages when context is undefined. |
| string value | Prefix each line of exception output with this format. |
| string value | Defines the format string for %(user_identity)s that is used in logging_context_format_string. |
| integer value | Limits the maximum number of instance builds to run concurrently by nova-compute. Compute service can attempt to build an infinite number of instances, if asked to do so. This limit is enforced to avoid building unlimited instance concurrently on a compute node. This value can be set per compute node. Possible Values:
|
| integer value | Maximum number of live migrations to run concurrently. This limit is enforced to avoid outbound live migrations overwhelming the host/network and causing failures. It is not recommended that you change this unless you are very sure that doing so is safe and stable in your environment. Possible values:
|
| integer value | Maximum number of devices that will result in a local image being created on the hypervisor node. A negative number means unlimited. Setting max_local_block_devices to 0 means that any request that attempts to create a local disk will fail. This option is meant to limit the number of local discs (so root local disc that is the result of --image being used, and any other ephemeral and swap disks). 0 does not mean that images will be automatically converted to volumes and boot instances from volumes - it just means that all requests that attempt to create a local disk will fail. Possible values:
|
| integer value | The number of times to attempt to reap an instance’s files. This option specifies the maximum number of retry attempts that can be made. Possible values:
Related options: * |
| string value | This option determines the IP address for the network metadata API server. This is really the client side of the metadata host equation that allows nova-network to find the metadata server when doing a default multi host networking. Possible values:
Related options:
|
| string value | IP address on which the metadata API will listen. The metadata API service listens on this IP address for incoming requests. |
| port value | Port on which the metadata API will listen. The metadata API service listens on this port number for incoming requests. |
| port value | This option determines the port used for the metadata API server. Related options:
|
| integer value | Number of workers for metadata service. If not specified the number of available CPUs will be used. The metadata service can be configured to run as multi-process (workers). This overcomes the problem of reduction in throughput when API request concurrency increases. The metadata service will run in the specified number of processes. Possible Values:
|
| integer value | Number of times to retry live-migration before failing. Possible values:
|
| string value | Name or path of the tool used for ISO image creation Use the mkisofs_cmd flag to set the path where you install the genisoimage program. If genisoimage is on the system path, you do not need to change the default value. To use configuration drive with Hyper-V, you must set the mkisofs_cmd value to the full path to an mkisofs.exe installation. Additionally, you must set the qemu_img_cmd value in the hyperv configuration section to the full path to an qemu-img command installation. Possible values:
Related options:
|
| boolean value | Determine if monkey patching should be applied. Related options:
|
| list value | List of modules/decorators to monkey patch. This option allows you to patch a decorator for all functions in specified modules. Possible values:
Related options:
|
| boolean value | Default value for multi_host in networks. nova-network service can operate in a multi-host or single-host mode. In multi-host mode each compute node runs a copy of nova-network and the instances on that compute node use the compute node as a gateway to the Internet. Where as in single-host mode, a central server runs the nova-network service. All compute nodes forward traffic from the instances to the cloud controller which then forwards traffic to the Internet. If this options is set to true, some rpc network calls will be sent directly to host. Note that this option is only used when using nova-network instead of Neutron in your deployment. Related options:
|
| string value | When creating multiple instances with a single request using the os-multiple-create API extension, this template will be used to build the display name for each instance. The benefit is that the instances end up with different hostnames. Example display names when creating two VM’s: name-1, name-2. Possible values:
|
| string value | The IP address which is used to connect to the block storage network. Possible values:
Related options:
|
| string value | The IP address which the host is using to connect to the management network. Possible values:
Related options:
|
| integer value | Number of times to retry network allocation. It is required to attempt network allocation retries if the virtual interface plug fails. Possible values:
|
| string value | Driver to use for network creation. Network driver initializes (creates bridges and so on) only when the first VM lands on a host node. All network managers configure the network using network drivers. The driver is not tied to any particular network manager. The default Linux driver implements vlans, bridges, and iptables rules using linux utilities. Note that this option is only used when using nova-network instead of Neutron in your deployment. Related options:
|
| string value | Full class name for the Manager for network |
| integer value | This option determines the number of addresses in each private subnet. Please note that this option is only used when using nova-network instead of Neutron in your deployment. Possible values:
Related options:
|
| string value | The location where the network configuration files will be kept. The default is the networks directory off of the location where nova’s Python module is installed. Possible values
|
| list value | Image properties that should not be inherited from the instance when taking a snapshot. This option gives an opportunity to select which image-properties should not be inherited by newly created snapshots. Possible values:
|
| integer value |
This option represents the number of networks to create if not explicitly specified when the network is created. The only time this is used is if a CIDR is specified, but an explicit network_size is not. In that case, the subnets are created by diving the IP address space of the CIDR by num_networks. The resulting subnet sizes cannot be larger than the configuration option Please note that this option is only used when using nova-network instead of Neutron in your deployment. Possible values:
Related options:
|
| string value | IP address on which the OpenStack API will listen. The OpenStack API service listens on this IP address for incoming requests. |
| port value | Port on which the OpenStack API will listen. The OpenStack API service listens on this port number for incoming requests. |
`osapi_compute_unique_server_name_scope = ` | string value | Sets the scope of the check for unique instance names. The default doesn’t check for unique names. If a scope for the name check is set, a launch of a new instance or an update of an existing instance with a duplicate name will result in an 'InstanceExists' error. The uniqueness is case-insensitive. Setting this option can increase the usability for end users as they don’t have to distinguish among instances with the same name by their IDs. Possible values:
|
| integer value | Number of workers for OpenStack API service. The default will be the number of CPUs available. OpenStack API services can be configured to run as multi-process (workers). This overcomes the problem of reduction in throughput when API request concurrency increases. OpenStack API service will run in the specified number of processes. Possible Values:
|
| integer value | This option represents the period of time, in seconds, that the ovs_vsctl calls will wait for a response from the database before timing out. A setting of 0 means that the utility should wait forever for a response. Possible values:
|
| integer value | Length of generated instance admin passwords. |
| boolean value | Enable periodic tasks. If set to true, this option allows services to periodically run tasks on the manager. In case of running multiple schedulers or conductors you may want to run periodic tasks on only one host - in this case disable this option for all hosts but one. |
| integer value | Number of seconds to randomly delay when starting the periodic task scheduler to reduce stampeding. When compute workers are restarted in unison across a cluster, they all end up running the periodic tasks at the same time causing problems for the external services. To mitigate this behavior, periodic_fuzzy_delay option allows you to introduce a random initial delay when starting the periodic task scheduler. Possible Values:
|
| string value | Generic property to specify the pointer type. Input devices allow interaction with a graphical framebuffer. For example to provide a graphic tablet for absolute cursor movement. If set, the hw_pointer_model image property takes precedence over this configuration option. Possible values:
Related options:
|
| string value | The image preallocation mode to use. Image preallocation allows storage for instance images to be allocated up front when the instance is initially provisioned. This ensures immediate feedback is given if enough space isn’t available. In addition, it should significantly improve performance on writes to new blocks and may even improve I/O performance to prewritten blocks due to reduced fragmentation. Possible values:
|
| string value | This is the name of the network interface for public IP addresses. The default is eth0. Possible values:
|
| boolean value | Enables or disables publication of error events. |
| string value | The directory where the Nova python modules are installed. This directory is used to store template files for networking and remote console access. It is also the default path for other config options which need to persist Nova internal data. It is very unlikely that you need to change this option from its default value. Possible values:
Related options:
|
| integer value | This option controls the number of private networks that can be created per project (or per tenant). Related options:
|
| floating point value | This option helps you specify virtual RAM to physical RAM allocation ratio. From Ocata (15.0.0) this is used to influence the hosts selected by the Placement API. Note that when Placement is used, the RamFilter is redundant, because the Placement API will have already filtered out hosts that would have failed the RamFilter. This configuration specifies ratio for RamFilter which can be set per compute node. For AggregateRamFilter, it will fall back to this configuration value if no per-aggregate setting found. Note This can be set per-compute, or if set to 0.0, the value set on the scheduler node(s) or compute node(s) will be used and defaulted to 1.5. Once set to a non-default value, it is not possible to "unset" the config to get back to the default behavior. If you want to reset back to the default, explicitly specify 1.5. Note As of the 16.0.0 Pike release, this configuration option is ignored for the ironic.IronicDriver compute driver and is hardcoded to 1.0. Possible values:
|
| integer value | Maximum number of logged messages per rate_limit_interval. |
| string value | Log level name used by rate limiting: CRITICAL, ERROR, INFO, WARNING, DEBUG or empty string. Logs with level greater or equal to rate_limit_except_level are not filtered. An empty string means that all levels are filtered. |
| integer value | Interval, number of seconds, of log rate limiting. |
| integer value | Time interval after which an instance is hard rebooted automatically. When doing a soft reboot, it is possible that a guest kernel is completely hung in a way that causes the soft reboot task to not ever finish. Setting this option to a time period in seconds will automatically hard reboot an instance if it has been stuck in a rebooting state longer than N seconds. Possible values:
|
| integer value | Interval for reclaiming deleted instances. A value greater than 0 will enable SOFT_DELETE of instances. This option decides whether the server to be deleted will be put into the SOFT_DELETED state. If this value is greater than 0, the deleted server will not be deleted immediately, instead it will be put into a queue until it’s too old (deleted time greater than the value of reclaim_instance_interval). The server can be recovered from the delete queue by using the restore action. If the deleted server remains longer than the value of reclaim_instance_interval, it will be deleted by a periodic task in the compute service automatically. Note that this option is read from both the API and compute nodes, and must be set globally otherwise servers could be put into a soft deleted state in the API and never actually reclaimed (deleted) on the compute node. Possible values:
|
| string value | Filename that will be used for storing websocket frames received and sent by a proxy service (like VNC, spice, serial) running on this host. If this is not set, no recording will be done. |
| boolean value | Should unused base images be removed? |
| integer value | Unused unresized base images younger than this will not be removed. |
| integer value | Number of seconds indicating how frequently the state of services on a given hypervisor is reported. Nova needs to know this to determine the overall health of the deployment. Related Options:
|
| integer value | Interval to wait before un-rescuing an instance stuck in RESCUE. Possible values:
|
| integer value | Number of physical CPUs to reserve for the host. The host resources usage is reported back to the scheduler continuously from nova-compute running on the compute node. To prevent the host CPU from being considered as available, this option is used to reserve random pCPU(s) for the host. Possible values:
|
| integer value | Amount of disk resources in MB to make them always available to host. The disk usage gets reported back to the scheduler from nova-compute running on the compute nodes. To prevent the disk resources from being considered as available, this option can be used to reserve disk space for that host. Possible values:
|
| integer value | Amount of memory in MB to reserve for the host so that it is always available to host processes. The host resources usage is reported back to the scheduler continuously from nova-compute running on the compute node. To prevent the host memory from being considered as available, this option is used to reserve memory for the host. Possible values:
|
| dict value | Number of huge/large memory pages to reserved per NUMA host cell. Possible values:
|
| integer value | Automatically confirm resizes after N seconds. Resize functionality will save the existing server before resizing. After the resize completes, user is requested to confirm the resize. The user has the opportunity to either confirm or revert all changes. Confirm resize removes the original server and changes server status from resized to active. Setting this option to a time period (in seconds) will automatically confirm the resize if the server is in resized state longer than that time. Possible values:
|
| boolean value | Enable resizing of filesystems via a block device. If enabled, attempt to resize the filesystem by accessing the image over a block device. This is done by the host and may not be necessary if the image contains a recent version of cloud-init. Possible mechanisms require the nbd driver (for qcow and raw), or loop (for raw). |
| boolean value | This option specifies whether to start guests that were running before the host rebooted. It ensures that all of the instances on a Nova compute node resume their state each time the compute node boots or restarts. |
| string value | Path to the rootwrap configuration file. Goal of the root wrapper is to allow a service-specific unprivileged user to run a number of actions as the root user in the safest manner possible. The configuration file used here must match the one defined in the sudoers entry. |
| string value | The public IP address of the network host. This is used when creating an SNAT rule. Possible values:
Related options:
|
| integer value | Number of seconds to wait for an ack from a cast/call. After each retry attempt this timeout is multiplied by some specified multiplier. |
| integer value | Number to multiply base ack timeout by after each retry attempt. |
| string value | The messaging driver to use, defaults to rabbit. Other drivers include amqp and zmq. |
| integer value | Size of RPC connection pool. |
| integer value | Expiration timeout in seconds of a sent/received message after which it is not tracked anymore by a client/server. |
| integer value | The default number of seconds that poll should wait. Poll raises timeout exception when timeout expired. |
| integer value | Seconds to wait for a response from a call. |
| integer value | Default number of message sending attempts in case of any problems occurred: positive value N means at most N retries, 0 means no retries, None or -1 (or any other negative values) mean to retry forever. This option is used only if acknowledgments are enabled. |
| integer value | Maximum number of (green) threads to work concurrently. |
| boolean value | Wait for message acknowledgements from receivers. This mechanism works only via proxy without PUB/SUB. |
| string value | ZeroMQ bind address. Should be a wildcard (*), an ethernet interface, or IP. The "host" option should point or resolve to this address. |
| integer value | Number of retries to find free port number before fail with ZMQBindError. |
| integer value | Number of ZeroMQ contexts, defaults to 1. |
| string value | Name of this node. Must be a valid hostname, FQDN, or IP address. Must match "host" option, if running Nova. |
| string value | Directory for holding IPC sockets. |
| string value | MatchMaker driver. |
| integer value | Maximal port number for random ports range. |
| port value | Minimal port number for random ports range. |
| string value | Default serialization mechanism for serializing/deserializing outgoing/incoming messages |
| integer value | Maximum number of ingress messages to locally buffer per topic. Default is unlimited. |
| boolean value | Some periodic tasks can be run in a separate process. Should we run them here? |
| string value | The compute service periodically checks for instances that have been deleted in the database but remain running on the compute node. The above option enables action to be taken when such instances are identified. Possible values:
Related options:
|
| integer value | Time interval in seconds to wait between runs for the clean up action. If set to 0, above check will be disabled. If "running_deleted_instance _action" is set to "log" or "reap", a value greater than 0 must be set. Possible values:
Related options:
|
| integer value | Time interval in seconds to wait for the instances that have been marked as deleted in database to be eligible for cleanup. Possible values:
Related options:
|
| integer value | Interval between sending the scheduler a list of current instance UUIDs to verify that its view of instances is in sync with nova. If the CONF option scheduler_tracks_instance_changes is False, the sync calls will not be made. So, changing this option will have no effect. If the out of sync situations are not very common, this interval can be increased to lower the number of RPC messages being sent. Likewise, if sync issues turn out to be a problem, the interval can be lowered to check more frequently. Possible values:
Related options:
|
| boolean value | When True, when a device starts up, and upon binding floating IP addresses, arp messages will be sent to ensure that the arp caches on the compute hosts are up-to-date. Related options:
|
| integer value | When arp messages are configured to be sent, they will be sent with the count set to the value of this option. Of course, if this is set to zero, no arp messages will be sent. Possible values:
Related options:
|
| integer value | Maximum time in seconds since last check-in for up service Each compute node periodically updates their database status based on the specified report interval. If the compute node hasn’t updated the status for more than service_down_time, then the compute node is considered down. Related Options:
|
| string value | This option specifies the driver to be used for the servicegroup service. ServiceGroup API in nova enables checking status of a compute node. When a compute worker running the nova-compute daemon starts, it calls the join API to join the compute group. Services like nova scheduler can query the ServiceGroup API to check if a node is alive. Internally, the ServiceGroup client driver automatically updates the compute worker status. There are multiple backend implementations for this service: Database ServiceGroup driver and Memcache ServiceGroup driver. Possible Values:
Related Options:
|
| boolean value | THIS VALUE SHOULD BE SET WHEN CREATING THE NETWORK. If True in multi_host mode, all compute hosts share the same dhcp address. The same IP address used for DHCP will be added on each nova-network node which is only visible to the VMs on the same host. The use of this configuration has been deprecated and may be removed in any release after Mitaka. It is recommended that instead of relying on this option, an explicit value should be passed to create_networks() as a keyword argument with the name share_address. |
| integer value | Time before a shelved instance is eligible for removal from a host. By default this option is set to 0 and the shelved instance will be removed from the hypervisor immediately after shelve operation. Otherwise, the instance will be kept for the value of shelved_offload_time(in seconds) so that during the time period the unshelve action will be faster, then the periodic task will remove the instance from hypervisor after shelved_offload_time passes. Possible values:
|
| integer value | Interval for polling shelved instances to offload. The periodic task runs for every shelved_poll_interval number of seconds and checks if there are any shelved instances. If it finds a shelved instance, based on the shelved_offload_time config value it offloads the shelved instances. Check shelved_offload_time config option description for details. Possible values:
Related options:
|
| integer value | Total time to wait in seconds for an instance toperform a clean shutdown. It determines the overall period (in seconds) a VM is allowed to perform a clean shutdown. While performing stop, rescue and shelve, rebuild operations, configuring this option gives the VM a chance to perform a controlled shutdown before the instance is powered off. The default timeout is 60 seconds. The timeout value can be overridden on a per image basis by means of os_shutdown_timeout that is an image metadata setting allowing different types of operating systems to specify how much time they need to shut down cleanly. Possible values:
|
| boolean value | Set to True if source host is addressed with IPv6. |
| boolean value | Disallow non-encrypted connections. |
| string value | The top-level directory for maintaining Nova’s state.
This directory is used to store Nova’s internal state. It is used by a variety of other config options which derive from this. In some scenarios (for example migrations) it makes sense to use a storage location which is shared between multiple compute hosts (for example via NFS). Unless the option Possible values:
|
| list value | List of publisher hosts SubConsumer can subscribe on. This option has higher priority then the default publishers list taken from the matchmaker. |
| integer value | Interval to sync power states between the database and the hypervisor. The interval that Nova checks the actual virtual machine power state and the power state that Nova has in its database. If a user powers down their VM, Nova updates the API to report the VM has been powered down. Should something turn on the VM unexpectedly, Nova will turn the VM back off to keep the system in the expected state. Possible values:
Related options:
|
| integer value | Number of greenthreads available for use to sync power states. This option can be used to reduce the number of concurrent requests made to the hypervisor or system with real instance power states for performance reasons, for example, with Ironic. Possible values:
|
| string value | Syslog facility to receive log lines. This option is ignored if log_config_append is set. |
| boolean value | Determines whether unused gateway devices, both VLAN and bridge, are deleted if the network is in nova-network VLAN mode and is multi-hosted. Related options:
|
| string value | Explicitly specify the temporary working directory. |
| integer value | Amount of time, in seconds, to wait for NBD device start up. |
| string value | The network address and optional user credentials for connecting to the messaging backend, in URL format. The expected format is: driver://[user:pass@]host:port[,[userN:passN@]hostN:portN]/virtual_host?query Example: rabbit://rabbitmq:password@127.0.0.1:5672// For full details on the fields in the URL see the documentation of oslo_messaging.TransportURL at https://docs.openstack.org/oslo.messaging/latest/reference/transport.html |
| boolean value | When this option is True, whenever a DNS entry must be updated, a fanout cast message is sent to all network hosts to update their DNS entries in multi-host mode. Related options:
|
| integer value | Interval for updating compute resources. This option specifies how often the update_available_resources periodic task should run. A number less than 0 means to disable the task completely. Leaving this at the default of 0 will cause this to run at the default periodic interval. Setting it to any positive value will cause it to run at approximately that number of seconds. Possible values:
|
| boolean value | Enable journald for logging. If running in a systemd environment you may wish to enable journal support. Doing so will use the journal native protocol which includes structured metadata in addition to log messages.This option is ignored if log_config_append is set. |
| boolean value | Use JSON formatting for logging. This option is ignored if log_config_append is set. |
| boolean value | Use syslog for logging. Existing syslog format is DEPRECATED and will be changed later to honor RFC5424. This option is ignored if log_config_append is set. |
| boolean value | Enable use of copy-on-write (cow) images. QEMU/KVM allow the use of qcow2 as backing files. By disabling this, backing files will not be used. |
| boolean value | This option makes direct connections dynamic or static. It makes sense only with use_router_proxy=False which means to use direct connections for direct message types (ignored otherwise). |
| boolean value | Assign IPv6 and IPv4 addresses when creating instances. Related options:
|
| boolean value |
When this option is set to True, the dns1 and dns2 servers for the network specified by the user on boot will be used for DNS, as well as any specified in the Related options:
|
| boolean value | Enable neutron as the backend for networking. Determine whether to use Neutron or Nova Network as the back end. Set to true to use neutron. |
| boolean value | Use PUB/SUB pattern for fanout methods. PUB/SUB always uses proxy. |
| boolean value | Start and use a daemon that can run the commands that need to be run with root privileges. This option is usually enabled on nodes that run nova compute processes. |
| boolean value | Use ROUTER remote proxy. |
| boolean value | When set to True, only the firt nic of a VM will get its default gateway from the DHCP server. |
| boolean value | Log output to standard error. This option is ignored if log_config_append is set. |
| string value | Defines which physical CPUs (pCPUs) can be used by instance virtual CPUs (vCPUs). Possible values:
|
| boolean value | Determine if instance should boot or fail on VIF plugging timeout. Nova sends a port update to Neutron after an instance has been scheduled, providing Neutron with the necessary information to finish setup of the port. Once completed, Neutron notifies Nova that it has finished setting up the port, at which point Nova resumes the boot of the instance since network connectivity is now supposed to be present. A timeout will occur if the reply is not received after a given interval. This option determines what Nova does when the VIF plugging timeout event happens. When enabled, the instance will error out. When disabled, the instance will continue to boot on the assumption that the port is ready. Possible values:
|
| integer value | Timeout for Neutron VIF plugging event message arrival. Number of seconds to wait for Neutron vif plugging events to arrive before continuing or failing (see vif_plugging_is_fatal). Related options:
|
| multi valued | Name of the mkfs commands for ephemeral device. The format is <os_type>=<mkfs command> |
| string value | This option is the name of the virtual interface of the VM on which the VLAN bridge will be built. While it was originally designed to be used only by nova-network, it is also used by libvirt and xenapi for the bridge interface name.
Please note that this setting will be ignored in nova-network if the configuration option for Possible values:
|
| integer value | This is the VLAN number used for private networks. Note that the when creating the networks, if the specified number has already been assigned, nova-network will increment this number until it finds an available VLAN.
Please note that this option is only used when using nova-network instead of Neutron in your deployment. It also will be ignored if the configuration option for Possible values:
Related options:
|
| integer value | Interval for gathering volume usages. This option updates the volume usage cache for every volume_usage_poll_interval number of seconds. Possible values:
|
| string value | This option is no longer used since the /os-cloudpipe API was removed in the 16.0.0 Pike release. This is the public IP address for the cloudpipe VPN servers. It defaults to the IP address of the host.
Please note that this option is only used when using nova-network instead of Neutron in your deployment. It also will be ignored if the configuration option for Possible values:
Related options:
|
| port value | This is the port number to use as the first VPN port for private networks.
Please note that this option is only used when using nova-network instead of Neutron in your deployment. It also will be ignored if the configuration option for Possible values:
Related options:
|
| boolean value | Uses logging handler designed to watch file system. When log file is moved or removed this handler will open a new log file with specified path instantaneously. It makes sense only if log_file option is specified and Linux platform is used. This option is ignored if log_config_append is set. |
| string value | Path to directory with content which will be served by a web server. |
| integer value | How many additional connections to a host will be made for failover reasons. This option is actual only in dynamic connections mode. |
| boolean value | This option configures round-robin mode in zmq socket. True means not keeping a queue when server side disconnects. False means to keep queue and messages even if server is disconnected, when the server appears we send all accumulated messages to it. |
| integer value | Number of seconds to wait before all pending messages will be sent after closing a socket. The default value of -1 specifies an infinite linger period. The value of 0 specifies no linger period. Pending messages shall be discarded immediately when the socket is closed. Positive values specify an upper bound for the linger period. |
| integer value | Expiration timeout in seconds of a name service record about existing target ( < 0 means no timeout). |
| integer value | Update period in seconds of a name service record about existing target. |
| integer value | Enable/disable TCP keepalive (KA) mechanism. The default value of -1 (or any other negative value) means to skip any overrides and leave it to OS default; 0 and 1 (or any other positive value) mean to disable and enable the option respectively. |
| integer value | The number of retransmissions to be carried out before declaring that remote end is not available. The default value of -1 (or any other negative value and 0) means to skip any overrides and leave it to OS default. |
| integer value | The duration between two keepalive transmissions in idle condition. The unit is platform dependent, for example, seconds in Linux, milliseconds in Windows etc. The default value of -1 (or any other negative value and 0) means to skip any overrides and leave it to OS default. |
| integer value | The duration between two successive keepalive retransmissions, if acknowledgement to the previous keepalive transmission is not received. The unit is platform dependent, for example, seconds in Linux, milliseconds in Windows etc. The default value of -1 (or any other negative value and 0) means to skip any overrides and leave it to OS default. |
11.1.2. api
The following table outlines the options available under the [api]
group in the /etc/nova/nova.conf
file.
Configuration option = Default value | Type | Description |
---|---|---|
| boolean value | Operators can turn off the ability for a user to take snapshots of their instances by setting this option to False. When disabled, any attempt to take a snapshot will result in a HTTP 400 response ("Bad Request"). |
| string value | This determines the strategy to use for authentication: keystone or noauth2. noauth2 is designed for testing only, as it does no actual credential checking. noauth2 provides administrative credentials only if admin is specified as the username. |
| string value | This string is prepended to the normal URL that is returned in links to the OpenStack Compute API. If it is empty (the default), the URLs are returned unchanged. Possible values:
|
| string value | When gathering the existing metadata for a config drive, the EC2-style metadata is returned for all versions that don’t appear in this option. As of the Liberty release, the available versions are:
The option is in the format of a single string, with each version separated by a space. Possible values:
|
| boolean value | Enables returning of the instance password by the relevant server API calls such as create, rebuild, evacuate, or rescue. If the hypervisor does not support password injection, then the password returned will not be correct, so if your hypervisor does not support password injection, set this to False. |
| string value | The full path to the fping binary. |
| string value | This string is prepended to the normal URL that is returned in links to Glance resources. If it is empty (the default), the URLs are returned unchanged. Possible values:
|
| list value | This option is a list of all instance states for which network address information should not be returned from the API. Possible values: A list of strings, where each string is a valid VM state, as defined in nova/compute/vm_states.py. As of the Newton release, they are:
|
| integer value | As a query can potentially return many thousands of items, you can limit the maximum number of items in a single response by setting this option. |
| integer value | This option is the time (in seconds) to cache metadata. When set to 0, metadata caching is disabled entirely; this is generally not recommended for performance reasons. Increasing this setting should improve response times of the metadata API when under heavy load. Higher values may increase memory usage, and result in longer times for host metadata changes to take effect. |
| string value | Tenant ID for getting the default network from Neutron API (also referred in some places as the project ID) to use. Related options:
|
| boolean value | When True, the X-Forwarded-For header is treated as the canonical remote address. When False (the default), the remote_address header is used. You should only enable this if you have an HTML sanitizing proxy. |
| boolean value | When True, the TenantNetworkController will query the Neutron API to get the default networks to use. Related options:
|
| integer value | Maximum wait time for an external REST service to connect. Possible values:
Related options:
|
| boolean value | Should failures to fetch dynamic vendordata be fatal to instance boot? Related options:
|
| integer value | Maximum wait time for an external REST service to return data once connected. Possible values:
Related options:
|
`vendordata_dynamic_ssl_certfile = ` | string value | Path to an optional certificate file or CA bundle to verify dynamic vendordata REST services ssl certificates against. Possible values:
Related options:
|
| list value | A list of targets for the dynamic vendordata provider. These targets are of the form <name>@<url>. The dynamic vendordata provider collects metadata by contacting external REST services and querying them for information about the instance. This behaviour is documented in the vendordata.rst file in the nova developer reference. |
| string value | Cloud providers may store custom data in vendor data file that will then be available to the instances via the metadata service, and to the rendering of config-drive. The default class for this, JsonFileVendorData, loads this information from a JSON file, whose path is configured by this option. If there is no path set by this option, the class returns an empty dictionary. Possible values:
|
| list value | A list of vendordata providers. vendordata providers are how deployers can provide metadata via configdrive and metadata that is specific to their deployment. There are currently two supported providers: StaticJSON and DynamicJSON. StaticJSON reads a JSON file configured by the flag vendordata_jsonfile_path and places the JSON from that file into vendor_data.json and vendor_data2.json. DynamicJSON is configured via the vendordata_dynamic_targets flag, which is documented separately. For each of the endpoints specified in that flag, a section is added to the vendor_data2.json. For more information on the requirements for implementing a vendordata dynamic endpoint, please see the vendordata.rst file in the nova developer reference. Possible values:
Related options:
|
11.1.3. api_database
The following table outlines the options available under the [api_database]
group in the /etc/nova/nova.conf
file.
Configuration option = Default value | Type | Description |
---|---|---|
| string value | The SQLAlchemy connection string to use to connect to the database. |
| integer value | Verbosity of SQL debugging information: 0=None, 100=Everything. |
| integer value | Connections which have been present in the connection pool longer than this number of seconds will be replaced with a new one the next time they are checked out from the pool. |
| boolean value | Add Python stack traces to SQL as comment strings. |
| integer value | If set, use this value for max_overflow with SQLAlchemy. |
| integer value | Maximum number of SQL connections to keep open in a pool. Setting a value of 0 indicates no limit. |
| integer value | Maximum number of database connection retries during startup. Set to -1 to specify an infinite retry count. |
| string value | The SQL mode to be used for MySQL sessions. This option, including the default, overrides any server-set SQL mode. To use whatever SQL mode is set by the server configuration, set this to no value. Example: mysql_sql_mode= |
| integer value | If set, use this value for pool_timeout with SQLAlchemy. |
| integer value | Interval between retries of opening a SQL connection. |
| string value | The SQLAlchemy connection string to use to connect to the slave database. |
| boolean value | If True, SQLite uses synchronous mode. |
11.1.4. barbican
The following table outlines the options available under the [barbican]
group in the /etc/nova/nova.conf
file.
Configuration option = Default value | Type | Description |
---|---|---|
| string value | Use this endpoint to connect to Keystone |
| string value | Version of the Barbican API, for example: "v1" |
| string value | Use this endpoint to connect to Barbican, for example: "http://localhost:9311/" |
| integer value | Number of times to retry poll for key creation completion |
| integer value | Number of seconds to wait before retrying poll for key creation completion |
| boolean value | Specifies if insecure TLS (https) requests. If False, the server’s certificate will not be validated |
11.1.5. cache
The following table outlines the options available under the [cache]
group in the /etc/nova/nova.conf
file.
Configuration option = Default value | Type | Description |
---|---|---|
| string value | Cache backend module. For eventlet-based or environments with hundreds of threaded servers, Memcache with pooling (oslo_cache.memcache_pool) is recommended. For environments with less than 100 threaded servers, Memcached (dogpile.cache.memcached) or Redis (dogpile.cache.redis) is recommended. Test environments with a single instance of the server can use the dogpile.cache.memory backend. |
| multi valued | Arguments supplied to the backend module. Specify this option once per argument to be passed to the dogpile.cache backend. Example format: "<argname>:<value>". |
| string value | Prefix for building the configuration dictionary for the cache region. This should not need to be changed unless there is another dogpile.cache region with the same configuration name. |
| boolean value | Extra debugging from the cache backend (cache keys, get/set/delete/etc calls). This is only really useful if you need to see the specific cache-backend get/set/delete calls with the keys/values. Typically this should be left set to false. |
| boolean value | Global toggle for caching. |
| integer value | Default TTL, in seconds, for any cached item in the dogpile.cache region. This applies to any cached method that doesn’t have an explicit cache expiration time defined for it. |
| integer value | Number of seconds memcached server is considered dead before it is tried again. (dogpile.cache.memcache and oslo_cache.memcache_pool backends only). |
| integer value | Number of seconds that an operation will wait to get a memcache client connection. |
| integer value | Max total number of open connections to every memcached server. (oslo_cache.memcache_pool backend only). |
| integer value | Number of seconds a connection to memcached is held unused in the pool before it is closed. (oslo_cache.memcache_pool backend only). |
| list value | Memcache servers in the format of "host:port". (dogpile.cache.memcache and oslo_cache.memcache_pool backends only). |
| integer value | Timeout in seconds for every call to a server. (dogpile.cache.memcache and oslo_cache.memcache_pool backends only). |
| list value | Proxy classes to import that will affect the way the dogpile.cache backend functions. See the dogpile.cache documentation on changing-backend-behavior. |
11.1.6. cells
The following table outlines the options available under the [cells]
group in the /etc/nova/nova.conf
file.
Configuration option = Default value | Type | Description |
---|---|---|
| integer value | Bandwidth update interval. Seconds between bandwidth usage cache updates for cells. Possible values:
|
| integer value | Call timeout. Cell messaging module waits for response(s) to be put into the eventlet queue. This option defines the seconds waited for response from a call to a cell. Possible values:
|
| list value | Cell capabilities. List of arbitrary key=value pairs defining capabilities of the current cell to be sent to the parent cells. These capabilities are intended to be used in cells scheduler filters/weighers. Possible values:
|
| string value | Type of cell.
When cells feature is enabled the hosts in the OpenStack Compute cloud are partitioned into groups. Cells are configured as a tree. The top-level cell’s cell_type must be set to Related option:
|
| string value | Optional cells configuration. Configuration file from which to read cells configuration. If given, overrides reading cells from the database. Cells store all inter-cell communication data, including user names and passwords, in the database. Because the cells data is not updated very frequently, use this option to specify a JSON file to store cells data. With this configuration, the database is no longer consulted when reloading the cells data. The file must have columns present in the Cell model (excluding common database fields and the id column). You must specify the queue connection information through a transport_url field, instead of username, password, and so on. The transport_url has the following form: rabbit://USERNAME:PASSWORD@HOSTNAME:PORT/VIRTUAL_HOST Possible values: The scheme can be either qpid or rabbit, the following sample shows this optional configuration: { "parent": { "name": "parent", "api_url": "http://api.example.com:8774", "transport_url": "rabbit://rabbit.example.com", "weight_offset": 0.0, "weight_scale": 1.0, "is_parent": true }, "cell1": { "name": "cell1", "api_url": "http://api.example.com:8774", "transport_url": "rabbit://rabbit1.example.com", "weight_offset": 0.0, "weight_scale": 1.0, "is_parent": false }, "cell2": { "name": "cell2", "api_url": "http://api.example.com:8774", "transport_url": "rabbit://rabbit2.example.com", "weight_offset": 0.0, "weight_scale": 1.0, "is_parent": false } } |
| integer value | DB check interval. Cell state manager updates cell status for all cells from the DB only after this particular interval time is passed. Otherwise cached status are used. If this value is 0 or negative all cell status are updated from the DB whenever a state is needed. Possible values:
|
| boolean value | Enable cell v1 functionality. Note that cells v1 is considered experimental and not recommended for new Nova deployments. Cells v1 is being replaced by cells v2 which starting in the 15.0.0 Ocata release, all Nova deployments are at least a cells v2 cell of one. Setting this option, or any other options in the [cells] group, is not required for cells v2. When this functionality is enabled, it lets you to scale an OpenStack Compute cloud in a more distributed fashion without having to use complicated technologies like database and message queue clustering. Cells are configured as a tree. The top-level cell should have a host that runs a nova-api service, but no nova-compute services. Each child cell should run all of the typical nova-* services in a regular Compute cloud except for nova-api. You can think of cells as a normal Compute deployment in that each cell has its own database server and message queue broker. Related options:
|
| integer value | Instance update num instances On every run of the periodic task, nova cells manager will attempt to sync instance_updated_at_threshold number of instances. When the manager gets the list of instances, it shuffles them so that multiple nova-cells services do not attempt to sync the same instances in lockstep. Possible values:
Related options:
|
| integer value | Instance update sync database limit. Number of instances to pull from the database at one time for a sync. If there are more instances to update the results will be paged through. Possible values:
|
| integer value | Instance updated at threshold Number of seconds after an instance was updated or deleted to continue to update cells. This option lets cells manager to only attempt to sync instances that have been updated recently. i.e., a threshold of 3600 means to only update instances that have modified in the last hour. Possible values:
Related options:
|
| integer value | Maximum hop count When processing a targeted message, if the local cell is not the target, a route is defined between neighbouring cells. And the message is processed across the whole routing path. This option defines the maximum hop counts until reaching the target. Possible values:
|
| integer value | Mute child interval. Number of seconds after which a lack of capability and capacity update the child cell is to be treated as a mute cell. Then the child cell will be weighed as recommend highly that it be skipped. Possible values:
|
| floating point value | Mute weight multiplier. Multiplier used to weigh mute children. Mute children cells are recommended to be skipped so their weight is multiplied by this negative value. Possible values:
|
| string value | Name of the current cell. This value must be unique for each cell. Name of a cell is used as its id, leaving this option unset or setting the same name for two or more cells may cause unexpected behaviour. Related options:
|
| floating point value | Offset weight multiplier Multiplier used to weigh offset weigher. Cells with higher weight_offsets in the DB will be preferred. The weight_offset is a property of a cell stored in the database. It can be used by a deployer to have scheduling decisions favor or disfavor cells based on the setting. Possible values:
|
| floating point value | Ram weight multiplier. Multiplier used for weighing ram. Negative numbers indicate that Compute should stack VMs on one host instead of spreading out new VMs to more hosts in the cell. Possible values:
|
| floating point value | Reserve percentage Percentage of cell capacity to hold in reserve, so the minimum amount of free resource is considered to be; min_free = total * (reserve_percent / 100.0) This option affects both memory and disk utilization. The primary purpose of this reserve is to ensure some space is available for users who want to resize their instance to be larger. Note that currently once the capacity expands into this reserve space this option is ignored. Possible values:
|
| string value | RPC driver queue base. When sending a message to another cell by JSON-ifying the message and making an RPC cast to process_message, a base queue is used. This option defines the base queue name to be used when communicating between cells. Various topics by message type will be appended to this. Possible values:
|
| string value | Cells scheduler. The class of the driver used by the cells scheduler. This should be the full Python path to the class to be used. If nothing is specified in this option, the CellsScheduler is used. |
| list value | Scheduler filter classes. Filter classes the cells scheduler should use. An entry of "nova.cells.filters.all_filters" maps to all cells filters included with nova. As of the Mitaka release the following filter classes are available: Different cell filter: A scheduler hint of different_cell with a value of a full cell name may be specified to route a build away from a particular cell. Image properties filter: Image metadata named hypervisor_version_requires with a version specification may be specified to ensure the build goes to a cell which has hypervisors of the required version. If either the version requirement on the image or the hypervisor capability of the cell is not present, this filter returns without filtering out the cells. Target cell filter: A scheduler hint of target_cell with a value of a full cell name may be specified to route a build to a particular cell. No error handling is done as there’s no way to know whether the full path is a valid. As an admin user, you can also add a filter that directs builds to a particular cell. |
| integer value | Scheduler retries. How many retries when no cells are available. Specifies how many times the scheduler tries to launch a new instance when no cells are available. Possible values:
Related options:
|
| integer value | Scheduler retry delay.
Specifies the delay (in seconds) between scheduling retries when no cell can be found to place the new instance on. When the instance could not be scheduled to a cell after Possible values:
Related options:
|
| list value | Scheduler weight classes. Weigher classes the cells scheduler should use. An entry of "nova.cells.weights.all_weighers" maps to all cell weighers included with nova. As of the Mitaka release the following weight classes are available: mute_child: Downgrades the likelihood of child cells being chosen for scheduling requests, which haven’t sent capacity or capability updates in a while. Options include mute_weight_multiplier (multiplier for mute children; value should be negative). ram_by_instance_type: Select cells with the most RAM capacity for the instance type being requested. Because higher weights win, Compute returns the number of available units for the instance type requested. The ram_weight_multiplier option defaults to 10.0 that adds to the weight by a factor of 10. Use a negative number to stack VMs on one host instead of spreading out new VMs to more hosts in the cell. weight_offset: Allows modifying the database to weight a particular cell. The highest weight will be the first cell to be scheduled for launching an instance. When the weight_offset of a cell is set to 0, it is unlikely to be picked but it could be picked if other cells have a lower weight, like if they’re full. And when the weight_offset is set to a very high value (for example, 999999999999999), it is likely to be picked if another cell do not have a higher weight. |
11.1.7. cinder
The following table outlines the options available under the [cinder]
group in the /etc/nova/nova.conf
file.
Configuration option = Default value | Type | Description |
---|---|---|
| string value | Authentication URL |
| string value | Config Section from which to load plugin specific options |
| string value | Authentication type to load |
| string value | PEM encoded Certificate Authority to use when verifying HTTPs connections. |
| string value | Info to match when looking for cinder in the service catalog. Possible values:
Note: Nova does not support the Cinder v2 API since the Nova 17.0.0 Queens release. Related options:
|
| string value | PEM encoded client certificate cert file |
| boolean value | Allow attach between instance and volume in different availability zones. If False, volumes attached to an instance must be in the same availability zone in Cinder as the instance availability zone in Nova. This also means care should be taken when booting an instance from a volume where source is not "volume" because Nova will attempt to create a volume using the same availability zone as what is assigned to the instance. If that AZ is not in Cinder (or allow_availability_zone_fallback=False in cinder.conf), the volume create request will fail and the instance will fail the build request. By default there is no availability zone restriction on volume attach. |
| string value | Optional domain ID to use with v3 and v2 parameters. It will be used for both the user and project domain in v3 and ignored in v2 authentication. |
| string value | Optional domain name to use with v3 API and v2 parameters. It will be used for both the user and project domain in v3 and ignored in v2 authentication. |
| string value | Domain ID to scope to |
| string value | Domain name to scope to |
| string value | If this option is set then it will override service catalog lookup with this template for cinder endpoint Possible values:
Note: Nova does not support the Cinder v2 API since the Nova 17.0.0 Queens release. Related options:
|
| integer value | Number of times cinderclient should retry on any failed http call. 0 means connection is attempted only once. Setting it to any positive integer means that on failure connection is retried that many times e.g. setting it to 3 means total attempts to connect will be 4. Possible values:
|
| boolean value | Verify HTTPS connections. |
| string value | PEM encoded client certificate key file |
| string value | Region name of this node. This is used when picking the URL in the service catalog. Possible values:
|
| string value | User’s password |
| string value | Domain ID containing project |
| string value | Domain name containing project |
| string value | Project ID to scope to |
| string value | Project name to scope to |
| string value | Scope for system operations |
| string value | Tenant ID |
| string value | Tenant Name |
| integer value | Timeout value for http requests |
| string value | Trust ID |
| string value | User’s domain id |
| string value | User’s domain name |
| string value | User ID |
| string value | Username |
11.1.8. compute
The following table outlines the options available under the [compute]
group in the /etc/nova/nova.conf
file.
Configuration option = Default value | Type | Description |
---|---|---|
| integer value | Enables reporting of build failures to the scheduler. Any nonzero value will enable sending build failure statistics to the scheduler for use by the BuildFailureWeigher. Possible values:
Related options:
|
| string value | Defines which physical CPUs (pCPUs) will be used for best-effort guest vCPU resources. Currently only used by libvirt driver to place guest emulator threads when hw:emulator_threads_policy:share.
|
| boolean value |
Determine if the source compute host should wait for a If you set this option the same on all of your compute hosts, which you should do if you use the same networking backend universally, you do not have to worry about this.
Before starting the transfer of the guest, some setup occurs on the destination compute host, including plugging virtual interfaces. Depending on the networking backend on the destination host, a By default, this is False for two reasons:
Possible values:
Related options:
|
| integer value | Interval for updating nova-compute-side cache of the compute node resource provider’s aggregates and traits info. This option specifies the number of seconds between attempts to update a provider’s aggregates and traits information in the local cache of the compute node. Possible values:
|
11.1.9. conductor
The following table outlines the options available under the [conductor]
group in the /etc/nova/nova.conf
file.
Configuration option = Default value | Type | Description |
---|---|---|
| string value | Topic exchange name on which conductor nodes listen. |
| integer value | Number of workers for OpenStack Conductor service. The default will be the number of CPUs available. |
11.1.10. console
The following table outlines the options available under the [console]
group in the /etc/nova/nova.conf
file.
Configuration option = Default value | Type | Description |
---|---|---|
| list value | Adds list of allowed origins to the console websocket proxy to allow connections from other origin hostnames. Websocket proxy matches the host header with the origin header to prevent cross-site requests. This list specifies if any there are values other than host are allowed in the origin header. Possible values:
|
11.1.11. consoleauth
The following table outlines the options available under the [consoleauth]
group in the /etc/nova/nova.conf
file.
Configuration option = Default value | Type | Description |
---|---|---|
| integer value | The lifetime of a console auth token (in seconds). A console auth token is used in authorizing console access for a user. Once the auth token time to live count has elapsed, the token is considered expired. Expired tokens are then deleted. |
11.1.12. cors
The following table outlines the options available under the [cors]
group in the /etc/nova/nova.conf
file.
Configuration option = Default value | Type | Description |
---|---|---|
| boolean value | Indicate that the actual request can include user credentials |
| list value | Indicate which header field names may be used during the actual request. |
| list value | Indicate which methods can be used during the actual request. |
| list value | Indicate whether this resource may be shared with the domain received in the requests "origin" header. Format: "<protocol>://<host>[:<port>]", no trailing slash. Example: https://horizon.example.com |
| list value | Indicate which headers are safe to expose to the API. Defaults to HTTP Simple Headers. |
| integer value | Maximum cache age of CORS preflight requests. |
11.1.13. crypto
The following table outlines the options available under the [crypto]
group in the /etc/nova/nova.conf
file.
Configuration option = Default value | Type | Description |
---|---|---|
| string value | Filename of root CA (Certificate Authority). This is a container format and includes root certificates. Possible values:
Related options:
|
| string value | Directory path where root CA is located. Related options:
|
| string value | Filename of root Certificate Revocation List (CRL). This is a list of certificates that have been revoked, and therefore, entities presenting those (revoked) certificates should no longer be trusted. Related options:
|
| string value | Filename of a private key. Related options:
|
| string value | Directory path where keys are located. Related options:
|
| string value | Subject for certificate for projects, %s for project, timestamp |
| boolean value | Option to enable/disable use of CA for each project. |
| string value | Subject for certificate for users, %s for project, user, timestamp |
11.1.14. database
The following table outlines the options available under the [database]
group in the /etc/nova/nova.conf
file.
Configuration option = Default value | Type | Description |
---|---|---|
| string value | The back end to use for the database. |
| string value | The SQLAlchemy connection string to use to connect to the database. |
| integer value | Verbosity of SQL debugging information: 0=None, 100=Everything. |
| integer value | Connections which have been present in the connection pool longer than this number of seconds will be replaced with a new one the next time they are checked out from the pool. |
| boolean value | Add Python stack traces to SQL as comment strings. |
| boolean value | If True, increases the interval between retries of a database operation up to db_max_retry_interval. |
| integer value | Maximum retries in case of connection error or deadlock error before error is raised. Set to -1 to specify an infinite retry count. |
| integer value | If db_inc_retry_interval is set, the maximum seconds between retries of a database operation. |
| integer value | Seconds between retries of a database transaction. |
| integer value | If set, use this value for max_overflow with SQLAlchemy. |
| integer value | Maximum number of SQL connections to keep open in a pool. Setting a value of 0 indicates no limit. |
| integer value | Maximum number of database connection retries during startup. Set to -1 to specify an infinite retry count. |
| integer value | Minimum number of SQL connections to keep open in a pool. |
| boolean value | If True, transparently enables support for handling MySQL Cluster (NDB). |
| string value | The SQL mode to be used for MySQL sessions. This option, including the default, overrides any server-set SQL mode. To use whatever SQL mode is set by the server configuration, set this to no value. Example: mysql_sql_mode= |
| integer value | If set, use this value for pool_timeout with SQLAlchemy. |
| integer value | Interval between retries of opening a SQL connection. |
| string value | The SQLAlchemy connection string to use to connect to the slave database. |
| boolean value | If True, SQLite uses synchronous mode. |
| boolean value | Enable the experimental use of database reconnect on connection lost. |
| boolean value | Enable the experimental use of thread pooling for all DB API calls |
11.1.15. devices
The following table outlines the options available under the [devices]
group in the /etc/nova/nova.conf
file.
Configuration option = Default value | Type | Description |
---|---|---|
| list value | A list of the vGPU types enabled in the compute node. Some pGPUs (e.g. NVIDIA GRID K1) support different vGPU types. User can use this option to specify a list of enabled vGPU types that may be assigned to a guest instance. But please note that Nova only supports a single type in the Queens release. If more than one vGPU type is specified (as a comma-separated list), only the first one will be used. An example is as the following: [devices] enabled_vgpu_types = GRID K100,Intel GVT-g,MxGPU.2,nvidia-11 |
11.1.16. ephemeral_storage_encryption
The following table outlines the options available under the [ephemeral_storage_encryption]
group in the /etc/nova/nova.conf
file.
Configuration option = Default value | Type | Description |
---|---|---|
| string value | Cipher-mode string to be used. The cipher and mode to be used to encrypt ephemeral storage. The set of cipher-mode combinations available depends on kernel support. According to the dm-crypt documentation, the cipher is expected to be in the format: "<cipher>-<chainmode>-<ivmode>". Possible values:
|
| boolean value | Enables/disables LVM ephemeral storage encryption. |
| integer value | Encryption key length in bits. The bit length of the encryption key to be used to encrypt ephemeral storage. In XTS mode only half of the bits are used for encryption key. |
11.1.17. filter_scheduler
The following table outlines the options available under the [filter_scheduler]
group in the /etc/nova/nova.conf
file.
Configuration option = Default value | Type | Description |
---|---|---|
| string value | Image property namespace for use in the host aggregate. Images and hosts can be configured so that certain images can only be scheduled to hosts in a particular aggregate. This is done with metadata values set on the host aggregate that are identified by beginning with the value of this option. If the host is part of an aggregate with such a metadata key, the image in the request spec must have the value of that metadata in its properties in order for the scheduler to consider the host as acceptable. This option is only used by the FilterScheduler and its subclasses; if you use a different scheduler, this option has no effect. Also note that this setting only affects scheduling if the aggregate_image_properties_isolation filter is enabled. Possible values:
Related options:
|
| string value | Separator character(s) for image property namespace and name. When using the aggregate_image_properties_isolation filter, the relevant metadata keys are prefixed with the namespace defined in the aggregate_image_properties_isolation_namespace configuration option plus a separator. This option defines the separator to be used. This option is only used by the FilterScheduler and its subclasses; if you use a different scheduler, this option has no effect. Also note that this setting only affects scheduling if the aggregate_image_properties_isolation filter is enabled. Possible values:
Related options:
|
| multi valued | Filters that the scheduler can use. An unordered list of the filter classes the nova scheduler may apply. Only the filters specified in the enabled_filters option will be used, but any filter appearing in that option must also be included in this list. By default, this is set to all filters that are included with nova. This option is only used by the FilterScheduler and its subclasses; if you use a different scheduler, this option has no effect. Possible values:
Related options:
|
| list value | Filters used for filtering baremetal hosts. Filters are applied in order, so place your most restrictive filters first to make the filtering process more efficient. This option is only used by the FilterScheduler and its subclasses; if you use a different scheduler, this option has no effect. Possible values:
Related options:
|
| floating point value | Multiplier used for weighing hosts that have had recent build failures. This option determines how much weight is placed on a compute node with recent build failures. Build failures may indicate a failing, misconfigured, or otherwise ailing compute node, and avoiding it during scheduling may be beneficial. The weight is inversely proportional to the number of recent build failures the compute node has experienced. This value should be set to some high value to offset weight given by other enabled weighers due to available resources. To disable weighing compute hosts by the number of recent failures, set this to zero. This option is only used by the FilterScheduler and its subclasses; if you use a different scheduler, this option has no effect. Possible values:
Related options:
|
| floating point value | CPU weight multiplier ratio. Multiplier used for weighting free vCPUs. Negative numbers indicate stacking rather than spreading. This option is only used by the FilterScheduler and its subclasses; if you use a different scheduler, this option has no effect. Also note that this setting only affects scheduling if the cpu weigher is enabled. Possible values:
Related options:
|
| floating point value | Disk weight multipler ratio. Multiplier used for weighing free disk space. Negative numbers mean to stack vs spread. This option is only used by the FilterScheduler and its subclasses; if you use a different scheduler, this option has no effect. Also note that this setting only affects scheduling if the disk weigher is enabled. Possible values:
|
| list value | Filters that the scheduler will use. An ordered list of filter class names that will be used for filtering hosts. These filters will be applied in the order they are listed so place your most restrictive filters first to make the filtering process more efficient. This option is only used by the FilterScheduler and its subclasses; if you use a different scheduler, this option has no effect. Possible values:
Related options:
|
| integer value | Size of subset of best hosts selected by scheduler. New instances will be scheduled on a host chosen randomly from a subset of the N best hosts, where N is the value set by this option. Setting this to a value greater than 1 will reduce the chance that multiple scheduler processes handling similar requests will select the same host, creating a potential race condition. By selecting a host randomly from the N hosts that best fit the request, the chance of a conflict is reduced. However, the higher you set this value, the less optimal the chosen host may be for a given request. This option is only used by the FilterScheduler and its subclasses; if you use a different scheduler, this option has no effect. Possible values:
|
| string value | The default architecture to be used when using the image properties filter. When using the ImagePropertiesFilter, it is possible that you want to define a default architecture to make the user experience easier and avoid having something like x86_64 images landing on aarch64 compute nodes because the user did not specify the hw_architecture property in Glance. Possible values:
|
| floating point value | IO operations weight multipler ratio. This option determines how hosts with differing workloads are weighed. Negative values, such as the default, will result in the scheduler preferring hosts with lighter workloads whereas positive values will prefer hosts with heavier workloads. Another way to look at it is that positive values for this option will tend to schedule instances onto hosts that are already busy, while negative values will tend to distribute the workload across more hosts. The absolute value, whether positive or negative, controls how strong the io_ops weigher is relative to other weighers. This option is only used by the FilterScheduler and its subclasses; if you use a different scheduler, this option has no effect. Also note that this setting only affects scheduling if the io_ops weigher is enabled. Possible values:
|
| list value | List of hosts that can only run certain images. If there is a need to restrict some images to only run on certain designated hosts, list those host names here. This option is only used by the FilterScheduler and its subclasses; if you use a different scheduler, this option has no effect. Also note that this setting only affects scheduling if the IsolatedHostsFilter filter is enabled. Possible values:
Related options:
|
| list value | List of UUIDs for images that can only be run on certain hosts. If there is a need to restrict some images to only run on certain designated hosts, list those image UUIDs here. This option is only used by the FilterScheduler and its subclasses; if you use a different scheduler, this option has no effect. Also note that this setting only affects scheduling if the IsolatedHostsFilter filter is enabled. Possible values:
Related options:
|
| integer value | Maximum number of instances that be active on a host. If you need to limit the number of instances on any given host, set this option to the maximum number of instances you want to allow. The num_instances_filter will reject any host that has at least as many instances as this option’s value. This option is only used by the FilterScheduler and its subclasses; if you use a different scheduler, this option has no effect. Also note that this setting only affects scheduling if the num_instances_filter filter is enabled. Possible values:
|
| integer value | The number of instances that can be actively performing IO on a host. Instances performing IO includes those in the following states: build, resize, snapshot, migrate, rescue, unshelve. This option is only used by the FilterScheduler and its subclasses; if you use a different scheduler, this option has no effect. Also note that this setting only affects scheduling if the io_ops_filter filter is enabled. Possible values:
|
| floating point value | PCI device affinity weight multiplier.
The PCI device affinity weighter computes a weighting based on the number of PCI devices on the host and the number of PCI devices requested by the instance. The https://docs.openstack.org/nova/latest/user/filter-scheduler.html Possible values:
|
| floating point value | RAM weight multipler ratio. This option determines how hosts with more or less available RAM are weighed. A positive value will result in the scheduler preferring hosts with more available RAM, and a negative number will result in the scheduler preferring hosts with less available RAM. Another way to look at it is that positive values for this option will tend to spread instances across many hosts, while negative values will tend to fill up (stack) hosts as much as possible before scheduling to a less-used host. The absolute value, whether positive or negative, controls how strong the RAM weigher is relative to other weighers. This option is only used by the FilterScheduler and its subclasses; if you use a different scheduler, this option has no effect. Also note that this setting only affects scheduling if the ram weigher is enabled. Possible values:
|
| boolean value | Prevent non-isolated images from being built on isolated hosts. This option is only used by the FilterScheduler and its subclasses; if you use a different scheduler, this option has no effect. Also note that this setting only affects scheduling if the IsolatedHostsFilter filter is enabled. Even then, this option doesn’t affect the behavior of requests for isolated images, which will always be restricted to isolated hosts. Related options:
|
| boolean value | Enable spreading the instances between hosts with the same best weight. Enabling it is beneficial for cases when host_subset_size is 1 (default), but there is a large number of hosts with same maximal weight. This scenario is common in Ironic deployments where there are typically many baremetal nodes with identical weights returned to the scheduler. In such case enabling this option will reduce contention and chances for rescheduling events. At the same time it will make the instance packing (even in unweighed case) less dense. |
| floating point value | Multiplier used for weighing hosts for group soft-affinity. Possible values:
|
| floating point value | Multiplier used for weighing hosts for group soft-anti-affinity. Possible values:
|
| boolean value | Enable querying of individual hosts for instance information. The scheduler may need information about the instances on a host in order to evaluate its filters and weighers. The most common need for this information is for the (anti-)affinity filters, which need to choose a host based on the instances already running on a host. If the configured filters and weighers do not need this information, disabling this option will improve performance. It may also be disabled when the tracking overhead proves too heavy, although this will cause classes requiring host usage data to query the database on each request instead. This option is only used by the FilterScheduler and its subclasses; if you use a different scheduler, this option has no effect. Note In a multi-cell (v2) setup where the cell MQ is separated from the top-level, computes cannot directly communicate with the scheduler. Thus, this option cannot be enabled in that scenario. See also the [workarounds]/disable_group_policy_check_upcall option. |
| boolean value | Enable baremetal filters. Set this to True to tell the nova scheduler that it should use the filters specified in the baremetal_enabled_filters option. If you are not scheduling baremetal nodes, leave this at the default setting of False. This option is only used by the FilterScheduler and its subclasses; if you use a different scheduler, this option has no effect. Related options:
|
| list value | Weighers that the scheduler will use. Only hosts which pass the filters are weighed. The weight for any host starts at 0, and the weighers order these hosts by adding to or subtracting from the weight assigned by the previous weigher. Weights may become negative. An instance will be scheduled to one of the N most-weighted hosts, where N is scheduler_host_subset_size. By default, this is set to all weighers that are included with Nova. This option is only used by the FilterScheduler and its subclasses; if you use a different scheduler, this option has no effect. Possible values:
|
11.1.18. glance
The following table outlines the options available under the [glance]
group in the /etc/nova/nova.conf
file.
Configuration option = Default value | Type | Description |
---|---|---|
| list value | List of url schemes that can be directly accessed. This option specifies a list of url schemes that can be downloaded directly via the direct_url. This direct_URL can be fetched from Image metadata which can be used by nova to get the image more efficiently. nova-compute could benefit from this by invoking a copy when it has access to the same file system as glance. Possible values:
|
| list value | List of glance api servers endpoints available to nova. https is used for ssl-based glance api servers. Note The preferred mechanism for endpoint discovery is via keystoneauth1 loading options. Only use api_servers if you need multiple endpoints and are unable to use a load balancer for some reason. Possible values:
|
| string value | PEM encoded Certificate Authority to use when verifying HTTPs connections. |
| string value | PEM encoded client certificate cert file |
| boolean value | Enable or disable debug logging with glanceclient. |
| list value | List of certificate IDs for certificates that should be trusted. May be used as a default list of trusted certificate IDs for certificate validation. The value of this option will be ignored if the user provides a list of trusted certificate IDs with an instance API request. The value of this option will be persisted with the instance data if signature verification and certificate validation are enabled and if the user did not provide an alternative list. If left empty when certificate validation is enabled the user must provide a list of trusted certificate IDs otherwise certificate validation will fail. Related options:
|
| boolean value | Enable certificate validation for image signature verification. During image signature verification nova will first verify the validity of the image’s signing certificate using the set of trusted certificates associated with the instance. If certificate validation fails, signature verification will not be performed and the image will be placed into an error state. This provides end users with stronger assurances that the image data is unmodified and trustworthy. If left disabled, image signature verification can still occur but the end user will not have any assurance that the signing certificate used to generate the image signature is still trustworthy. Related options:
|
| string value |
Always use this endpoint URL for requests for this client. NOTE: The unversioned endpoint should be specified here; to request a particular API version, use the |
| boolean value | Verify HTTPS connections. |
| string value | PEM encoded client certificate key file |
| integer value | Enable glance operation retries. Specifies the number of retries when uploading / downloading an image to / from glance. 0 means no retries. |
| string value | The default region_name for endpoint URL discovery. |
| string value | The default service_name for endpoint URL discovery. |
| string value | The default service_type for endpoint URL discovery. |
| integer value | Timeout value for http requests |
| list value | List of interfaces, in order of preference, for endpoint URL. |
| boolean value | Enable image signature verification. nova uses the image signature metadata from glance and verifies the signature of a signed image while downloading that image. If the image signature cannot be verified or if the image signature metadata is either incomplete or unavailable, then nova will not boot the image and instead will place the instance into an error state. This provides end users with stronger assurances of the integrity of the image data they are using to create servers. Related options:
|
11.1.19. guestfs
The following table outlines the options available under the [guestfs]
group in the /etc/nova/nova.conf
file.
Configuration option = Default value | Type | Description |
---|---|---|
| boolean value | Enable/disables guestfs logging. This configures guestfs to debug messages and push them to OpenStack logging system. When set to True, it traces libguestfs API calls and enable verbose debug messages. In order to use the above feature, "libguestfs" package must be installed. Related options: Since libguestfs access and modifies VM’s managed by libvirt, below options should be set to give access to those VM’s.
|
11.1.20. healthcheck
The following table outlines the options available under the [healthcheck]
group in the /etc/nova/nova.conf
file.
Configuration option = Default value | Type | Description |
---|---|---|
| list value | Additional backends that can perform health checks and report that information back as part of a request. |
| boolean value | Show more detailed information as part of the response |
| string value | Check the presence of a file to determine if an application is running on a port. Used by DisableByFileHealthcheck plugin. |
| list value | Check the presence of a file based on a port to determine if an application is running on a port. Expects a "port:path" list of strings. Used by DisableByFilesPortsHealthcheck plugin. |
| string value | The path to respond to healtcheck requests on. |
11.1.21. hyperv
The following table outlines the options available under the [hyperv]
group in the /etc/nova/nova.conf
file.
Configuration option = Default value | Type | Description |
---|---|---|
| boolean value | Configuration drive cdrom OpenStack can be configured to write instance metadata to a configuration drive, which is then attached to the instance before it boots. The configuration drive can be attached as a disk drive (default) or as a CD drive. Possible values:
Related options:
|
| boolean value | Configuration drive inject password Enables setting the admin password in the configuration drive image. Related options:
|
| floating point value | Dynamic memory ratio Enables dynamic memory allocation (ballooning) when set to a value greater than 1. The value expresses the ratio between the total RAM assigned to an instance and its startup RAM amount. For example a ratio of 2.0 for an instance with 1024MB of RAM implies 512MB of RAM allocated at startup. Possible values:
|
| boolean value | Enable instance metrics collection Enables metrics collections for an instance by using Hyper-V’s metric APIs. Collected data can be retrieved by other apps and services, e.g.: Ceilometer. |
| boolean value | Enable RemoteFX feature This requires at least one DirectX 11 capable graphics adapter for Windows / Hyper-V Server 2012 R2 or newer and RDS-Virtualization feature has to be enabled. Instances with RemoteFX can be requested with the following flavor extra specs:
os:vram. Guest VM VRAM amount. Only available on Windows / Hyper-V Server 2016. Acceptable values:: 64, 128, 256, 512, 1024 |
`instances_path_share = ` | string value | Instances path share The name of a Windows share mapped to the "instances_path" dir and used by the resize feature to copy files to the target host. If left blank, an administrative share (hidden network share) will be used, looking for the same "instances_path" used locally. Possible values:
Related options:
|
| list value | List of iSCSI initiators that will be used for estabilishing iSCSI sessions. If none are specified, the Microsoft iSCSI initiator service will choose the initiator. |
| boolean value | Limit CPU features This flag is needed to support live migration to hosts with different CPU features and checked during instance creation in order to limit the CPU features used by the instance. |
| integer value | Mounted disk query retry count The number of times to retry checking for a mounted disk. The query runs until the device can be found or the retry count is reached. Possible values:
Related options:
|
| integer value | Mounted disk query retry interval Interval between checks for a mounted disk, in seconds. Possible values:
Related options:
|
| integer value | Power state check timeframe The timeframe to be checked for instance power state changes. This option is used to fetch the state of the instance from Hyper-V through the WMI interface, within the specified timeframe. Possible values:
|
| integer value | Power state event polling interval Instance power state change event polling frequency. Sets the listener interval for power state events to the given value. This option enhances the internal lifecycle notifications of instances that reboot themselves. It is unlikely that an operator has to change this value. Possible values:
|
| string value | qemu-img command qemu-img is required for some of the image related operations like converting between different image types. You can get it from here: (http://qemu.weilnetz.de/) or you can install the Cloudbase OpenStack Hyper-V Compute Driver (https://cloudbase.it/openstack-hyperv-driver/) which automatically sets the proper path for this config option. You can either give the full path of qemu-img.exe or set its path in the PATH environment variable and leave this option to the default value. Possible values:
Related options:
|
| boolean value | Use multipath connections when attaching iSCSI or FC disks. This requires the Multipath IO Windows feature to be enabled. MPIO must be configured to claim such devices. |
| integer value | Volume attach retry count The number of times to retry attaching a volume. Volume attachment is retried until success or the given retry count is reached. Possible values:
Related options:
|
| integer value | Volume attach retry interval Interval between volume attachment attempts, in seconds. Possible values:
Related options:
|
| string value | External virtual switch name The Hyper-V Virtual Switch is a software-based layer-2 Ethernet network switch that is available with the installation of the Hyper-V server role. The switch includes programmatically managed and extensible capabilities to connect virtual machines to both virtual networks and the physical network. In addition, Hyper-V Virtual Switch provides policy enforcement for security, isolation, and service levels. The vSwitch represented by this config option must be an external one (not internal or private). Possible values:
|
| integer value | Wait soft reboot seconds Number of seconds to wait for instance to shut down after soft reboot request is made. We fall back to hard reboot if instance does not shutdown within this window. Possible values:
|
11.1.22. ironic
The following table outlines the options available under the [ironic]
group in the /etc/nova/nova.conf
file.
Configuration option = Default value | Type | Description |
---|---|---|
| uri value | URL override for the Ironic API endpoint. |
| integer value | The number of times to retry when a request conflicts. If set to 0, only try once, no retries. Related options:
|
| integer value | The number of seconds to wait before retrying the request. Related options:
|
| string value | Authentication URL |
| string value | Config Section from which to load plugin specific options |
| string value | Authentication type to load |
| string value | PEM encoded Certificate Authority to use when verifying HTTPs connections. |
| string value | PEM encoded client certificate cert file |
| string value | Domain ID to scope to |
| string value | Domain name to scope to |
| string value |
Always use this endpoint URL for requests for this client. NOTE: The unversioned endpoint should be specified here; to request a particular API version, use the |
| boolean value | Verify HTTPS connections. |
| string value | PEM encoded client certificate key file |
| string value | User’s password |
| string value | Domain ID containing project |
| string value | Domain name containing project |
| string value | Project ID to scope to |
| string value | Project name to scope to |
| string value | The default region_name for endpoint URL discovery. |
| integer value | Timeout (seconds) to wait for node serial console state changed. Set to 0 to disable timeout. |
| string value | The default service_name for endpoint URL discovery. |
| string value | The default service_type for endpoint URL discovery. |
| string value | Scope for system operations |
| integer value | Timeout value for http requests |
| string value | Trust ID |
| string value | User’s domain id |
| string value | User’s domain name |
| string value | User ID |
| string value | Username |
| list value | List of interfaces, in order of preference, for endpoint URL. |
11.1.23. key_manager
The following table outlines the options available under the [key_manager]
group in the /etc/nova/nova.conf
file.
Configuration option = Default value | Type | Description |
---|---|---|
| string value | The type of authentication credential to create. Possible values are token, password, keystone_token, and keystone_password. Required if no context is passed to the credential factory. |
| string value | Use this endpoint to connect to Keystone. |
| string value | Specify the key manager implementation. Options are "barbican" and "vault". Default is "barbican". Will support the values earlier set using [key_manager]/api_class for some time. |
| string value | Domain ID for domain scoping. Optional for keystone_token and keystone_password auth_type. |
| string value | Domain name for domain scoping. Optional for keystone_token and keystone_password auth_type. |
| string value | Fixed key returned by key manager, specified in hex. Possible values:
|
| string value | Password for authentication. Required for password and keystone_password auth_type. |
| string value | Project’s domain ID for project. Optional for keystone_token and keystone_password auth_type. |
| string value | Project’s domain name for project. Optional for keystone_token and keystone_password auth_type. |
| string value | Project ID for project scoping. Optional for keystone_token and keystone_password auth_type. |
| string value | Project name for project scoping. Optional for keystone_token and keystone_password auth_type. |
| boolean value | Allow fetching a new token if the current one is going to expire. Optional for keystone_token and keystone_password auth_type. |
| string value | Token for authentication. Required for token and keystone_token auth_type if no context is passed to the credential factory. |
| string value | Trust ID for trust scoping. Optional for keystone_token and keystone_password auth_type. |
| string value | User’s domain ID for authentication. Optional for keystone_token and keystone_password auth_type. |
| string value | User’s domain name for authentication. Optional for keystone_token and keystone_password auth_type. |
| string value | User ID for authentication. Optional for keystone_token and keystone_password auth_type. |
| string value | Username for authentication. Required for password auth_type. Optional for the keystone_password auth_type. |
11.1.24. keystone
The following table outlines the options available under the [keystone]
group in the /etc/nova/nova.conf
file.
Configuration option = Default value | Type | Description |
---|---|---|
| string value | PEM encoded Certificate Authority to use when verifying HTTPs connections. |
| string value | PEM encoded client certificate cert file |
| string value |
Always use this endpoint URL for requests for this client. NOTE: The unversioned endpoint should be specified here; to request a particular API version, use the |
| boolean value | Verify HTTPS connections. |
| string value | PEM encoded client certificate key file |
| string value | The default region_name for endpoint URL discovery. |
| string value | The default service_name for endpoint URL discovery. |
| string value | The default service_type for endpoint URL discovery. |
| integer value | Timeout value for http requests |
| list value | List of interfaces, in order of preference, for endpoint URL. |
11.1.25. keystone_authtoken
The following table outlines the options available under the [keystone_authtoken]
group in the /etc/nova/nova.conf
file.
Configuration option = Default value | Type | Description |
---|---|---|
| string value | Config Section from which to load plugin specific options |
| string value | Authentication type to load |
| string value | Complete "public" Identity API endpoint. This endpoint should not be an "admin" endpoint, as it should be accessible by all end users. Unauthenticated clients are redirected to this endpoint to authenticate. Although this endpoint should ideally be unversioned, client support in the wild varies. If you’re using a versioned v2 endpoint here, then this should not be the same endpoint the service user utilizes for validating tokens, because normal end users may not be able to reach that endpoint. This option is deprecated in favor of www_authenticate_uri and will be removed in the S release. |
| string value | API version of the admin Identity API endpoint. |
| string value |
Request environment key where the Swift cache object is stored. When auth_token middleware is deployed with a Swift cache, use this option to have the middleware share a caching backend with swift. Otherwise, use the |
| string value | A PEM encoded Certificate Authority to use when verifying HTTPs connections. Defaults to system CAs. |
| string value | Required if identity server requires client certificate |
| boolean value | If true, the revocation list will be checked for cached tokens. This requires that PKI tokens are configured on the identity server. |
| boolean value | Do not handle authorization requests within the middleware, but delegate the authorization decision to downstream WSGI components. |
| string value | Used to control the use and type of token binding. Can be set to: "disabled" to not check token binding. "permissive" (default) to validate binding information if the bind type is of a form known to the server and ignore it if not. "strict" like "permissive" but if the bind type is unknown the token will be rejected. "required" any form of token binding is needed to be allowed. Finally the name of a binding method that must be present in tokens. |
| list value | Hash algorithms to use for hashing PKI tokens. This may be a single algorithm or multiple. The algorithms are those supported by Python standard hashlib.new(). The hashes will be tried in the order given, so put the preferred one first for performance. The result of the first hash will be stored in the cache. This will typically be set to multiple values only while migrating from a less secure algorithm to a more secure one. Once all the old tokens are expired this option should be set to a single value for better performance. |
| integer value | Request timeout value for communicating with Identity API server. |
| integer value | How many times are we trying to reconnect when communicating with Identity API Server. |
| boolean value | (Optional) Indicate whether to set the X-Service-Catalog header. If False, middleware will not ask for service catalog on token validation and will not set the X-Service-Catalog header. |
| boolean value | Verify HTTPS connections. |
| string value | Required if identity server requires client certificate |
| integer value | (Optional) Number of seconds that an operation will wait to get a memcached client connection from the pool. |
| integer value | (Optional) Number of seconds memcached server is considered dead before it is tried again. |
| integer value | (Optional) Maximum total number of open connections to every memcached server. |
| integer value | (Optional) Socket timeout in seconds for communicating with a memcached server. |
| integer value | (Optional) Number of seconds a connection to memcached is held unused in the pool before it is closed. |
| string value | (Optional, mandatory if memcache_security_strategy is defined) This string is used for key derivation. |
| string value | (Optional) If defined, indicate whether token data should be authenticated or authenticated and encrypted. If MAC, token data is authenticated (with HMAC) in the cache. If ENCRYPT, token data is encrypted and authenticated in the cache. If the value is not one of these options or empty, auth_token will raise an exception on initialization. |
| boolean value | (Optional) Use the advanced (eventlet safe) memcached client pool. The advanced pool will only work under python 2.x. |
| list value | Optionally specify a list of memcached server(s) to use for caching. If left undefined, tokens will instead be cached in-process. |
| string value | The region in which the identity server can be found. |
| integer value | Determines the frequency at which the list of revoked tokens is retrieved from the Identity service (in seconds). A high number of revocation events combined with a low cache duration may significantly reduce performance. Only valid for PKI tokens. This option has been deprecated in the Ocata release and will be removed in the P release. |
| list value | A choice of roles that must be present in a service token. Service tokens are allowed to request that an expired token can be used and so this check should tightly control that only actual services should be sending this token. Roles here are applied as an ANY check so any role in this list must be present. For backwards compatibility reasons this currently only affects the allow_expired check. |
| boolean value | For backwards compatibility reasons we must let valid service tokens pass that don’t pass the service_token_roles check as valid. Setting this true will become the default in a future release and should be enabled if possible. |
| string value | Directory used to cache files related to PKI tokens. This option has been deprecated in the Ocata release and will be removed in the P release. |
| integer value | In order to prevent excessive effort spent validating tokens, the middleware caches previously-seen tokens for a configurable duration (in seconds). Set to -1 to disable caching completely. |
| string value | Complete "public" Identity API endpoint. This endpoint should not be an "admin" endpoint, as it should be accessible by all end users. Unauthenticated clients are redirected to this endpoint to authenticate. Although this endpoint should ideally be unversioned, client support in the wild varies. If you’re using a versioned v2 endpoint here, then this should not be the same endpoint the service user utilizes for validating tokens, because normal end users may not be able to reach that endpoint. |
11.1.26. libvirt
The following table outlines the options available under the [libvirt]
group in the /etc/nova/nova.conf
file.
Configuration option = Default value | Type | Description |
---|---|---|
| boolean value | Write a checksum for files in _base to disk |
| integer value | How frequently to checksum base images |
`connection_uri = ` | string value | Overrides the default libvirt URI of the chosen virtualization type. If set, Nova will use this URI to connect to libvirt. Possible values:
Related options:
|
| string value | Is used to set the CPU mode an instance should have. If virt_type="kvm|qemu", it will default to "host-model", otherwise it will default to "none". Possible values:
Related options:
|
| string value | Set the name of the libvirt CPU model the instance should use. Possible values:
Related options:
|
| list value |
This allows specifying granular CPU feature flags when configuring CPU models. For example, to explicitly specify the [libvirt] cpu_mode = custom cpu_model = IvyBridge cpu_model_extra_flags = pcid
To specify multiple CPU flags (e.g. the Intel [libvirt] cpu_mode = custom cpu_model = Haswell-noTSX-IBRS cpu_model_extra_flags = PCID, VMX, pdpe1gb
As it can be noticed from above, the
The possible values for
Note that when using this config attribute to set the PCID CPU flag with the
For more information about https://www.us-cert.gov/ncas/alerts/TA18-141A https://www.redhat.com/archives/libvir-list/2018-May/msg01562.html https://www.redhat.com/archives/libvir-list/2018-June/msg01111.html
Besides Related options:
|
| list value | Specific cache modes to use for different disk types. For example: file=directsync,block=none,network=writeback For local or direct-attached storage, it is recommended that you use writethrough (default) mode, as it ensures data integrity and has acceptable I/O performance for applications running in the guest, especially for read operations. However, caching mode none is recommended for remote NFS storage, because direct I/O operations (O_DIRECT) perform better than synchronous I/O operations (with O_SYNC). Caching mode none effectively turns all guest I/O operations into direct I/O operations on the host, which is the NFS client in this environment. Possible cache modes:
|
| string value | Override the default disk prefix for the devices attached to an instance. If set, this is used to identify a free disk device name for a bus. Possible values:
Related options:
|
| list value |
This is a performance event list which could be used as monitor. These events will be passed to libvirt domain xml while creating a new instances. Then event statistics data can be collected from libvirt. The minimum libvirt version is 2.0.0. For more information about
Possible values: * A string list. For example: |
| list value | List of guid targets and ranges.Syntax is guest-gid:host-gid:countMaximum of 5 allowed. |
| string value | Discard option for nova managed disks. Requires:
|
| list value | For qemu or KVM guests, set this option to specify a default machine type per host architecture. You can find a list of supported machine types in your environment by checking the output of the "virsh capabilities"command. The format of the value for this config option is host-arch=machine-type. For example: x86_64=machinetype1,armv7l=machinetype2 |
| string value | Allows image information files to be stored in non-standard locations |
`images_rbd_ceph_conf = ` | string value | Path to the ceph configuration file to use |
| string value | The RADOS pool in which rbd volumes are stored |
| string value | VM Images format. If default is specified, then use_cow_images flag is used instead of this one. Related options:
|
| string value | LVM Volume Group that is used for VM images, when you specify images_type=lvm Related options:
|
| boolean value | Allow the injection of an SSH key at boot time.
There is no agent needed within the image to do this. If libguestfs is available on the host, it will be used. Otherwise nbd is used. The file system of the image will be mounted and the SSH key, which is provided in the REST API call will be injected as SSH key for the root user and appended to the This config option will enable directly modifying the instance disk and does not affect what cloud-init may do using data from config_drive option or the metadata service. Related options:
|
| integer value | Determines the way how the file system is chosen to inject data into it. libguestfs will be used a first solution to inject data. If that’s not available on the host, the image will be locally mounted on the host as a fallback solution. If libguestfs is not able to determine the root partition (because there are more or less than one root partition) or cannot mount the file system it will result in an error and the instance won’t be boot. Possible values:
Related options:
|
| boolean value |
Allow the injection of an admin password for instance only at There is no agent needed within the image to do this. If libguestfs is available on the host, it will be used. Otherwise nbd is used. The file system of the image will be mounted and the admin password, which is provided in the REST API call will be injected as password for the root user. If no root user is available, the instance won’t be launched and an error is thrown. Be aware that the injection is not possible when the instance gets launched from a volume. Possible values:
Related options:
|
| string value | The iSCSI transport iface to use to connect to target in case offload support is desired. Default format is of the form <transport_name>.<hwaddress> where <transport_name> is one of (be2iscsi, bnx2i, cxgb3i, cxgb4i, qla4xxx, ocs) and <hwaddress> is the MAC address of the interface and can be generated via the iscsiadm -m iface command. Do not confuse the iscsi_iface parameter to be provided here with the actual transport name. |
| boolean value | Use multipath connection of the iSER volume. iSER volumes can be connected as multipath devices. This will provide high availability and fault tolerance. |
| integer value | Maximum bandwidth(in MiB/s) to be used during migration. If set to 0, the hypervisor will choose a suitable default. Some hypervisors do not support this feature and will return an error if bandwidth is not 0. Please refer to the libvirt documentation for further details. |
| integer value | Time to wait, in seconds, for migration to successfully complete transferring data before aborting the operation. Value is per GiB of guest RAM + disk to be transferred, with lower bound of a minimum of 2 GiB. Should usually be larger than downtime delay * downtime steps. Set to 0 to disable timeouts. Related options:
|
| integer value | Maximum permitted downtime, in milliseconds, for live migration switchover. Will be rounded up to a minimum of 100ms. You can increase this value if you want to allow live-migrations to complete faster, or avoid live-migration timeout errors by allowing the guest to be paused for longer during the live-migration switch over. Related options:
|
| integer value | Time to wait, in seconds, between each step increase of the migration downtime. Minimum delay is 3 seconds. Value is per GiB of guest RAM + disk to be transferred, with lower bound of a minimum of 2 GiB per device. |
| integer value | Number of incremental steps to reach max downtime value. Will be rounded up to a minimum of 3 steps. |
| string value | The IP address or hostname to be used as the target for live migration traffic. If this option is set to None, the hostname of the migration target compute node will be used. This option is useful in environments where the live-migration traffic can impact the network plane significantly. A separate network for live-migration traffic can then use this config option and avoids the impact on the management network. Possible values:
Related options:
|
| boolean value | This option allows nova to start live migration with auto converge on. Auto converge throttles down CPU if a progress of on-going live migration is slow. Auto converge will only be used if this flag is set to True and post copy is not permitted or post copy is unavailable due to the version of libvirt and QEMU in use. Related options:
|
| boolean value | This option allows nova to switch an on-going live migration to post-copy mode, i.e., switch the active VM to the one on the destination node before the migration is complete, therefore ensuring an upper bound on the memory that needs to be transferred. Post-copy requires libvirt>=1.3.3 and QEMU>=2.5.0. When permitted, post-copy mode will be automatically activated if a live-migration memory copy iteration does not make percentage increase of at least 10% over the last iteration. The live-migration force complete API also uses post-copy when permitted. If post-copy mode is not available, force complete falls back to pausing the VM to ensure the live-migration operation will complete. When using post-copy mode, if the source and destination hosts loose network connectivity, the VM being live-migrated will need to be rebooted. For more details, please see the Administration guide. Related options:
|
| integer value | Time to wait, in seconds, for migration to make forward progress in transferring data before aborting the operation. Set to 0 to disable timeouts. This is deprecated, and now disabled by default because we have found serious bugs in this feature that caused false live-migration timeout failures. This feature will be removed or replaced in a future release. |
| string value | URI scheme used for live migration. Override the default libvirt live migration scheme (which is dependent on virt_type). If this option is set to None, nova will automatically choose a sensible default based on the hypervisor. It is not recommended that you change this unless you are very sure that hypervisor supports a particular scheme. Related options:
|
| boolean value | Enable tunnelled migration. This option enables the tunnelled migration feature, where migration data is transported over the libvirtd connection. If enabled, we use the VIR_MIGRATE_TUNNELLED migration flag, avoiding the need to configure the network to allow direct hypervisor to hypervisor communication. If False, use the native transport. If not set, Nova will choose a sensible default based on, for example the availability of native encryption support in the hypervisor. Enabling this option will definitely impact performance massively. Note that this option is NOT compatible with use of block migration. Related options:
|
| string value | Live migration target URI to use. Override the default libvirt live migration target URI (which is dependent on virt_type). Any included "%s" is replaced with the migration target hostname.
If this option is set to None (which is the default), Nova will automatically generate the
Related options:
|
| integer value | A number of seconds to memory usage statistics period. Zero or negative value mean to disable memory usage statistics. |
| string value | Mount options passed to the NFS client. See section of the nfs man page for details. Mount options controls the way the filesystem is mounted and how the NFS client behaves when accessing files on this mount point. Possible values:
|
| string value | Directory where the NFS volume is mounted on the compute node. The default is mnt directory of the location where nova’s Python module is installed. NFS provides shared storage for the OpenStack Block Storage service. Possible values:
|
| integer value | Number of times to rediscover AoE target to find volume. Nova provides support for block storage attaching to hosts via AOE (ATA over Ethernet). This option allows the user to specify the maximum number of retry attempts that can be made to discover the AoE device. |
| integer value | Number of times to scan iSER target to find volume. iSER is a server network protocol that extends iSCSI protocol to use Remote Direct Memory Access (RDMA). This option allows the user to specify the maximum number of scan attempts that can be made to find iSER volume. |
| integer value | Number of times to scan given storage protocol to find volume. |
| string value | Path to a Quobyte Client configuration file. |
| string value | Directory where the Quobyte volume is mounted on the compute node. Nova supports Quobyte volume driver that enables storing Block Storage service volumes on a Quobyte storage back end. This Option specifies the path of the directory where Quobyte volume is mounted. Possible values:
|
| string value | The libvirt UUID of the secret for the rbd_user volumes. |
| string value | The RADOS client name for accessing rbd(RADOS Block Devices) volumes. Libvirt will refer to this user when connecting and authenticating with the Ceph RBD server. |
| integer value | In a realtime host context vCPUs for guest will run in that scheduling priority. Priority depends on the host kernel (usually 1-99) |
| string value | libvirt’s transport method for remote file operations. Because libvirt cannot use RPC to copy files over network to/from other compute nodes, other method must be used for:
|
| integer value | Unused resized base images younger than this will not be removed |
| string value | The ID of the image to boot from to rescue data from a corrupted instance. If the rescue REST API operation doesn’t provide an ID of an image to use, the image which is referenced by this ID is used. If this option is not set, the image from the instance is used. Possible values:
Related options:
|
| string value | The ID of the kernel (AKI) image to use with the rescue image. If the chosen rescue image allows the separate definition of its kernel disk, the value of this option is used, if specified. This is the case when Amazon's AMI/AKI/ARI image format is used for the rescue image. Possible values:
Related options:
|
| string value | The ID of the RAM disk (ARI) image to use with the rescue image. If the chosen rescue image allows the separate definition of its RAM disk, the value of this option is used, if specified. This is the case when Amazon's AMI/AKI/ARI image format is used for the rescue image. Possible values:
Related options:
|
| string value |
The path to an RNG (Random Number Generator) device that will be used as the source of entropy on the host. Since libvirt 1.3.4, any path (that returns random numbers when read) is accepted. The recommended source of entropy is |
| integer value | Configure virtio rx queue size. This option is only usable for virtio-net device with vhost and vhost-user backend. Available only with QEMU/KVM. Requires libvirt v2.3 QEMU v2.7. |
`smbfs_mount_options = ` | string value | Mount options passed to the SMBFS client.
Provide SMBFS options as a single string containing all parameters. See mount.cifs man page for details. Note that the libvirt-qemu |
| string value | Directory where the SMBFS shares are mounted on the compute node. |
| boolean value |
Enable snapshot compression for
Note: you can set Related options:
|
| string value | Determine the snapshot image format when sending to the image service. If set, this decides what format is used when sending the snapshot to the image service. If not set, defaults to same type as source image. Possible values:
|
| string value | Location where libvirt driver will store snapshots before uploading them to image service |
| boolean value | Create sparse logical volumes (with virtualsize) if this flag is set to True. |
| string value | The data source used to the populate the host "serial" UUID exposed to guest in the virtual BIOS. |
| integer value | Configure virtio tx queue size. This option is only usable for virtio-net device with vhost-user backend. Available only with QEMU/KVM. Requires libvirt v3.7 QEMU v2.10. |
| list value | List of uid targets and ranges.Syntax is guest-uid:host-uid:countMaximum of 5 allowed. |
| boolean value | Enable a mouse cursor within a graphical VNC or SPICE sessions. This will only be taken into account if the VM is fully virtualized and VNC and/or SPICE is enabled. If the node doesn’t support a graphical framebuffer, then it is valid to set this to False.
Related options: * |
| boolean value | Use virtio for bridge interfaces with KVM/QEMU |
| string value | Describes the virtualization type (or so called domain type) libvirt should use. The choice of this type must match the underlying virtualization strategy you have chosen for this host. Possible values:
Related options:
|
| string value | Method used to wipe ephemeral disks when they are deleted. Only takes effect if LVM is set as backing storage. Possible values:
Related options:
|
| integer value |
Size of area in MiB, counting from the beginning of the allocated volume, that will be cleared using method set in Possible values:
Related options:
|
| boolean value | Use multipath connection of the iSCSI or FC volume Volumes can be connected in the LibVirt as multipath devices. This will provide high availability and fault tolerance. |
| string value | Path to the SSD cache file. You can attach an SSD drive to a client and configure the drive to store a local cache of frequently accessed data. By having a local cache on a client’s SSD drive, you can increase the overall cluster performance by up to 10 and more times. WARNING! There is a lot of SSD models which are not server grade and may loose arbitrary set of data changes on power loss. Such SSDs should not be used in Vstorage and are dangerous as may lead to data corruptions and inconsistencies. Please consult with the manual on which SSD models are known to be safe or verify it using vstorage-hwflush-check(1) utility. This option defines the path which should include "%(cluster_name)s" template to separate caches from multiple shares. Related options:
|
| string value | Path to vzstorage client log. This option defines the log of cluster operations, it should include "%(cluster_name)s" template to separate logs from multiple shares. Related options:
|
| string value | Mount owner group name. This option defines the owner group of Vzstorage cluster mountpoint. Related options:
|
| list value | Extra mount options for pstorage-mount For full description of them, see https://static.openvz.org/vz-man/man1/pstorage-mount.1.gz.html Format is a python string representation of arguments list, like: "[-v, -R, 500]" Shouldn’t include -c, -l, -C, -u, -g and -m as those have explicit vzstorage_* options. Related options:
|
| string value | Mount access mode. This option defines the access bits of Vzstorage cluster mountpoint, in the format similar to one of chmod(1) utility, like this: 0770. It consists of one to four digits ranging from 0 to 7, with missing lead digits assumed to be 0’s. Related options:
|
| string value | Directory where the Virtuozzo Storage clusters are mounted on the compute node. This option defines non-standard mountpoint for Vzstorage cluster. Related options:
|
| string value | Mount owner user name. This option defines the owner user of Vzstorage cluster mountpoint. Related options:
|
| integer value | Number of seconds to wait for instance to shut down after soft reboot request is made. We fall back to hard reboot if instance does not shutdown within this window. |
| string value | Location where the Xen hvmloader is kept |
11.1.27. matchmaker_redis
The following table outlines the options available under the [matchmaker_redis]
group in the /etc/nova/nova.conf
file.
Configuration option = Default value | Type | Description |
---|---|---|
| integer value | Time in ms to wait before the transaction is killed. |
| string value | Host to locate redis. |
`password = ` | string value | Password for Redis server (optional). |
| port value | Use this port to connect to redis host. |
| string value | Redis replica set name. |
| list value | List of Redis Sentinel hosts (fault tolerance mode), e.g., [host:port, host1:port … ] |
| integer value | Timeout in ms on blocking socket operations. |
| integer value | Time in ms to wait between connection attempts. |
11.1.28. metrics
The following table outlines the options available under the [metrics]
group in the /etc/nova/nova.conf
file.
Configuration option = Default value | Type | Description |
---|---|---|
| boolean value | This setting determines how any unavailable metrics are treated. If this option is set to True, any hosts for which a metric is unavailable will raise an exception, so it is recommended to also use the MetricFilter to filter out those hosts before weighing. This option is only used by the FilterScheduler and its subclasses; if you use a different scheduler, this option has no effect. Possible values:
Related options:
|
| floating point value | When using metrics to weight the suitability of a host, you can use this option to change how the calculated weight influences the weight assigned to a host as follows:
This option is only used by the FilterScheduler and its subclasses; if you use a different scheduler, this option has no effect. Possible values:
Related options:
|
| floating point value | When any of the following conditions are met, this value will be used in place of any actual metric value:
This option is only used by the FilterScheduler and its subclasses; if you use a different scheduler, this option has no effect. Possible values:
Related options:
|
| list value | This setting specifies the metrics to be weighed and the relative ratios for each metric. This should be a single string value, consisting of a series of one or more name=ratio pairs, separated by commas, where name is the name of the metric to be weighed, and ratio is the relative weight for that metric. Note that if the ratio is set to 0, the metric value is ignored, and instead the weight will be set to the value of the weight_of_unavailable option. As an example, let’s consider the case where this option is set to: `name1=1.0, name2=-1.3` The final weight will be: `(name1.value * 1.0) + (name2.value * -1.3)` This option is only used by the FilterScheduler and its subclasses; if you use a different scheduler, this option has no effect. Possible values:
Related options:
|
11.1.29. mks
The following table outlines the options available under the [mks]
group in the /etc/nova/nova.conf
file.
Configuration option = Default value | Type | Description |
---|---|---|
| boolean value | Enables graphical console access for virtual machines. |
| uri value | Location of MKS web console proxy The URL in the response points to a WebMKS proxy which starts proxying between client and corresponding vCenter server where instance runs. In order to use the web based console access, WebMKS proxy should be installed and configured Possible values:
|
11.1.30. neutron
The following table outlines the options available under the [neutron]
group in the /etc/nova/nova.conf
file.
Configuration option = Default value | Type | Description |
---|---|---|
| string value | Authentication URL |
| string value | Config Section from which to load plugin specific options |
| string value | Authentication type to load |
| string value | PEM encoded Certificate Authority to use when verifying HTTPs connections. |
| string value | PEM encoded client certificate cert file |
| string value | Optional domain ID to use with v3 and v2 parameters. It will be used for both the user and project domain in v3 and ignored in v2 authentication. |
| string value | Optional domain name to use with v3 API and v2 parameters. It will be used for both the user and project domain in v3 and ignored in v2 authentication. |
| string value | Default name for the floating IP pool. Specifies the name of floating IP pool used for allocating floating IPs. This option is only used if Neutron does not specify the floating IP pool name in port binding reponses. |
| string value | Domain ID to scope to |
| string value | Domain name to scope to |
| string value |
Always use this endpoint URL for requests for this client. NOTE: The unversioned endpoint should be specified here; to request a particular API version, use the |
| integer value | Integer value representing the number of seconds to wait before querying Neutron for extensions. After this number of seconds the next time Nova needs to create a resource in Neutron it will requery Neutron for the extensions that it has loaded. Setting value to 0 will refresh the extensions with no wait. |
| boolean value | Verify HTTPS connections. |
| string value | PEM encoded client certificate key file |
`metadata_proxy_shared_secret = ` | string value | This option holds the shared secret string used to validate proxy requests to Neutron metadata requests. In order to be used, the X-Metadata-Provider-Signature header must be supplied in the request. Related options:
|
| string value | Default name for the Open vSwitch integration bridge. Specifies the name of an integration bridge interface used by OpenvSwitch. This option is only used if Neutron does not specify the OVS bridge name in port binding responses. |
| string value | User’s password |
| list value | List of physnets present on this host.
For each physnet listed, an additional section, [neutron] physnets = foo, bar [neutron_physnet_foo] numa_nodes = 0 [neutron_physnet_bar] numa_nodes = 0,1 Any physnet that is not listed using this option will be treated as having no particular NUMA node affinity.
Tunnelled networks (VXLAN, GRE, …) cannot be accounted for in this way and are instead configured using the [neutron_tunnel] numa_nodes = 1 Related options:
|
| string value | Domain ID containing project |
| string value | Domain name containing project |
| string value | Project ID to scope to |
| string value | Project name to scope to |
| string value | The default region_name for endpoint URL discovery. |
| string value | The default service_name for endpoint URL discovery. |
| string value | The default service_type for endpoint URL discovery. |
| boolean value | When set to True, this option indicates that Neutron will be used to proxy metadata requests and resolve instance ids. Otherwise, the instance ID must be passed to the metadata request in the X-Instance-ID header. Related options:
|
| string value | Scope for system operations |
| string value | Tenant ID |
| string value | Tenant Name |
| integer value | Timeout value for http requests |
| string value | Trust ID |
| uri value | This option specifies the URL for connecting to Neutron. Possible values:
|
| string value | User’s domain id |
| string value | User’s domain name |
| string value | User ID |
| string value | Username |
| list value | List of interfaces, in order of preference, for endpoint URL. |
11.1.31. notifications
The following table outlines the options available under the [notifications]
group in the /etc/nova/nova.conf
file.
Configuration option = Default value | Type | Description |
---|---|---|
| boolean value | If enabled, include block device information in the versioned notification payload. Sending block device information is disabled by default as providing that information can incur some overhead on the system since the information may need to be loaded from the database. |
| string value | Default notification level for outgoing notifications. |
| string value | Default publisher_id for outgoing notifications. If you consider routing notifications using different publisher, change this value accordingly. Possible values:
Related options:
|
| string value | Specifies which notification format shall be used by nova. The default value is fine for most deployments and rarely needs to be changed. This value can be set to versioned once the infrastructure moves closer to consuming the newer format of notifications. After this occurs, this option will be removed.
Note that notifications can be completely disabled by setting Possible values: * unversioned: Only the legacy unversioned notifications are emitted. * versioned: Only the new versioned notifications are emitted. * both: Both the legacy unversioned and the new versioned notifications are emitted. (Default) The list of versioned notifications is visible in https://docs.openstack.org/nova/latest/reference/notifications.html |
| string value | If set, send compute.instance.update notifications on instance state changes. Please refer to https://docs.openstack.org/nova/latest/reference/notifications.html for additional information on notifications. Possible values:
|
| list value | Specifies the topics for the versioned notifications issued by nova. The default value is fine for most deployments and rarely needs to be changed. However, if you have a third-party service that consumes versioned notifications, it might be worth getting a topic for that service. Nova will send a message containing a versioned notification payload to each topic queue in this list. The list of versioned notifications is visible in https://docs.openstack.org/nova/latest/reference/notifications.html |
11.1.32. osapi_v21
The following table outlines the options available under the [osapi_v21]
group in the /etc/nova/nova.conf
file.
Configuration option = Default value | Type | Description |
---|---|---|
| string value | This option is a string representing a regular expression (regex) that matches the project_id as contained in URLs. If not set, it will match normal UUIDs created by keystone. Possible values:
|
11.1.33. oslo_concurrency
The following table outlines the options available under the [oslo_concurrency]
group in the /etc/nova/nova.conf
file.
Configuration option = Default value | Type | Description |
---|---|---|
| boolean value | Enables or disables inter-process locks. |
| string value | Directory to use for lock files. For security, the specified directory should only be writable by the user running the processes that need locking. Defaults to environment variable OSLO_LOCK_PATH. If external locks are used, a lock path must be set. |
11.1.34. oslo_messaging_amqp
The following table outlines the options available under the [oslo_messaging_amqp]
group in the /etc/nova/nova.conf
file.
Configuration option = Default value | Type | Description |
---|---|---|
| string value | Indicates the addressing mode used by the driver. Permitted values: legacy - use legacy non-routable addressing routable - use routable addresses dynamic - use legacy addresses if the message bus does not support routing otherwise use routable addressing |
| boolean value | Accept clients using either SSL or plain TCP |
| string value | Appended to the address prefix when sending to a group of consumers. Used by the message bus to identify messages that should be delivered in a round-robin fashion across consumers. |
| string value | address prefix used when broadcasting to all servers |
| integer value | Increase the connection_retry_interval by this many seconds after each unsuccessful failover attempt. |
| integer value | Seconds to pause before attempting to re-connect. |
| integer value | Maximum limit for connection_retry_interval + connection_retry_backoff |
| string value | Name for the AMQP container. must be globally unique. Defaults to a generated UUID |
| string value | Exchange name used in notification addresses. Exchange name resolution precedence: Target.exchange if set else default_notification_exchange if set else control_exchange if set else notify |
| integer value | The deadline for a sent notification message delivery. Only used when caller does not provide a timeout expiry. |
| integer value | The maximum number of attempts to re-send a reply message which failed due to a recoverable error. |
| integer value | The deadline for an rpc reply message delivery. |
| string value | Exchange name used in RPC addresses. Exchange name resolution precedence: Target.exchange if set else default_rpc_exchange if set else control_exchange if set else rpc |
| integer value | The deadline for an rpc cast or call message delivery. Only used when caller does not provide a timeout expiry. |
| integer value | The duration to schedule a purge of idle sender links. Detach link after expiry. |
| string value | address prefix when sending to any server in group |
| integer value | Timeout for inactive connections (in seconds) |
| integer value | Time to pause between re-connecting an AMQP 1.0 link that failed due to a recoverable error. |
| string value | Appended to the address prefix when sending a fanout message. Used by the message bus to identify fanout messages. |
| string value | Address prefix for all generated Notification addresses |
| integer value | Window size for incoming Notification messages |
`password = ` | string value | Password for message broker authentication |
| multi valued | Send messages of this type pre-settled. Pre-settled messages will not receive acknowledgement from the peer. Note well: pre-settled messages may be silently discarded if the delivery fails. Permitted values: rpc-call - send RPC Calls pre-settled rpc-reply- send RPC Replies pre-settled rpc-cast - Send RPC Casts pre-settled notify - Send Notifications pre-settled |
| boolean value | Enable virtual host support for those message buses that do not natively support virtual hosting (such as qpidd). When set to true the virtual host name will be added to all message bus addresses, effectively creating a private subnet per virtual host. Set to False if the message bus supports virtual hosting using the hostname field in the AMQP 1.0 Open performative as the name of the virtual host. |
| integer value | Window size for incoming RPC Reply messages. |
| string value | Address prefix for all generated RPC addresses |
| integer value | Window size for incoming RPC Request messages |
`sasl_config_dir = ` | string value | Path to directory that contains the SASL configuration |
`sasl_config_name = ` | string value | Name of configuration file (without .conf suffix) |
`sasl_default_realm = ` | string value | SASL realm to use if no realm present in username |
`sasl_mechanisms = ` | string value | Space separated list of acceptable SASL mechanisms |
| string value | address prefix used when sending to a specific server |
| boolean value | Attempt to connect via SSL. If no other ssl-related parameters are given, it will use the system’s CA-bundle to verify the server’s certificate. |
`ssl_ca_file = ` | string value | CA certificate PEM file used to verify the server’s certificate |
`ssl_cert_file = ` | string value | Self-identifying certificate PEM file for client authentication |
`ssl_key_file = ` | string value | Private key PEM file used to sign ssl_cert_file certificate (optional) |
| string value | Password for decrypting ssl_key_file (if encrypted) |
| boolean value | By default SSL checks that the name in the server’s certificate matches the hostname in the transport_url. In some configurations it may be preferable to use the virtual hostname instead, for example if the server uses the Server Name Indication TLS extension (rfc6066) to provide a certificate per virtual host. Set ssl_verify_vhost to True if the server’s SSL certificate uses the virtual host name instead of the DNS name. |
| boolean value | Debug: dump AMQP frames to stdout |
| string value | Appended to the address prefix when sending to a particular RPC/Notification server. Used by the message bus to identify messages sent to a single destination. |
`username = ` | string value | User name for message broker authentication |
11.1.35. oslo_messaging_kafka
The following table outlines the options available under the [oslo_messaging_kafka]
group in the /etc/nova/nova.conf
file.
Configuration option = Default value | Type | Description |
---|---|---|
| integer value | The pool size limit for connections expiration policy |
| integer value | The time-to-live in sec of idle connections in the pool |
| string value | Group id for Kafka consumer. Consumers in one group will coordinate message consumption |
| floating point value | Default timeout(s) for Kafka consumers |
| string value | Default Kafka broker Host |
| port value | Default Kafka broker Port |
| integer value | Max fetch bytes of Kafka consumer |
| integer value | Pool Size for Kafka Consumers |
| integer value | Size of batch for the producer async send |
| floating point value | Upper bound on the delay for KafkaProducer batching in seconds |
11.1.36. oslo_messaging_notifications
The following table outlines the options available under the [oslo_messaging_notifications]
group in the /etc/nova/nova.conf
file.
Configuration option = Default value | Type | Description |
---|---|---|
| multi valued | The Drivers(s) to handle sending notifications. Possible values are messaging, messagingv2, routing, log, test, noop |
| integer value | The maximum number of attempts to re-send a notification message which failed to be delivered due to a recoverable error. 0 - No retry, -1 - indefinite |
| list value | AMQP topic used for OpenStack notifications. |
| string value | A URL representing the messaging driver to use for notifications. If not set, we fall back to the same configuration used for RPC. |
11.1.37. oslo_messaging_rabbit
The following table outlines the options available under the [oslo_messaging_rabbit]
group in the /etc/nova/nova.conf
file.
Configuration option = Default value | Type | Description |
---|---|---|
| boolean value | Auto-delete queues in AMQP. |
| boolean value | Use durable queues in AMQP. |
| integer value | Maximum number of channels to allow |
| string value | Connection factory implementation |
| string value | Exchange name for sending notifications |
| integer value | Reconnecting retry count in case of connectivity problem during sending notification, -1 means infinite retry. |
| string value | Exchange name for sending RPC messages |
| integer value | Reconnecting retry count in case of connectivity problem during sending RPC message, -1 means infinite retry. If actual retry attempts in not 0 the rpc request could be processed more than one time |
| string value | Default serialization mechanism for serializing/deserializing outgoing/incoming messages |
| boolean value | Deprecated, use rpc_backend=kombu+memory or rpc_backend=fake |
| integer value | The maximum byte size for an AMQP frame |
| integer value | How often to send heartbeats for consumer’s connections |
| integer value | How often times during the heartbeat_timeout_threshold we check the heartbeat. |
| integer value | Number of seconds after which the Rabbit broker is considered down if heartbeat’s keep-alive fails (0 disable the heartbeat). EXPERIMENTAL |
| floating point value | Set delay for reconnection to some host which has connection error |
| string value | EXPERIMENTAL: Possible values are: gzip, bz2. If not set compression will not be used. This option may not be available in future versions. |
| string value | Determines how the next RabbitMQ node is chosen in case the one we are currently connected to becomes unavailable. Takes effect only if more than one RabbitMQ node is provided in config. |
| integer value | How long to wait a missing client before abandoning to send it its replies. This value should not be longer than rpc_response_timeout. |
| floating point value | How long to wait before reconnecting in response to an AMQP consumer cancel notification. |
| integer value | Max number of not acknowledged message which RabbitMQ can send to notification listener. |
| boolean value | Persist notification messages. |
| floating point value | Reconnecting retry delay in case of connectivity problem during sending notification message |
| integer value |
Maximum number of connections to create above |
| integer value | Maximum number of connections to keep queued. |
| integer value | Lifetime of a connection (since creation) in seconds or None for no recycling. Expired connections are closed on acquire. |
| integer value | Threshold at which inactive (since release) connections are considered stale in seconds or None for no staleness. Stale connections are closed on acquire. |
| integer value | Default number of seconds to wait for a connections to available |
| boolean value | Try to use HA queues in RabbitMQ (x-ha-policy: all). If you change this option, you must wipe the RabbitMQ database. In RabbitMQ 3.0, queue mirroring is no longer controlled by the x-ha-policy argument when declaring a queue. If you just want to make sure that all queues (except those with auto-generated names) are mirrored across all nodes, run: "rabbitmqctl set_policy HA ^(?!amq\.).* {"ha-mode": "all"} " |
| string value | The RabbitMQ broker address where a single node is used. |
| list value | RabbitMQ HA cluster host:port pairs. |
| integer value | Maximum interval of RabbitMQ connection retries. Default is 30 seconds. |
| string value | The RabbitMQ login method. |
| integer value | Maximum number of RabbitMQ connection retries. Default is 0 (infinite retry count). |
| string value | The RabbitMQ password. |
| port value | The RabbitMQ broker port where a single node is used. |
| integer value | Specifies the number of messages to prefetch. Setting to zero allows unlimited messages. |
| integer value | How long to backoff for between retries when connecting to RabbitMQ. |
| integer value | How frequently to retry connecting with RabbitMQ. |
| integer value | Positive integer representing duration in seconds for queue TTL (x-expires). Queues which are unused for the duration of the TTL are automatically deleted. The parameter affects only reply and fanout queues. |
| string value | The RabbitMQ userid. |
| string value | The RabbitMQ virtual host. |
| integer value | Max number of not acknowledged message which RabbitMQ can send to rpc listener. |
| integer value | Time to live for rpc queues without consumers in seconds. |
| string value | Exchange name for receiving RPC replies |
| integer value | Max number of not acknowledged message which RabbitMQ can send to rpc reply listener. |
| integer value | Reconnecting retry count in case of connectivity problem during sending reply. -1 means infinite retry during rpc_timeout |
| floating point value | Reconnecting retry delay in case of connectivity problem during sending reply. |
| floating point value | Reconnecting retry delay in case of connectivity problem during sending RPC message |
| floating point value | Set socket timeout in seconds for connection’s socket |
| boolean value | Enable SSL |
`ssl_ca_file = ` | string value | SSL certification authority file (valid only if SSL enabled). |
`ssl_cert_file = ` | string value | SSL cert file (valid only if SSL enabled). |
`ssl_key_file = ` | string value | SSL key file (valid only if SSL enabled). |
| dict value | Arguments passed to ssl.wrap_socket |
`ssl_version = ` | string value | SSL version to use (valid only if SSL enabled). Valid values are TLSv1 and SSLv23. SSLv2, SSLv3, TLSv1_1, and TLSv1_2 may be available on some distributions. |
| floating point value | Set TCP_USER_TIMEOUT in seconds for connection’s socket |
11.1.38. oslo_messaging_zmq
The following table outlines the options available under the [oslo_messaging_zmq]
group in the /etc/nova/nova.conf
file.
Configuration option = Default value | Type | Description |
---|---|---|
| integer value | Number of seconds to wait for an ack from a cast/call. After each retry attempt this timeout is multiplied by some specified multiplier. |
| integer value | Number to multiply base ack timeout by after each retry attempt. |
| integer value | Expiration timeout in seconds of a sent/received message after which it is not tracked anymore by a client/server. |
| integer value | The default number of seconds that poll should wait. Poll raises timeout exception when timeout expired. |
| integer value | Default number of message sending attempts in case of any problems occurred: positive value N means at most N retries, 0 means no retries, None or -1 (or any other negative values) mean to retry forever. This option is used only if acknowledgments are enabled. |
| integer value | Maximum number of (green) threads to work concurrently. |
| boolean value | Wait for message acknowledgements from receivers. This mechanism works only via proxy without PUB/SUB. |
| string value | ZeroMQ bind address. Should be a wildcard (*), an ethernet interface, or IP. The "host" option should point or resolve to this address. |
| integer value | Number of retries to find free port number before fail with ZMQBindError. |
| integer value | Number of ZeroMQ contexts, defaults to 1. |
| string value | Name of this node. Must be a valid hostname, FQDN, or IP address. Must match "host" option, if running Nova. |
| string value | Directory for holding IPC sockets. |
| string value | MatchMaker driver. |
| integer value | Maximal port number for random ports range. |
| port value | Minimal port number for random ports range. |
| string value | Default serialization mechanism for serializing/deserializing outgoing/incoming messages |
| integer value | Maximum number of ingress messages to locally buffer per topic. Default is unlimited. |
| list value | List of publisher hosts SubConsumer can subscribe on. This option has higher priority then the default publishers list taken from the matchmaker. |
| boolean value | This option makes direct connections dynamic or static. It makes sense only with use_router_proxy=False which means to use direct connections for direct message types (ignored otherwise). |
| boolean value | Use PUB/SUB pattern for fanout methods. PUB/SUB always uses proxy. |
| boolean value | Use ROUTER remote proxy. |
| integer value | How many additional connections to a host will be made for failover reasons. This option is actual only in dynamic connections mode. |
| boolean value | This option configures round-robin mode in zmq socket. True means not keeping a queue when server side disconnects. False means to keep queue and messages even if server is disconnected, when the server appears we send all accumulated messages to it. |
| integer value | Number of seconds to wait before all pending messages will be sent after closing a socket. The default value of -1 specifies an infinite linger period. The value of 0 specifies no linger period. Pending messages shall be discarded immediately when the socket is closed. Positive values specify an upper bound for the linger period. |
| integer value | Expiration timeout in seconds of a name service record about existing target ( < 0 means no timeout). |
| integer value | Update period in seconds of a name service record about existing target. |
| integer value | Enable/disable TCP keepalive (KA) mechanism. The default value of -1 (or any other negative value) means to skip any overrides and leave it to OS default; 0 and 1 (or any other positive value) mean to disable and enable the option respectively. |
| integer value | The number of retransmissions to be carried out before declaring that remote end is not available. The default value of -1 (or any other negative value and 0) means to skip any overrides and leave it to OS default. |
| integer value | The duration between two keepalive transmissions in idle condition. The unit is platform dependent, for example, seconds in Linux, milliseconds in Windows etc. The default value of -1 (or any other negative value and 0) means to skip any overrides and leave it to OS default. |
| integer value | The duration between two successive keepalive retransmissions, if acknowledgement to the previous keepalive transmission is not received. The unit is platform dependent, for example, seconds in Linux, milliseconds in Windows etc. The default value of -1 (or any other negative value and 0) means to skip any overrides and leave it to OS default. |
11.1.39. oslo_middleware
The following table outlines the options available under the [oslo_middleware]
group in the /etc/nova/nova.conf
file.
Configuration option = Default value | Type | Description |
---|---|---|
| boolean value | Whether the application is behind a proxy or not. This determines if the middleware should parse the headers or not. |
| integer value | The maximum body size for each request, in bytes. |
| string value | The HTTP Header that will be used to determine what the original request protocol scheme was, even if it was hidden by a SSL termination proxy. |
11.1.40. oslo_policy
The following table outlines the options available under the [oslo_policy]
group in the /etc/nova/nova.conf
file.
Configuration option = Default value | Type | Description |
---|---|---|
| boolean value |
This option controls whether or not to enforce scope when evaluating policies. If |
| string value | Default rule. Enforced when a requested rule is not found. |
| multi valued | Directories where policy configuration files are stored. They can be relative to any directory in the search path defined by the config_dir option, or absolute paths. The file defined by policy_file must exist for these directories to be searched. Missing or empty directories are ignored. |
| string value | The file that defines policies. |
| string value | Content Type to send and receive data for REST based policy check |
| string value | Absolute path to ca cert file for REST based policy check |
| string value | Absolute path to client cert for REST based policy check |
| string value | Absolute path client key file REST based policy check |
| boolean value | server identity verification for REST based policy check |
11.1.41. pci
The following table outlines the options available under the [pci]
group in the /etc/nova/nova.conf
file.
Configuration option = Default value | Type | Description |
---|---|---|
| multi valued | An alias for a PCI passthrough device requirement. This allows users to specify the alias in the extra specs for a flavor, without needing to repeat all the PCI property requirements. Possible Values:
|
| multi valued | White list of PCI devices available to VMs. Possible values:
|
11.1.42. placement
The following table outlines the options available under the [placement]
group in the /etc/nova/nova.conf
file.
Configuration option = Default value | Type | Description |
---|---|---|
| string value | Authentication URL |
| string value | Config Section from which to load plugin specific options |
| string value | Authentication type to load |
| string value | PEM encoded Certificate Authority to use when verifying HTTPs connections. |
| string value | PEM encoded client certificate cert file |
| string value | Optional domain ID to use with v3 and v2 parameters. It will be used for both the user and project domain in v3 and ignored in v2 authentication. |
| string value | Optional domain name to use with v3 API and v2 parameters. It will be used for both the user and project domain in v3 and ignored in v2 authentication. |
| string value | Domain ID to scope to |
| string value | Domain name to scope to |
| string value |
Always use this endpoint URL for requests for this client. NOTE: The unversioned endpoint should be specified here; to request a particular API version, use the |
| boolean value | Verify HTTPS connections. |
| string value | PEM encoded client certificate key file |
| string value | Endpoint interface for this node. This is used when picking the URL in the service catalog. |
| string value | Region name of this node. This is used when picking the URL in the service catalog. Possible values:
|
| string value | User’s password |
| string value | Domain ID containing project |
| string value | Domain name containing project |
| string value | Project ID to scope to |
| string value | Project name to scope to |
| boolean value | If True, when limiting allocation candidate results, the results will be a random sampling of the full result set. If False, allocation candidates are returned in a deterministic but undefined order. That is, all things being equal, two requests for allocation candidates will return the same results in the same order; but no guarantees are made as to how that order is determined. |
| string value | The default region_name for endpoint URL discovery. |
| string value | The default service_name for endpoint URL discovery. |
| string value | The default service_type for endpoint URL discovery. |
| string value | Scope for system operations |
| string value | Tenant ID |
| string value | Tenant Name |
| integer value | Timeout value for http requests |
| string value | Trust ID |
| string value | User’s domain id |
| string value | User’s domain name |
| string value | User ID |
| string value | Username |
| list value | List of interfaces, in order of preference, for endpoint URL. |
11.1.43. quota
The following table outlines the options available under the [quota]
group in the /etc/nova/nova.conf
file.
Configuration option = Default value | Type | Description |
---|---|---|
| integer value | The number of instance cores or vCPUs allowed per project. Possible values:
|
| string value | The quota enforcer driver. Provides abstraction for quota checks. Users can configure a specific driver to use for quota checks. Possible values:
|
| integer value | The number of fixed IPs allowed per project. Unlike floating IPs, fixed IPs are allocated dynamically by the network component when instances boot up. This quota value should be at least the number of instances allowed Possible values:
|
| integer value | The number of floating IPs allowed per project. Floating IPs are not allocated to instances by default. Users need to select them from the pool configured by the OpenStack administrator to attach to their instances. Possible values:
|
| integer value | The number of bytes allowed per injected file. Possible values:
|
| integer value | The maximum allowed injected file path length. Possible values:
|
| integer value | The number of injected files allowed.
File injection allows users to customize the personality of an instance by injecting data into it upon boot. Only text file injection is permitted: binary or ZIP files are not accepted. During file injection, any existing files that match specified files are renamed to include Possible values:
|
| integer value | The number of instances allowed per project. Possible Values
|
| integer value | The maximum number of key pairs allowed per user. Users can create at least one key pair for each project and use the key pair for multiple instances that belong to that project. Possible values:
|
| integer value | The number of seconds between subsequent usage refreshes. This defaults to 0 (off) to avoid additional load but it is useful to turn on to help keep quota usage up-to-date and reduce the impact of out of sync usage issues. Note that quotas are not updated on a periodic task, they will update on a new reservation if max_age has passed since the last reservation. |
| integer value | The number of metadata items allowed per instance. Users can associate metadata with an instance during instance creation. This metadata takes the form of key-value pairs. Possible values:
|
| integer value | The number of megabytes of instance RAM allowed per project. Possible values:
|
| boolean value | Recheck quota after resource creation to prevent allowing quota to be exceeded. This defaults to True (recheck quota after resource creation) but can be set to False to avoid additional load if allowing quota to be exceeded because of racing requests is considered acceptable. For example, when set to False, if a user makes highly parallel REST API requests to create servers, it will be possible for them to create more servers than their allowed quota during the race. If their quota is 10 servers, they might be able to create 50 during the burst. After the burst, they will not be able to create any more servers but they will be able to keep their 50 servers until they delete them. The initial quota check is done before resources are created, so if multiple parallel requests arrive at the same time, all could pass the quota check and create resources, potentially exceeding quota. When recheck_quota is True, quota will be checked a second time after resources have been created and if the resource is over quota, it will be deleted and OverQuota will be raised, usually resulting in a 403 response to the REST API user. This makes it impossible for a user to exceed their quota with the caveat that it will, however, be possible for a REST API user to be rejected with a 403 response in the event of a collision close to reaching their quota limit, even if the user has enough quota available when they made the request. |
| integer value | The number of seconds until a reservation expires. This quota represents the time period for invalidating quota reservations. |
| integer value | The number of security rules per security group. The associated rules in each security group control the traffic to instances in the group. Possible values:
|
| integer value | The number of security groups per project. Possible values:
|
| integer value | The maximum number of servers per server group. Possible values:
|
| integer value | The maxiumum number of server groups per project. Server groups are used to control the affinity and anti-affinity scheduling policy for a group of servers or instances. Reducing the quota will not affect any existing group, but new servers will not be allowed into groups that have become over quota. Possible values:
|
| integer value | The count of reservations until usage is refreshed. This defaults to 0 (off) to avoid additional load but it is useful to turn on to help keep quota usage up-to-date and reduce the impact of out of sync usage issues. |
11.1.44. rdp
The following table outlines the options available under the [rdp]
group in the /etc/nova/nova.conf
file.
Configuration option = Default value | Type | Description |
---|---|---|
| boolean value | Enable Remote Desktop Protocol (RDP) related features. Hyper-V, unlike the majority of the hypervisors employed on Nova compute nodes, uses RDP instead of VNC and SPICE as a desktop sharing protocol to provide instance console access. This option enables RDP for graphical console access for virtual machines created by Hyper-V. Note: RDP should only be enabled on compute nodes that support the Hyper-V virtualization platform. Related options:
|
| uri value | The URL an end user would use to connect to the RDP HTML5 console proxy. The console proxy service is called with this token-embedded URL and establishes the connection to the proper instance. An RDP HTML5 console proxy service will need to be configured to listen on the address configured here. Typically the console proxy service would be run on a controller node. The localhost address used as default would only work in a single node environment i.e. devstack. An RDP HTML5 proxy allows a user to access via the web the text or graphical console of any Windows server or workstation using RDP. RDP HTML5 console proxy services include FreeRDP, wsgate. See https://github.com/FreeRDP/FreeRDP-WebConnect Possible values:
Related options:
|
11.1.45. remote_debug
The following table outlines the options available under the [remote_debug]
group in the /etc/nova/nova.conf
file.
Configuration option = Default value | Type | Description |
---|---|---|
| unknown value | Debug host (IP or name) to connect to. This command line parameter is used when you want to connect to a nova service via a debugger running on a different host. Note that using the remote debug option changes how Nova uses the eventlet library to support async IO. This could result in failures that do not occur under normal operation. Use at your own risk. Possible Values:
|
| port value | Debug port to connect to. This command line parameter allows you to specify the port you want to use to connect to a nova service via a debugger running on different host. Note that using the remote debug option changes how Nova uses the eventlet library to support async IO. This could result in failures that do not occur under normal operation. Use at your own risk. Possible Values:
|
11.1.46. scheduler
The following table outlines the options available under the [scheduler]
group in the /etc/nova/nova.conf
file.
Configuration option = Default value | Type | Description |
---|---|---|
| integer value | Periodic task interval. This value controls how often (in seconds) the scheduler should attempt to discover new hosts that have been added to cells. If negative (the default), no automatic discovery will occur. Deployments where compute nodes come and go frequently may want this enabled, where others may prefer to manually discover hosts when one is added to avoid any overhead from constantly checking. If enabled, every time this runs, we will select any unmapped hosts out of each cell database on every run. |
| string value | The class of the driver used by the scheduler. This should be chosen from one of the entrypoints under the namespace nova.scheduler.driver of file setup.cfg. If nothing is specified in this option, the filter_scheduler is used. Other options are:
Possible values:
Related options:
|
| string value | The scheduler host manager to use. The host manager manages the in-memory picture of the hosts that the scheduler uses. The options values are chosen from the entry points under the namespace nova.scheduler.host_manager in setup.cfg. Note The "ironic_host_manager" option is deprecated as of the 17.0.0 Queens release. |
| integer value | This is the maximum number of attempts that will be made for a given instance build/move operation. It limits the number of alternate hosts returned by the scheduler. When that list of hosts is exhausted, a MaxRetriesExceeded exception is raised and the instance is set to an error state. Possible values:
|
| integer value | This setting determines the maximum limit on results received from the placement service during a scheduling operation. It effectively limits the number of hosts that may be considered for scheduling requests that match a large number of candidates. A value of 1 (the minimum) will effectively defer scheduling to the placement service strictly on "will it fit" grounds. A higher value will put an upper cap on the number of results the scheduler will consider during the filtering and weighing process. Large deployments may need to set this lower than the total number of hosts available to limit memory consumption, network traffic, etc. of the scheduler. This option is only used by the FilterScheduler; if you use a different scheduler, this option has no effect. |
| integer value | Periodic task interval. This value controls how often (in seconds) to run periodic tasks in the scheduler. The specific tasks that are run for each period are determined by the particular scheduler being used. If this is larger than the nova-service service_down_time setting, Nova may report the scheduler service as down. This is because the scheduler driver is responsible for sending a heartbeat and it will only do that as often as this option allows. As each scheduler can work a little differently than the others, be sure to test this with your selected scheduler. Possible values:
Related options:
|
| integer value | Number of workers for the nova-scheduler service. The default will be the number of CPUs available if using the "filter_scheduler" scheduler driver, otherwise the default will be 1. |
11.1.47. serial_console
The following table outlines the options available under the [serial_console]
group in the /etc/nova/nova.conf
file.
Configuration option = Default value | Type | Description |
---|---|---|
| uri value |
The URL an end user would use to connect to the
The Related options:
|
| boolean value | Enable the serial console feature.
In order to use this feature, the service |
| string value | A range of TCP ports a guest can use for its backend. Each instance which gets created will use one port out of this range. If the range is not big enough to provide another port for an new instance, this instance won’t get launched. Possible values:
|
| string value |
The IP address to which proxy clients (like
This is typically the IP address of the host of a |
| string value |
The IP address which is used by the
The Related options:
|
| port value |
The port number which is used by the
The Related options:
|
11.1.48. service_user
The following table outlines the options available under the [service_user]
group in the /etc/nova/nova.conf
file.
Configuration option = Default value | Type | Description |
---|---|---|
| string value | Authentication URL |
| string value | Config Section from which to load plugin specific options |
| string value | Authentication type to load |
| string value | PEM encoded Certificate Authority to use when verifying HTTPs connections. |
| string value | PEM encoded client certificate cert file |
| string value | Optional domain ID to use with v3 and v2 parameters. It will be used for both the user and project domain in v3 and ignored in v2 authentication. |
| string value | Optional domain name to use with v3 API and v2 parameters. It will be used for both the user and project domain in v3 and ignored in v2 authentication. |
| string value | Domain ID to scope to |
| string value | Domain name to scope to |
| boolean value | Verify HTTPS connections. |
| string value | PEM encoded client certificate key file |
| string value | User’s password |
| string value | Domain ID containing project |
| string value | Domain name containing project |
| string value | Project ID to scope to |
| string value | Project name to scope to |
| boolean value | When True, if sending a user token to a REST API, also send a service token. Nova often reuses the user token provided to the nova-api to talk to other REST APIs, such as Cinder, Glance and Neutron. It is possible that while the user token was valid when the request was made to Nova, the token may expire before it reaches the other service. To avoid any failures, and to make it clear it is Nova calling the service on the user’s behalf, we include a service token along with the user token. Should the user’s token have expired, a valid service token ensures the REST API request will still be accepted by the keystone middleware. |
| string value | Scope for system operations |
| string value | Tenant ID |
| string value | Tenant Name |
| integer value | Timeout value for http requests |
| string value | Trust ID |
| string value | User’s domain id |
| string value | User’s domain name |
| string value | User ID |
| string value | Username |
11.1.49. spice
The following table outlines the options available under the [spice]
group in the /etc/nova/nova.conf
file.
Configuration option = Default value | Type | Description |
---|---|---|
| boolean value | Enable the SPICE guest agent support on the instances. The Spice agent works with the Spice protocol to offer a better guest console experience. However, the Spice console can still be used without the Spice Agent. With the Spice agent installed the following features are enabled:
|
| boolean value | Enable SPICE related features. Related options:
|
| uri value | Location of the SPICE HTML5 console proxy.
End user would use this URL to connect to the
In order to use SPICE console, the service Possible values:
Related options:
|
| unknown value |
IP address or a hostname on which the Related options:
|
| port value |
Port on which the Related options:
|
| string value | A keyboard layout which is supported by the underlying hypervisor on this node. Possible values: * This is usually an IETF language tag (default is en-us). If you use QEMU as hypervisor, you should find the list of supported keyboard layouts at /usr/share/qemu/keymaps. |
| string value | The address where the SPICE server running on the instances should listen.
Typically, the Possible values:
|
| string value |
The address used by
Typically, the Possible values:
Related options:
|
11.1.50. upgrade_levels
The following table outlines the options available under the [upgrade_levels]
group in the /etc/nova/nova.conf
file.
Configuration option = Default value | Type | Description |
---|---|---|
| string value | Base API RPC API version cap |
| string value | Cells RPC API version cap |
| string value | Cert RPC API version cap |
| string value | Compute RPC API version cap. By default, we always send messages using the most recent version the client knows about. Where you have old and new compute services running, you should set this to the lowest deployed version. This is to guarantee that all services never send messages that one of the compute nodes can’t understand. Note that we only support upgrading from release N to release N+1. Set this option to "auto" if you want to let the compute RPC module automatically determine what version to use based on the service versions in the deployment. Possible values:
|
| string value | Conductor RPC API version cap |
| string value | Console RPC API version cap |
| string value | Consoleauth RPC API version cap |
| string value | Intercell RPC API version cap |
| string value | Network RPC API version cap |
| string value | Scheduler RPC API version cap |
11.1.51. vault
The following table outlines the options available under the [vault]
group in the /etc/nova/nova.conf
file.
Configuration option = Default value | Type | Description |
---|---|---|
| string value | root token for vault |
| string value | Absolute path to ca cert file |
| boolean value | SSL Enabled/Disabled |
| string value | Use this endpoint to connect to Vault, for example: "http://127.0.0.1:8200" |
11.1.52. vendordata_dynamic_auth
The following table outlines the options available under the [vendordata_dynamic_auth]
group in the /etc/nova/nova.conf
file.
Configuration option = Default value | Type | Description |
---|---|---|
| string value | Authentication URL |
| string value | Config Section from which to load plugin specific options |
| string value | Authentication type to load |
| string value | PEM encoded Certificate Authority to use when verifying HTTPs connections. |
| string value | PEM encoded client certificate cert file |
| string value | Optional domain ID to use with v3 and v2 parameters. It will be used for both the user and project domain in v3 and ignored in v2 authentication. |
| string value | Optional domain name to use with v3 API and v2 parameters. It will be used for both the user and project domain in v3 and ignored in v2 authentication. |
| string value | Domain ID to scope to |
| string value | Domain name to scope to |
| boolean value | Verify HTTPS connections. |
| string value | PEM encoded client certificate key file |
| string value | User’s password |
| string value | Domain ID containing project |
| string value | Domain name containing project |
| string value | Project ID to scope to |
| string value | Project name to scope to |
| string value | Scope for system operations |
| string value | Tenant ID |
| string value | Tenant Name |
| integer value | Timeout value for http requests |
| string value | Trust ID |
| string value | User’s domain id |
| string value | User’s domain name |
| string value | User ID |
| string value | Username |
11.1.53. vmware
The following table outlines the options available under the [vmware]
group in the /etc/nova/nova.conf
file.
Configuration option = Default value | Type | Description |
---|---|---|
| integer value | Number of times VMware vCenter server API must be retried on connection failures, e.g. socket error, etc. |
| string value | Specifies the CA bundle file to be used in verifying the vCenter server certificate. |
| string value | This option adds a prefix to the folder where cached images are stored This is not the full path - just a folder prefix. This should only be used when a datastore cache is shared between compute nodes. Note: This should only be used when the compute nodes are running on same host or they have a shared file system. Possible values:
|
| string value | Name of a VMware Cluster ComputeResource. |
| integer value | This option sets the http connection pool size The connection pool size is the maximum number of connections from nova to vSphere. It should only be increased if there are warnings indicating that the connection pool is full, otherwise, the default should suffice. |
| integer value | Set this value if affected by an increased network latency causing repeated characters when typing in a remote console. |
| string value | Regular expression pattern to match the name of datastore. The datastore_regex setting specifies the datastores to use with Compute. For example, datastore_regex="nas.*" selects all the data stores that have a name starting with "nas". Note If no regex is given, it just picks the datastore with the most freespace. Possible values:
|
| unknown value | Hostname or IP address for connection to VMware vCenter host. |
| string value | Password for connection to VMware vCenter host. |
| port value | Port for connection to VMware vCenter host. |
| string value | Username for connection to VMware vCenter host. |
| boolean value | If true, the vCenter server certificate is not verified. If false, then the default CA truststore is used for verification. Related options: * ca_file: This option is ignored if "ca_file" is set. |
| string value | This option should be configured only when using the NSX-MH Neutron plugin. This is the name of the integration bridge on the ESXi server or host. This should not be set for any other Neutron plugin. Hence the default value is not set. Possible values:
|
| integer value | This option specifies the limit on the maximum number of objects to return in a single result. A positive value will cause the operation to suspend the retrieval when the count of objects reaches the specified limit. The server may still limit the count to something less than the configured value. Any remaining objects may be retrieved with additional requests. |
| string value | This option specifies the default policy to be used. If pbm_enabled is set and there is no defined storage policy for the specific request, then this policy will be used. Possible values:
Related options:
|
| boolean value | This option enables or disables storage policy based placement of instances. Related options:
|
| string value | This option specifies the PBM service WSDL file location URL. Setting this will disable storage policy based placement of instances. Possible values:
|
| string value | Specifies the directory where the Virtual Serial Port Concentrator is storing console log files. It should match the serial_log_dir config value of VSPC. |
| uri value | Identifies a proxy service that provides network access to the serial_port_service_uri. Possible values:
Related options: This option is ignored if serial_port_service_uri is not specified. * serial_port_service_uri |
| string value | Identifies the remote system where the serial port traffic will be sent. This option adds a virtual serial port which sends console output to a configurable service URI. At the service URI address there will be virtual serial port concentrator that will collect console logs. If this is not set, no serial ports will be added to the created VMs. Possible values:
|
| floating point value | Time interval in seconds to poll remote tasks invoked on VMware VC server. |
| boolean value | This option enables/disables the use of linked clone. The ESX hypervisor requires a copy of the VMDK file in order to boot up a virtual machine. The compute driver must download the VMDK via HTTP from the OpenStack Image service to a datastore that is visible to the hypervisor and cache it. Subsequent virtual machines that need the VMDK use the cached version and don’t have to copy the file again from the OpenStack Image service. If set to false, even with a cached VMDK, there is still a copy operation from the cache location to the hypervisor file directory in the shared datastore. If set to true, the above copy operation is avoided as it creates copy of the virtual machine that shares virtual disks with its parent VM. |
| string value | This option specifies the physical ethernet adapter name for VLAN networking. Set the vlan_interface configuration option to match the ESX host interface that handles VLAN-tagged VM traffic. Possible values:
|
| port value | This option specifies VNC starting port. Every VM created by ESX host has an option of enabling VNC client for remote connection. Above option vnc_port helps you to set default starting port for the VNC client. Possible values:
Related options: Below options should be set to enable VNC client. * vnc.enabled = True * vnc_port_total |
| integer value | Total number of VNC ports. |
11.1.54. vnc
The following table outlines the options available under the [vnc]
group in the /etc/nova/nova.conf
file.
Configuration option = Default value | Type | Description |
---|---|---|
| list value | The authentication schemes to use with the compute node. Control what RFB authentication schemes are permitted for connections between the proxy and the compute host. If multiple schemes are enabled, the first matching scheme will be used, thus the strongest schemes should be listed first. Possible values:
Related options:
|
| boolean value | Enable VNC related features. Guests will get created with graphical devices to support this. Clients (for example Horizon) can then establish a VNC connection to the guest. |
| string value | Keymap for VNC. The keyboard mapping (keymap) determines which keyboard layout a VNC session should use by default. Possible values:
|
| uri value | Public address of noVNC VNC console proxy. The VNC proxy is an OpenStack component that enables compute service users to access their instances through VNC clients. noVNC provides VNC support through a websocket-based client. This option sets the public base URL to which client systems will connect. noVNC clients can use this address to connect to the noVNC instance and, by extension, the VNC sessions. Related options:
|
| string value | IP address that the noVNC console proxy should bind to. The VNC proxy is an OpenStack component that enables compute service users to access their instances through VNC clients. noVNC provides VNC support through a websocket-based client. This option sets the private address to which the noVNC console proxy service should bind to. Related options:
|
| port value | Port that the noVNC console proxy should bind to. The VNC proxy is an OpenStack component that enables compute service users to access their instances through VNC clients. noVNC provides VNC support through a websocket-based client. This option sets the private port to which the noVNC console proxy service should bind to. Related options:
|
| unknown value | The IP address or hostname on which an instance should listen to for incoming VNC connection requests on this node. |
| unknown value | Private, internal IP address or hostname of VNC console proxy. The VNC proxy is an OpenStack component that enables compute service users to access their instances through VNC clients.
This option sets the private address to which proxy clients, such as |
| string value | The path to the CA certificate PEM file The fully qualified path to a PEM file containing one or more x509 certificates for the certificate authorities used by the compute node VNC server. Related options:
|
| string value | The path to the client key file (for x509) The fully qualified path to a PEM file containing the x509 certificate which the VNC proxy server presents to the compute node during VNC authentication. Realted options:
|
| string value | The path to the client certificate PEM file (for x509) The fully qualified path to a PEM file containing the private key which the VNC proxy server presents to the compute node during VNC authentication. Related options:
|
| uri value | Public URL address of XVP VNC console proxy. The VNC proxy is an OpenStack component that enables compute service users to access their instances through VNC clients. Xen provides the Xenserver VNC Proxy, or XVP, as an alternative to the websocket-based noVNC proxy used by Libvirt. In contrast to noVNC, XVP clients are Java-based. This option sets the public base URL to which client systems will connect. XVP clients can use this address to connect to the XVP instance and, by extension, the VNC sessions. Related options:
|
| unknown value | IP address or hostname that the XVP VNC console proxy should bind to. The VNC proxy is an OpenStack component that enables compute service users to access their instances through VNC clients. Xen provides the Xenserver VNC Proxy, or XVP, as an alternative to the websocket-based noVNC proxy used by Libvirt. In contrast to noVNC, XVP clients are Java-based. This option sets the private address to which the XVP VNC console proxy service should bind to. Related options:
|
| port value | Port that the XVP VNC console proxy should bind to. The VNC proxy is an OpenStack component that enables compute service users to access their instances through VNC clients. Xen provides the Xenserver VNC Proxy, or XVP, as an alternative to the websocket-based noVNC proxy used by Libvirt. In contrast to noVNC, XVP clients are Java-based. This option sets the private port to which the XVP VNC console proxy service should bind to. Related options:
|
11.1.55. workarounds
The following table outlines the options available under the [workarounds]
group in the /etc/nova/nova.conf
file.
Configuration option = Default value | Type | Description |
---|---|---|
| boolean value | Disable the server group policy check upcall in compute. In order to detect races with server group affinity policy, the compute service attempts to validate that the policy was not violated by the scheduler. It does this by making an upcall to the API database to list the instances in the server group for one that it is booting, which violates our api/cell isolation goals. Eventually this will be solved by proper affinity guarantees in the scheduler and placement service, but until then, this late check is needed to ensure proper affinity policy. Operators that desire api/cell isolation over this check should enable this flag, which will avoid making that upcall from compute. Related options:
|
| boolean value | Disable live snapshots when using the libvirt driver. Live snapshots allow the snapshot of the disk to happen without an interruption to the guest, using coordination with a guest agent to quiesce the filesystem. When using libvirt 1.2.2 live snapshots fail intermittently under load (likely related to concurrent libvirt/qemu operations). This config option provides a mechanism to disable live snapshot, in favor of cold snapshot, while this is resolved. Cold snapshot causes an instance outage while the guest is going through the snapshotting process. For more information, refer to the bug report: https://bugs.launchpad.net/nova/+bug/1334398 Possible values:
|
| boolean value | Use sudo instead of rootwrap. Allow fallback to sudo for performance reasons. For more information, refer to the bug report: https://bugs.launchpad.net/nova/+bug/1415106 Possible values:
Interdependencies to other options:
|
| boolean value | Ensure the instance directory is removed during clean up when using rbd.
When enabled this workaround will ensure that the instance directory is always removed during cleanup on hosts using https://bugs.launchpad.net/nova/+bug/1414895 https://bugs.launchpad.net/nova/+bug/1761062
Both of these bugs can then result in
Related options:
|
| boolean value | Enable handling of events emitted from compute drivers. Many compute drivers emit lifecycle events, which are events that occur when, for example, an instance is starting or stopping. If the instance is going through task state changes due to an API operation, like resize, the events are ignored. This is an advanced feature which allows the hypervisor to signal to the compute service that an unexpected state change has occurred in an instance and that the instance can be shutdown automatically. Unfortunately, this can race in some conditions, for example in reboot operations or when the compute service or when host is rebooted (planned or due to an outage). If such races are common, then it is advisable to disable this feature. Care should be taken when this feature is disabled and sync_power_state_interval is set to a negative value. In this case, any instances that get out of sync between the hypervisor and the Nova database will have to be synchronized manually. For more information, refer to the bug report: https://bugs.launchpad.net/bugs/1444630 Interdependencies to other options:
|
11.1.56. wsgi
The following table outlines the options available under the [wsgi]
group in the /etc/nova/nova.conf
file.
Configuration option = Default value | Type | Description |
---|---|---|
| string value | This option represents a file name for the paste.deploy config for nova-api. Possible values:
|
| integer value | This option specifies the timeout for client connections' socket operations. If an incoming connection is idle for this number of seconds it will be closed. It indicates timeout on individual read/writes on the socket connection. To wait forever set to 0. |
| integer value | This option specifies the size of the pool of greenthreads used by wsgi. It is possible to limit the number of concurrent connections using this option. |
| boolean value | This option allows using the same TCP connection to send and receive multiple HTTP requests/responses, as opposed to opening a new one for every single request/response pair. HTTP keep-alive indicates HTTP connection reuse. Possible values:
Related options:
|
| integer value | This option specifies the maximum line size of message headers to be accepted. max_header_line may need to be increased when using large tokens (typically those generated by the Keystone v3 API with big service catalogs). Since TCP is a stream based protocol, in order to reuse a connection, the HTTP has to have a way to indicate the end of the previous response and beginning of the next. Hence, in a keep_alive case, all messages must have a self-defined message length. |
| string value | This option specifies the HTTP header used to determine the protocol scheme for the original request, even if it was removed by a SSL terminating proxy. Possible values:
Warning Do not set this unless you know what you are doing. Make sure ALL of the following are true before setting this (assuming the values from the example above): * Your API is behind a proxy. * Your proxy strips the X-Forwarded-Proto header from all incoming requests. In other words, if end users include that header in their requests, the proxy will discard it. * Your proxy sets the X-Forwarded-Proto header and sends it to API, but only for requests that originally come in via HTTPS. If any of those are not true, you should keep this setting set to None. |
| string value | This option allows setting path to the CA certificate file that should be used to verify connecting clients. Possible values:
Related options:
|
| string value | This option allows setting path to the SSL certificate of API server. Possible values:
Related options:
|
| string value | This option specifies the path to the file where SSL private key of API server is stored when SSL is in effect. Possible values:
Related options:
|
| integer value | This option sets the value of TCP_KEEPIDLE in seconds for each server socket. It specifies the duration of time to keep connection active. TCP generates a KEEPALIVE transmission for an application that requests to keep connection active. Not supported on OS X. Related options:
|
| string value | It represents a python format string that is used as the template to generate log lines. The following values can be formatted into it: client_ip, date_time, request_line, status_code, body_length, wall_seconds. This option is used for building custom request loglines when running nova-api under eventlet. If used under uwsgi or apache, this option has no effect. Possible values:
|
11.1.57. xenserver
The following table outlines the options available under the [xenserver]
group in the /etc/nova/nova.conf
file.
Configuration option = Default value | Type | Description |
---|---|---|
| string value | Path to locate guest agent on the server. Specifies the path in which the XenAPI guest agent should be located. If the agent is present, network configuration is not injected into the image. Related options:
For this option to have an effect: * |
| integer value | Number of seconds to wait for agent’s reply to resetnetwork request.
This indicates the amount of time xapi agent plugin waits for the agent to respond to the resetnetwork request specifically. The generic timeout for agent communication |
| integer value | Number of seconds to wait for agent’s reply to a request. Nova configures/performs certain administrative actions on a server with the help of an agent that’s installed on the server. The communication between Nova and the agent is achieved via sharing messages, called records, over xenstore, a shared storage across all the domains on a Xenserver host. Operations performed by the agent on behalf of nova are: version,' key_init', password,resetnetwork,inject_file, and agentupdate. To perform one of the above operations, the xapi agent plugin writes the command and its associated parameters to a certain location known to the domain and awaits response. On being notified of the message, the agent performs appropriate actions on the server and writes the result back to xenstore. This result is then read by the xapi agent plugin to determine the success/failure of the operation. This config option determines how long the xapi agent plugin shall wait to read the response off of xenstore for a given request/command. If the agent on the instance fails to write the result in this time period, the operation is considered to have timed out. Related options:
|
| integer value | Number of seconds to wait for agent’t reply to version request.
This indicates the amount of time xapi agent plugin waits for the agent to respond to the version request specifically. The generic timeout for agent communication During the build process the version request is used to determine if the agent is available/operational to perform other requests such as resetnetwork, password, key_init and inject_file. If the version call fails, the other configuration is skipped. So, this configuration option can also be interpreted as time in which agent is expected to be fully operational. |
| integer value | Time in secs to wait for a block device to be created |
| string value | Cache glance images locally. The value for this option must be chosen from the choices listed here. Configuring a value other than these will default to all. Note: There is nothing that deletes these images. Possible values:
|
| boolean value | Ensure compute service is running on host XenAPI connects to. This option must be set to false if the independent_compute option is set to true. Possible values:
Related options:
|
| integer value | Maximum number of concurrent XenAPI connections. In nova, multiple XenAPI requests can happen at a time. Configuring this option will parallelize access to the XenAPI session, which allows you to make concurrent XenAPI connections. |
| string value | Password for connection to XenServer/Xen Cloud Platform |
| string value | URL for connection to XenServer/Xen Cloud Platform. A special value of unix://local can be used to connect to the local unix socket. Possible values:
|
| string value | Username for connection to XenServer/Xen Cloud Platform |
| string value | Publicly visible name for this console host. Possible values:
|
| string value | Default OS type used when uploading an image to glance |
| boolean value | Disables the use of XenAPI agent.
This configuration option suggests whether the use of agent should be enabled or not regardless of what image properties are present. Image properties have an effect only when this is set to Related options:
|
| integer value | Compression level for images. By setting this option we can configure the gzip compression level. This option sets GZIP environment variable before spawning tar -cz to force the compression level. It defaults to none, which means the GZIP environment variable is not set and the default (usually -6) is used. Possible values:
|
| string value | Dom0 plugin driver used to handle image uploads. Provide a string value representing a plugin driver required to handle the image uploading to GlanceStore. Images, and snapshots from XenServer need to be uploaded to the data store for use. image_upload_handler takes in a value for the Dom0 plugin driver. This driver is then called to uplaod images to the GlanceStore. |
| boolean value | Used to prevent attempts to attach VBDs locally, so Nova can be run in a VM on a different host. Related options:
|
| integer value | Number of seconds to wait for SR to settle if the VDI does not exist when first introduced. Some SRs, particularly iSCSI connections are slow to see the VDIs right after they got introduced. Setting this option to a time interval will make the SR to wait for that time period before raising VDI not found exception. |
| string value | URL to the iPXE boot menu. An iPXE ISO is a specially crafted ISO which supports iPXE booting. This feature gives a means to roll your own image. By default this option is not set. Enable this option to boot an iPXE ISO. Related Options:
|
| string value | Name and optionally path of the tool used for ISO image creation. An iPXE ISO is a specially crafted ISO which supports iPXE booting. This feature gives a means to roll your own image.
Note: By default Related Options:
|
| string value | Name of network to use for booting iPXE ISOs. An iPXE ISO is a specially crafted ISO which supports iPXE booting. This feature gives a means to roll your own image. By default this option is not set. Enable this option to boot an iPXE ISO. Related Options:
|
| integer value | Timeout in seconds for XenAPI login. |
| integer value | Maximum size in bytes of kernel or ramdisk images. Specifying the maximum size of kernel or ramdisk will avoid copying large files to dom0 and fill up /boot/guest. |
| integer value | Maximum number of retries to unplug VBD. If set to 0, should try once, no retries. |
| string value | The name of the integration Bridge that is used with xenapi when connecting with Open vSwitch. Note: The value of this config option is dependent on the environment, therefore this configuration value must be set accordingly if you are using XenAPI. Possible values:
|
| integer value | Wait time for instances to go to running state. Provide an integer value representing time in seconds to set the wait time for an instance to go to running state. When a request to create an instance is received by nova-api and communicated to nova-compute, the creation of the instance occurs through interaction with Xen via XenAPI in the compute node. Once the node on which the instance(s) are to be launched is decided by nova-schedule and the launch is triggered, a certain amount of wait time is involved until the instance(s) can become available and running. This wait time is defined by running_timeout. If the instances do not go to running state within this specified wait time, the launch expires and the instance(s) are set to error state. |
| boolean value | Whether to use sparse_copy for copying data on a resize down. (False will use standard dd). This speeds up resizes down considerably since large runs of zeros won’t have to be rsynced. |
| string value | Base path to the storage repository on the XenServer host. |
| string value | Filter for finding the SR to be used to install guest instances on. Possible values:
|
| unknown value | The iSCSI Target Host. This option represents the hostname or ip of the iSCSI Target. If the target host is not present in the connection information from the volume provider then the value from this option is taken. Possible values:
|
| port value | The iSCSI Target Port. This option represents the port of the iSCSI Target. If the target port is not present in the connection information from the volume provider then the value from this option is taken. |
| boolean value | Whether or not to use the agent by default when its usage is enabled but not indicated by the image.
The use of XenAPI agent can be disabled altogether using the configuration option
Note that if this configuration is set to Related options:
|
| boolean value | When adding new host to a pool, this will append a --force flag to the command, forcing hosts to join a pool, even if they have different CPUs. Since XenServer version 5.6 it is possible to create a pool of hosts that have different CPU capabilities. To accommodate CPU differences, XenServer limited features it uses to determine CPU compatibility to only the ones that are exposed by CPU and support for CPU masking was added. Despite this effort to level differences between CPUs, it is still possible that adding new host will fail, thus option to force join was introduced. |
| integer value | Max number of times to poll for VHD to coalesce. This option determines the maximum number of attempts that can be made for coalescing the VHD before giving up. Related opitons:
|
| floating point value | The interval used for polling of coalescing vhds. This is the interval after which the task of coalesce VHD is performed, until it reaches the max attempts that is set by vhd_coalesce_max_attempts. Related options:
|
| string value | The XenAPI VIF driver using XenServer Network APIs. Provide a string value representing the VIF XenAPI vif driver to use for plugging virtual network interfaces. Xen configuration uses bridging within the backend domain to allow all VMs to appear on the network as individual hosts. Bridge interfaces are used to create a XenServer VLAN network in which the VIFs for the VM instances are plugged. If no VIF bridge driver is plugged, the bridge is not made available. This configuration option takes in a value for the VIF driver. Possible values:
Related options:
|
11.1.58. xvp
The following table outlines the options available under the [xvp]
group in the /etc/nova/nova.conf
file.
Configuration option = Default value | Type | Description |
---|---|---|
| string value | Generated XVP conf file |
| string value | XVP conf template |
| string value | XVP log file |
| port value | Port for XVP to multiplex VNC connections on |
| string value | XVP master process pid file |