[openstack-announce] [release][nova] nova 13.0.0 release (mitaka)
no-reply at openstack.org
no-reply at openstack.org
Thu Apr 7 07:50:42 UTC 2016
We are pleased to announce the release of:
nova 13.0.0: Cloud computing fabric controller
This release is part of the mitaka release series.
For more details, please see below.
13.0.0
^^^^^^
Nova 13.0.0 release is including a lot of new features and bugfixes.
It can be extremely hard to mention all the changes we introduced
during that release but we beg you to read at least the upgrade
section which describes the required modifications that you need to do
for upgrading your cloud from 12.0.0 (Liberty) to 13.0.0 (Mitaka).
That said, a few major changes are worth to notice here. This is not
an exhaustive list of things to notice, rather just important things
you need to know :
* Latest API microversion supported for Mitaka is v2.25
* Nova now requires a second database (called 'API DB').
* A new nova-manage script allows you to perform all online DB
migrations once you upgrade your cloud
* EC2 API support is fully removed.
New Features
************
* Enables NUMA topology reporting on PowerPC architecture from the
libvirt driver in Nova but with a caveat as mentioned below. NUMA
cell affinity and dedicated cpu pinning code assumes that the host
operating system is exposed to threads. PowerPC based hosts use core
based scheduling for processes. Due to this, the cores on the
PowerPC architecture are treated as threads. Since cores are always
less than or equal to the threads on a system, this leads to non-
optimal resource usage while pinning. This feature is supported from
libvirt version 1.2.19 for PowerPC.
* A new REST API to cancel an ongoing live migration has been added
in microversion 2.24. Initially this operation will only work with
the libvirt virt driver.
* It is possible to call attach and detach volume API operations for
instances which are in shelved and shelved_offloaded state. For an
instance in shelved_offloaded state Nova will set to None the value
for the device_name field, the right value for that field will be
set once the instance will be unshelved as it will be managed by a
specific compute manager.
* It is possible to block live migrate instances with additional
cinder volumes attached. This requires libvirt version to be
>=1.2.17 and does not work when live_migration_tunnelled is set to
True.
* Project-id and user-id are now also returned in the return data of
os-server-groups APIs. In order to use this new feature, user have
to contain the header of request microversion v2.13 in the API
request.
* Add support for enabling uefi boot with libvirt.
* A new host_status attribute for servers/detail and
servers/{server_id}. In order to use this new feature, user have to
contain the header of request microversion v2.16 in the API request.
A new policy "os_compute_api:servers:show:host_status" added to
enable the feature. By default, this is only exposed to cloud
administrators.
* A new server action trigger_crash_dump has been added to the REST
API in microversion 2.17.
* When RBD is used for ephemeral disks and image storage, make
snapshot use Ceph directly, and update Glance with the new location.
In case of failure, it will gracefully fallback to the "generic"
snapshot method. This requires changing the typical permissions for
the Nova Ceph user (if using authx) to allow writing to the pool
where vm images are stored, and it also requires configuring Glance
to provide a v2 endpoint with direct_url support enabled (there are
security implications to doing this). See
http://docs.ceph.com/docs/master/rbd/rbd-openstack/ for more
information on configuring OpenStack with RBD.
* A new option "live_migration_inbound_addr" has been added in the
configuration file, set None as default value. If this option is
present in pre_migration_data, the ip address/hostname provided will
be used instead of the migration target compute node's hostname as
the uri for live migration, if it's None, then the mechanism remains
as it is before.
* Added support for CPU thread policies, which can be used to
control how the libvirt virt driver places guests with respect to
CPU SMT "threads". These are provided as instance and image metadata
options, 'hw:cpu_thread_policy' and 'hw_cpu_thread_policy'
respectively, and provide an additional level of control over CPU
pinning policy, when compared to the existing CPU policy feature.
These changes were introduced in commits '83cd67c' and 'aaaba4a'.
* Add support for enabling discard support for block devices with
libvirt. This will be enabled for Cinder volume attachments that
specify support for the feature in their connection properties. This
requires support to be present in the version of libvirt (v1.0.6+)
and qemu (v1.6.0+) used along with the configured virtual drivers
for the instance. The virtio-blk driver does not support this
functionality.
* A new "auto" value for the configuration option
"upgrade_levels.compute" is accepted, that allows automatic
determination of the compute service version to use for RPC
communication. By default, we still use the newest version if not
set in the config, a specific version if asked, and only do this
automatic behavior if 'auto' is configured. When 'auto' is used,
sending a SIGHUP to the service will cause the value to be re-
calculated. Thus, after an upgrade is complete, sending SIGHUP to
all services will cause them to start sending messages compliant
with the newer RPC version.
* Libvirt driver in Nova now supports Cinder DISCO volume driver.
* A disk space scheduling filter is now available, which prefers
compute nodes with the most available disk space. By default, free
disk space is given equal importance to available RAM. To increase
the priority of free disk space in scheduling, increase the
disk_weight_multiplier option.
* A new REST API to force live migration to complete has been added
in microversion 2.22.
* The os-instance-actions methods now read actions from deleted
instances. This means that 'GET /v2.1/{tenant-id}/servers/{server-id
}/os-instance-actions' and 'GET /v2.1/{tenant-id}/servers/{server-id
}/os-instance-actions/{req-id}' will return instance-action items
even if the instance corresponding to '{server-id}' has been
deleted.
* When booting an instance, its sanitized 'hostname' attribute is
now used to populate the 'dns_name' attribute of the Neutron ports
the instance is attached to. This functionality enables the Neutron
internal DNS service to know the ports by the instance's hostname.
As a consequence, commands like 'hostname -f' will work as expected
when executed in the instance. When a port's network has a non-blank
'dns_domain' attribute, the port's 'dns_name' combined with the
network's 'dns_domain' will be published by Neutron in an external
DNS as a service like Designate. As a consequence, the instance's
hostname is published in the external DNS as a service. This
functionality is added to Nova when the 'DNS Integration' extension
is enabled in Neutron. The publication of 'dns_name' and
'dns_domain' combinations to an external DNS as a service
additionaly requires the configuration of the appropriate driver in
Neutron. When the 'Port Binding' extension is also enabled in
Neutron, the publication of a 'dns_name' and 'dns_domain'
combination to the external DNS as a service will require one
additional update operation when Nova allocates the port during the
instance boot. This may have a noticeable impact on the performance
of the boot process.
* The libvirt driver now has a live_migration_tunnelled
configuration option which should be used where the
VIR_MIGRATE_TUNNELLED flag would previously have been set or unset
in the live_migration_flag and block_migration_flag configuration
options.
* For the libvirt driver, by default hardware properties will be
retrieved from the Glance image and if such haven't been provided,
it will use a libosinfo database to get those values. If users want
to force a specific guest OS ID for the image, they can now use a
new glance image property "os_distro" (eg. "--property
os_distro=fedora21"). In order to use the libosinfo database, you
need to separately install the related native package provided for
your operating system distribution.
* Add support for allowing Neutron to specify the bridge name for
the OVS, Linux Bridge, and vhost-user VIF types.
* Added a *nova-manage db online_data_migrations* command for
forcing online data migrations, which will run all registered
migrations for the release, instead of there being a separate
command for each logical data migration. Operators need to make sure
all data is migrated before upgrading to the next release, and the
new command provides a unified interface for doing it.
* Provides API 2.18, which makes the use of project_ids in API urls
optional.
* Libvirt with Virtuozzo virtualisation type now supports snapshot
operations
* Remove "onSharedStorage" parameter from server's evacuate action
in microversion 2.14. Nova will automatically detect if the instance
is on shared storage. Also adminPass is removed from the response
body which makes the response body empty. The user can get the
password with the server's os-server-password action.
* Add two new list/show API for server-migration. The list API will
return the in progress live migratons information of a server. The
show API will return a specified in progress live migration of a
server. This has been added in microversion 2.23.
* A new service.status versioned notification has been introduced.
When the status of the Service object is changed nova will send a
new service.update notification with versioned payload according to
bp versioned-notification-api. The new notification is documented in
http://docs.openstack.org/developer/nova/notifications.html
* Two new policies soft-affinty and soft-anti-affinity have been
implemented for the server-group feature of Nova. This means that
POST /v2.1/{tenant_id}/os-server-groups API resource now accepts
'soft-affinity' and 'soft-anti-affinity' as value of the 'policies'
key of the request body.
* In Nova Compute API microversion 2.19, you can specify a
"description" attribute when creating, rebuilding, or updating a
server instance. This description can be retrieved by getting
server details, or list details for servers. Refer to the Nova
Compute API documentation for more information. Note that the
description attribute existed in prior Nova versions, but was set to
the server name by Nova, and was not visible to the user. So,
servers you created with microversions prior to 2.19 will return the
description equals the name on server details microversion 2.19.
* As part of refactoring the notification interface of Nova a new
config option 'notification_format' has been added to specifies
which notification format shall be used by nova. The possible values
are 'unversioned' (e.g. legacy), 'versioned', 'both'. The default
value is 'both'. The new versioned notifications are documented in
http://docs.openstack.org/developer/nova/notifications.html
* For the VMware driver, the flavor extra specs for quotas has been
extended to support:
* quota:cpu_limit - The cpu of a virtual machine will not exceed
this limit, even if there are available resources. This is
typically used to ensure a consistent performance of virtual
machines independent of available resources. Units are MHz.
* quota:cpu_reservation - guaranteed minimum reservation (MHz)
* quota:cpu_shares_level - the allocation level. This can be
'custom', 'high', 'normal' or 'low'.
* quota:cpu_shares_share - in the event that 'custom' is used,
this is the number of shares.
* quota:memory_limit - The memory utilization of a virtual machine
will not exceed this limit, even if there are available resources.
This is typically used to ensure a consistent performance of
virtual machines independent of available resources. Units are MB.
* quota:memory_reservation - guaranteed minimum reservation (MB)
* quota:memory_shares_level - the allocation level. This can be
'custom', 'high', 'normal' or 'low'.
* quota:memory_shares_share - in the event that 'custom' is used,
this is the number of shares.
* quota:disk_io_limit - The I/O utilization of a virtual machine
will not exceed this limit. The unit is number of I/O per second.
* quota:disk_io_reservation - Reservation control is used to
provide guaranteed allocation in terms of IOPS
* quota:disk_io_shares_level - the allocation level. This can be
'custom', 'high', 'normal' or 'low'.
* quota:disk_io_shares_share - in the event that 'custom' is used,
this is the number of shares.
* quota:vif_limit - The bandwidth limit for the virtual network
adapter. The utilization of the virtual network adapter will not
exceed this limit, even if there are available resources. Units in
Mbits/sec.
* quota:vif_reservation - Amount of network bandwidth that is
guaranteed to the virtual network adapter. If utilization is less
than reservation, the resource can be used by other virtual
network adapters. Reservation is not allowed to exceed the value
of limit if limit is set. Units in Mbits/sec.
* quota:vif_shares_level - the allocation level. This can be
'custom', 'high', 'normal' or 'low'.
* quota:vif_shares_share - in the event that 'custom' is used,
this is the number of shares.
Upgrade Notes
*************
* All noVNC proxy configuration options have been added to the 'vnc'
group. They should no longer be included in the 'DEFAULT' group.
* All VNC XVP configuration options have been added to the 'vnc'
group. They should no longer be included in the 'DEFAULT' group.
* Upon first startup of the scheduler service in Mitaka, all defined
aggregates will have UUIDs generated and saved back to the database.
If you have a significant number of aggregates, this may delay
scheduler start as that work is completed, but it should be minor
for most deployments.
* During an upgrade to Mitaka, operators must create and initialize
a database for the API service. Configure this in
[api_database]/connection, and then run "nova-manage api_db sync"
* We can not use microversion 2.25 to do live-migration during
upgrade, nova-api will raise bad request if there is still old
compute node in the cluster.
* The option "scheduler_driver" is now changed to use entrypoint
instead of full class path. Set one of the entrypoints under the
namespace 'nova.scheduler.driver' in 'setup.cfg'. Its default value
is 'host_manager'. The full class path style is still supported in
current release. But it is not recommended because class path can be
changed and this support will be dropped in the next major release.
* The option "scheduler_host_manager" is now changed to use
entrypoint instead of full class path. Set one of the entrypoints
under the namespace 'nova.scheduler.host_manager' in 'setup.cfg'.
Its default value is 'host_manager'. The full class path style is
still supported in current release. But it is not recommended
because class path can be changed and this support will be dropped
in the next major release.
* The local conductor mode is now deprecated and may be removed as
early as the 14.0.0 release. If you are using local conductor mode,
plan on deploying remote conductor by the time you upgrade to the
14.0.0 release.
* The Extensible Resource Tracker is deprecated and will be removed
in the 14.0.0 release. If you use this functionality and have custom
resources that are managed by the Extensible Resource Tracker,
please contact the Nova development team by posting to the
openstack-dev mailing list. There is no future planned support for
the tracking of custom resources.
* For Liberty compute nodes, the disk_allocation_ratio works as
before, you must set it on the scheduler if you want to change it.
For Mitaka compute nodes, the disk_allocation_ratio set on the
compute nodes will be used only if the configuration is not set on
the scheduler. This is to allow, for backwards compatibility, the
ability to still override the disk allocation ratio by setting the
configuration on the scheduler node. In Newton, we plan to remove
the ability to set the disk allocation ratio on the scheduler, at
which point the compute nodes will always define the disk allocation
ratio, and pass that up to the scheduler. None of this changes the
default disk allocation ratio of 1.0. This matches the behaviour of
the RAM and CPU allocation ratios.
* (Only if you do continuous deployment)
1337890ace918fa2555046c01c8624be014ce2d8 drops support for an
instance major version, which means that you must have deployed at
least commit 713d8cb0777afb9fe4f665b9a40cac894b04aacb before
deploying this one.
* nova now requires ebtables 2.0.10 or later
* nova recommends libvirt 1.2.11 or later
* Filters internal interface changed using now the RequestSpec
NovaObject instead of an old filter_properties dictionary. In case
you run out-of-tree filters, you need to modify the host_passes()
method to accept a new RequestSpec object and modify the filter
internals to use that new object. You can see other in-tree filters
for getting the logic or ask for help in #openstack-nova IRC
channel.
* The "force_config_drive" configuration option provided an "always"
value which was deprecated in the previous release. That "always"
value is now no longer accepted and deployments using that value
have to change it to "True" before upgrading.
* Support for Windows / Hyper-V Server 2008 R2 has been deprecated
in Liberty (12.0.0) and it is no longer supported in Mitaka
(13.0.0). If you have compute nodes running that version, please
consider moving the running instances to other compute nodes before
upgrading those to Mitaka.
* The libvirt driver will now correct unsafe and invalid values for
the live_migration_flag and block_migration_flag configuration
options. The live_migration_flag must not contain
VIR_MIGRATE_SHARED_INC but block_migration_flag must contain it.
Both options must contain the VIR_MIGRATE_PEER2PEER, except when
using the 'xen' virt type this flag is not supported. Both flags
must contain the VIR_MIGRATE_UNDEFINE_SOURCE flag and not contain
the VIR_MIGRATE_PERSIST_DEST flag.
* The libvirt driver has changed the default value of the
'live_migration_uri' flag, that now is dependent on the 'virt_type'.
The old default 'qemu+tcp://%s/system' now is adjusted for each of
the configured hypervisors. For Xen this will be
'xenmigr://%s/system', for kvm/qemu this will be
'qemu+tcp://%s/system'.
* The minimum required libvirt is now version 0.10.2. The minimum
libvirt for the N release has been set to 1.2.1.
* In order to make project_id optional in urls, we must constrain
the set of allowed values for project_id in our urls. This defaults
to a regex of "[0-9a-f\-]+", which will match hex uuids (with /
without dashes), and integers. This covers all known project_id
formats in the wild. If your site uses other values for project_id,
you can set a site specific validation with "project_id_regex"
config variable.
* The old neutron communication options that were slated for removal
in Mitaka are no longer available. This means that going forward
communication to neutron will need to be configured using auth
plugins.
* All code and tests for Nova's EC2 and ObjectStore API support
which was deprecated in Kilo
(https://wiki.openstack.org/wiki/ReleaseNotes/Kilo#Upgrade_Notes_2)
has been completely removed in Mitaka. This has been replaced by the
new ec2-api project
(http://git.openstack.org/cgit/openstack/ec2-api/).
* The commit with change-id
Idd4bbbe8eea68b9e538fa1567efd304e9115a02a requires that the nova_api
database is setup and Nova is configured to use it. Instructions on
doing that are provided below.
Nova now requires that two databases are available and configured.
The existing nova database needs no changes, but a new nova_api
database needs to be setup. It is configured and managed very
similarly to the nova database. A new connection string
configuration option is available in the api_database group. An
example:
[api_database]
connection = mysql+pymysql://user:secret@127.0.0.1/nova_api?charset=utf8
And a new nova-manage command has been added to manage db migrations
for this database. "nova-manage api_db sync" and "nova-manage
api_db version" are available and function like the parallel "nova-
manage db ..." version.
* A new "use_neutron" option is introduced which replaces the obtuse
"network_api_class" option. This defaults to 'False' to match
existing defaults, however if "network_api_class" is set to the
known Neutron value Neutron networking will still be used as before.
* The FilterScheduler is now including disabled hosts. Make sure you
include the ComputeFilter in the "scheduler_default_filters" config
option to avoid placing instances on disabled hosts.
* Upgrade the rootwrap configuration for the compute service, so
that patches requiring new rootwrap configuration can be tested with
grenade.
* For backward compatible support the setting
"CONF.vmware.integration_bridge" needs to be set when using the
Neutron NSX|MH plugin. The default value has been set to "None".
* XenServer hypervisor type has been changed from "xen" to
"XenServer". It could impact your aggregate metadata or your flavor
extra specs if you provide only the former.
* The glance xenserver plugin has been bumped to version 1.3 which
includes new interfaces for referencing glance servers by url. All
dom0 will need to be upgraded with this plugin before upgrading the
nova code.
Deprecation Notes
*****************
* It is now deprecated to use [glance] api_servers without a
protocol scheme (http / https). This is required to support urls
throughout the system. Update any api_servers list with fully
qualified https / http urls.
* The conductor.manager configuration option is now deprecated and
will be removed.
* Deprecate "compute_stats_class" config option. This allowed
loading an alternate implementation for collecting statistics for
the local compute host. Deployments that felt the need to use this
facility are encoraged to propose additions upstream so we can
create a stable and supported interface here.
* Deprecate the "db_driver" config option. Previously this let you
replace our SQLAlchemy database layer with your own. This approach
is deprecated. Deployments that felt the need to use the facility
are encourage to work with upstream Nova to address db driver
concerns in the main SQLAlchemy code paths.
* The host, port, and protocol options in the [glance] configuration
section are deprecated, and will be removed in the N release. The
api_servers value should be used instead.
* Deprecate the use of nova.hooks. This facility used to let
arbitrary out of tree code be executed around certain internal
actions, but is unsuitable for having a well maintained API. Anyone
using this facility should bring forward their use cases in the
Newton cycle as nova-specs.
* Nova used to support the concept that "service managers" were
replaceable components. There are many config options where you can
replace a manager by specifying a new class. This concept is
deprecated in Mitaka as are the following config options.
* [cells] manager
* metadata_manager
* compute_manager
* console_manager
* consoleauth_manager
* cert_manager
* scheduler_manager
Many of these will be removed in Newton. Users of these options are
encouraged to work with Nova upstream on any features missing in the
default implementations that are needed.
* Deprecate "security_group_api" configuration option. The current
values are "nova" and "neutron". In future the correct
security_group_api option will be chosen based on the value of
"use_neutron" which provides a more coherent user experience.
* Deprecate the "vendordata_driver" config option. This allowed
creating a different class loader for defining vendordata metadata.
The default driver loads from a json file that can be arbitrarily
specified, so is still quite flexible. Deployments that felt the
need to use this facility are encoraged to propose additions
upstream so we can create a stable and supported interface here.
* The configuration option "api_version" in the "ironic" group was
marked as deprecated and will be removed in the future. The only
possible value for that configuration was "1" (because Ironic only
has 1 API version) and the Ironic team came to an agreement that
setting the API version via configuration option should not be
supported anymore. As the Ironic driver in Nova requests the Ironic
v1.8 API, that means that Nova 13.0.0 ("Mitaka") requires Ironic
4.0.0 ("Liberty") or newer if you want to use the Ironic driver.
* The libvirt live_migration_flag and block_migration_flag config
options are deprecated. These options gave too fine grained control
over the flags used and, in some cases, misconfigurations could have
dangerous side effects. Please note the availability of a new
live_migration_tunnelled configuration option.
* The "network_device_mtu" option in Nova is deprecated for removal
since network MTU should be specified when creating the network with
nova-network. With Neutron networks, the MTU value comes from the
"segment_mtu" configuration option in Neutron.
* The old top-level resource */os-migrations* is deprecated, it
won't be extended anymore. And migration_type for /os-migrations,
also add ref link to the /servers/{uuid}/migrations/{id} for it when
the migration is an in-progress live-migration. This has been added
in microversion 2.23.
* Deprecate "volume_api_class" and "network_api_class" config
options. We only have one sensible backend for either of these.
These options will be removed and turned into constants in Newton.
* Option "memcached_servers" is deprecated in Mitaka. Operators
should use oslo.cache configuration instead. Specifically "enabled"
option under [cache] section should be set to True and the url(s)
for the memcached servers should be in [cache]/memcache_servers
option.
* The Zookeeper Service Group driver has been removed.
The driver has no known users and is not actively mantained. A
warning log message about the driver's state was added for the Kilo
release. Also, evzookeeper library that the driver depends on is
unmaintained and incompatible with recent eventlet releases.
A future release of Nova will use the Tooz library to track service
liveliness, and Tooz supports Zookeeper.
(https://bugs.launchpad.net/nova/+bug/1443910)
(http://specs.openstack.org/openstack/nova-
specs/specs/liberty/approved/service-group-using-tooz.html)
Security Issues
***************
* [OSSA 2016-001] Nova host data leak through snapshot
(CVE-2015-7548)
* Bug 1524274 (https://bugs.launchpad.net/nova/+bug/1524274)
* Announcement
[OSSA 2016-002] Xen connection password leak in logs via
StorageError (CVE-2015-8749)
* Bug 1516765 (https://bugs.launchpad.net/nova/+bug/1516765)
* Announcement
[OSSA 2016-007] Host data leak during resize/migrate for raw-backed
instances (CVE-2016-2140)
* Bug 1548450 (https://bugs.launchpad.net/nova/+bug/1548450)
* Announcement
Bug Fixes
*********
* In a race condition if base image is deleted by ImageCacheManager
while imagebackend is copying the image to instance path, then the
instance goes in to error state. In this case when libvirt has
changed the base file ownership to libvirt-qemu while imagebackend
is copying the image, then we get permission denied error on
updating the file access time using os.utime. Fixed this issue by
updating the base file access time with root user privileges using
'touch' command.
* When plugging virtual interfaces of type vhost-user the MTU value
will not be applied to the interface by nova. vhost-user ports exist
only in userspace and are not backed by kernel netdevs, for this
reason it is not possible to set the mtu on a vhost-user interface
using standard tools such as ifconfig or ip link.
Other Notes
***********
* Conductor RPC API no longer supports v2.x.
* The service subcommand of nova-manage is deprecated. Use the nova
service-* commands from python-novaclient instead or the os-services
REST resource. The service subcommand will be removed in the 14.0
release.
* The Neutron network MTU value is now used when plugging virtual
interfaces in nova-compute. If the value is 0, which is the default
value for the "segment_mtu" configuration option in Neutron before
Mitaka, then the (deprecated) "network_device_mtu" configuration
option in Nova is used, which defaults to not setting an MTU value.
* The sample policy file shipped with Nova contained many policies
set to ""(allow all) which was not the proper default for many of
those checks. It was also a source of confusion as some people
thought "" meant to use the default rule. These empty policies have
been updated to be explicit in all cases. Many of them were changed
to match the default rule of "admin_or_owner" which is a more
restrictive policy check but does not change the restrictiveness of
the API calls overall because there are similar checks in the
database already. This does not affect any existing deployment, just
the sample file included for use by new deployments.
* Nova's EC2 API support which was deprecated in Kilo
(https://wiki.openstack.org/wiki/ReleaseNotes/Kilo#Upgrade_Notes_2)
is removed from Mitaka. This has been replaced by the new ec2-api
project (http://git.openstack.org/cgit/openstack/ec2-api/).
Changes in nova 13.0.0.0rc2..13.0.0
-----------------------------------
7105f88 Imported Translations from Zanata
5de98cb Imported Translations from Zanata
a9d5542 Fix detach SR-IOV when using LibvirtConfigGuestHostdevPCI
5b6ee70 Imported Translations from Zanata
29042e0 Imported Translations from Zanata
3e9819d Update cells blacklist regex for test_server_basic_ops
c71c4e0 Stop providing force_hosts to the scheduler for move ops
Diffstat (except docs and test files)
-------------------------------------
devstack/tempest-dsvm-cells-rc | 2 +-
nova/conductor/manager.py | 8 +
nova/conductor/tasks/live_migrate.py | 4 +
nova/locale/de/LC_MESSAGES/nova.po | 22 +-
nova/locale/fr/LC_MESSAGES/nova.po | 34 +-
nova/locale/ja/LC_MESSAGES/nova.po | 1315 ++++++++++++--
nova/locale/ko_KR/LC_MESSAGES/nova-log-warning.po | 1914 ++++++++++++++++++++
nova/objects/request_spec.py | 12 +
.../unit/conductor/tasks/test_live_migrate.py | 3 +
nova/virt/libvirt/driver.py | 13 +-
13 files changed, 3194 insertions(+), 194 deletions(-)
More information about the OpenStack-announce
mailing list