[Openstack] Volume backed live migration error
Jakub Pavlík
J.Pavlik at tcpisek.cz
Mon Jun 30 11:07:42 UTC 2014
Hi guys,
I have problem with live migration and block migration with IBM SVC plugin for Cinder at Fibre Channel. I have instance (boot from volume) running at compute node ch1nod2 and I am trying to migrate on second compute ch1nod3. I run the command and it takes about 1 second, after that instance is in state MIGRATION and at ch1nod3 I can see a new paused instance. It takes about 6 minutes and after I can see the errors bellow.
[root at ctl-1*12* ~]#nova list
+--------------------------------------+----------------+--------+------------+-------------+------------------------+
| ID | Name | Status | Task State | Power State | Networks |
+--------------------------------------+----------------+--------+------------+-------------+------------------------+
| 8a9979f0-8a27-40ee-bc9d-b8cf17dd7265 | WindowsTest | ACTIVE | - | Running | network1=192.168.6.54 |
| 9c5afd75-ab19-44c6-a630-458431ad4eda | centossnap2 | ACTIVE | - | Running | network2=192.168.7.252 |
| df725a9f-784a-4207-b7d0-da2a9f34eb9c | next | ACTIVE | - | NOSTATE | network1=192.168.6.253 |
| 2cb09534-5413-4975-8560-b05ff9645c35 | volumesnaphost | ACTIVE | - | Running | network2=192.168.7.253 |
+--------------------------------------+----------------+--------+------------+-------------+------------------------+ │
[root at ctl-1*12* ~]#nova live-migration 9c5afd75-ab19-44c6-a630-458431ad4ed ch1nod3.12.intra.cloudlab.cz
ch1nod2 - compute node
2014-06-30 11:02:17.376 3079 INFO nova.compute.manager [-] [instance: 9c5afd75-ab19-44c6-a630-458431ad4eda] During sync_power_state the instance has a pending task. Skip.
2014-06-30 11:08:11.606 3079 INFO nova.compute.resource_tracker [-] Compute_service record updated for ch1nod2.12.intra.cloudlab.cz:ch1nod2.12.intra.cloudlab.cz
2014-06-30 11:09:07.582 3079 INFO nova.compute.manager [-] Lifecycle event 3 on VM 9c5afd75-ab19-44c6-a630-458431ad4eda
2014-06-30 11:09:07.585 3079 ERROR nova.virt.libvirt.driver [-] [instance: 9c5afd75-ab19-44c6-a630-458431ad4eda] Live Migration failure: operation failed: migration job: unexpectedly failed
2014-06-30 11:09:11.670 3079 AUDIT nova.compute.resource_tracker [-] Auditing locally available compute resources
2014-06-30 11:09:11.873 3079 AUDIT nova.compute.resource_tracker [-] Free ram (MB): 189078
2014-06-30 11:09:11.874 3079 AUDIT nova.compute.resource_tracker [-] Free disk (GB): 156
2014-06-30 11:09:11.874 3079 AUDIT nova.compute.resource_tracker [-] Free VCPUS: 22
ch1nod3 - compute node
2014-06-30 11:09:07.578 29488 INFO nova.compute.manager [-] Lifecycle event 1 on VM 9c5afd75-ab19-44c6-a630-458431ad4eda
2014-06-30 11:09:07.738 29488 INFO nova.compute.manager [-] [instance: 9c5afd75-ab19-44c6-a630-458431ad4eda] During the sync_power process the instance has moved from host ch1nod3.12.intra.cloudlab.cz to host ch1nod2.12.intra.cloudlab.cz
2014-06-30 11:09:07.939 29488 AUDIT nova.compute.manager [req-6748d84f-4f0c-43ef-97b4-12e74a989b57 6836cb1afded478a802e2f28020b2bad e47d5141f5ac40f8a5fedf76bb40e904] [instance: 9c5afd75-ab19-44c6-a630-458431ad4eda] Detach volume 5c0160d7-f9e2-4089-9b0b-d3f3ad46006c from mountpoint vda
2014-06-30 11:09:07.941 29488 WARNING nova.compute.manager [req-6748d84f-4f0c-43ef-97b4-12e74a989b57 6836cb1afded478a802e2f28020b2bad e47d5141f5ac40f8a5fedf76bb40e904] [instance: 9c5afd75-ab19-44c6-a630-458431ad4eda] Detaching volume from unknown instance
2014-06-30 11:09:07.944 29488 ERROR nova.compute.manager [req-6748d84f-4f0c-43ef-97b4-12e74a989b57 6836cb1afded478a802e2f28020b2bad e47d5141f5ac40f8a5fedf76bb40e904] [instance: 9c5afd75-ab19-44c6-a630-458431ad4eda] Failed to detach volume 5c0160d7-f9e2-4089-9b0b-d3f3ad46006c from vda
2014-06-30 11:09:07.944 29488 TRACE nova.compute.manager [instance: 9c5afd75-ab19-44c6-a630-458431ad4eda] Traceback (most recent call last):
2014-06-30 11:09:07.944 29488 TRACE nova.compute.manager [instance: 9c5afd75-ab19-44c6-a630-458431ad4eda] File "/usr/lib/python2.6/site-packages/nova/compute/manager.py", line 3725, in _detach_volume
2014-06-30 11:09:07.944 29488 TRACE nova.compute.manager [instance: 9c5afd75-ab19-44c6-a630-458431ad4eda] encryption=encryption)
2014-06-30 11:09:07.944 29488 TRACE nova.compute.manager [instance: 9c5afd75-ab19-44c6-a630-458431ad4eda] File "/usr/lib/python2.6/site-packages/nova/virt/libvirt/driver.py", line 1202, in detach_volume
2014-06-30 11:09:07.944 29488 TRACE nova.compute.manager [instance: 9c5afd75-ab19-44c6-a630-458431ad4eda] virt_dom = self._lookup_by_name(instance_name)
2014-06-30 11:09:07.944 29488 TRACE nova.compute.manager [instance: 9c5afd75-ab19-44c6-a630-458431ad4eda] File "/usr/lib/python2.6/site-packages/nova/virt/libvirt/driver.py", line 3085, in _lookup_by_name
2014-06-30 11:09:07.944 29488 TRACE nova.compute.manager [instance: 9c5afd75-ab19-44c6-a630-458431ad4eda] raise exception.InstanceNotFound(instance_id=instance_name)
2014-06-30 11:09:07.944 29488 TRACE nova.compute.manager [instance: 9c5afd75-ab19-44c6-a630-458431ad4eda] InstanceNotFound: Instance instance-000009e0 could not be found.
2014-06-30 11:09:07.944 29488 TRACE nova.compute.manager [instance: 9c5afd75-ab19-44c6-a630-458431ad4eda]
[root at ch1nod2 ~]# grep "tls\|tcp" /etc/libvirt/libvirtd.conf | grep -v "^#"
listen_tls = 0
listen_tcp = 1
auth_tcp = "none"
nova.conf
# Migration flags to be set for live migration (string value)
#live_migration_flag=VIR_MIGRATE_UNDEFINE_SOURCE, VIR_MIGRATE_PEER2PEER
live_migration_flag=VIR_MIGRATE_UNDEFINE_SOURCE,VIR_MIGRATE_PEER2PEER,VIR_MIGRATE_LIVE
# Migration flags to be set for block migration (string value)
block_migration_flag=VIR_MIGRATE_UNDEFINE_SOURCE, VIR_MIGRATE_PEER2PEER, VIR_MIGRATE_NON_SHARED_INC
Can anybody help me with this problem?
Jakub
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.openstack.org/pipermail/openstack/attachments/20140630/9769878c/attachment.html>
More information about the Openstack
mailing list