[Openstack] [OpenStack][Nova] Live Migration Error

Razique Mahroua razique.mahroua at gmail.com
Tue Jul 10 10:05:39 UTC 2012


Ok it looks like Qemu is unable to access the instance state.
could you perform a $ virsh list --all from the second node and tell me what you see ?
as for the second message, make sure you installed the "dbus" package

Regards,
Razique

Nuage & Co - Razique Mahroua 
razique.mahroua at gmail.com



Le 10 juil. 2012 à 11:29, Leander Bessa Beernaert a écrit :

> Well i've checked the libvirt logs on both nodes and i found these two lines:
> 
> 2012-07-09 13:58:27.179+0000: 10227: warning : qemuDomainObjTaint:1134 : Domain id=2 name='instance-00000002' uuid=57aca8a6-d062-4a08-8d87-e4d11d259ac7 is tainted: high-privileges
> 2012-07-09 13:58:27.736+0000: 10226: error : qemuMonitorIORead:513 : Unable to read from monitor: Connection reset by peer
> 
> The log is alos filled with the message below, it gets repeated over and over and over.
> 
> 2012-07-10 09:26:02.244+0000: 10229: error : virExecWithHook:328 : Cannot find 'pm-is-supported' in path: No such file or directory
> 2012-07-10 09:26:02.244+0000: 10229: warning : qemuCapsInit:856 : Failed to get host power management capabilities
> 
> On Tue, Jul 10, 2012 at 8:16 AM, Razique Mahroua <razique.mahroua at gmail.com> wrote:
> Hi Leander,
> try to check libvirtd.log files, 
> is the instance still running on the first node while you are lauching the migration process ?
> 
> Razique
> Nuage & Co - Razique Mahroua 
> razique.mahroua at gmail.com
> 
> <NUAGECO-LOGO-Fblan_petit.jpg>
> 
> Le 9 juil. 2012 à 16:09, Leander Bessa Beernaert a écrit :
> 
>> Ok, so i've updated to the test packages from 
>> 
>> The migration still fails, but i see no errors in the logs. I'm trying to migrate a VM with the m1.tiny flavor from one machine to another. Their hardware are identical and they have more than enough resources to support the m1.tiny flavor:
>> 
>> cloud35         (total)                         4    3867     186
>> cloud35         (used_now)                      0     312       5
>> cloud35         (used_max)                      0       0       0
>> 
>> These are the logs from the origin compute node:  http://paste.openstack.org/show/19319/  and  the destination compute node: http://paste.openstack.org/show/19318/ . The scheduler's log has no visible errors or stack traces.
>> 
>> I'm still using nfsv4. 
>> 
>> Any ideas?
>> 
>> 
>> On Fri, Jul 6, 2012 at 7:57 PM, Leander Bessa Beernaert <leanderbb at gmail.com> wrote:
>> Thanks for the tip, it's a better than nothing :)
>> 
>> Regards,
>> Leander
>> 
>> On Fri, Jul 6, 2012 at 6:32 PM, Mandar Vaze / मंदार वझे <mandarvaze at gmail.com> wrote:
>> Not sure if you are able to debug this, but a while ago there was a bug where instance.id was passed where instance.uuid was expected. This used to cause some problem.
>> It looks like you are using distribution package rather than devstack installation, so it is likely that the issue is now fixed. Can you try latest packages (and/or try devstack if you can)
>> 
>> I wish I could help more.
>> 
>> -Mandar
>> 
>> 
>> On Fri, Jul 6, 2012 at 3:26 PM, Leander Bessa Beernaert <leanderbb at gmail.com> wrote:
>> Hello,
>> 
>> I've recently setup a system to test out the live migration feature. So far i've been able to launch the instances with the shared nfs folder. However, when i run the live-migration command i encounter this error in the destination compute node:
>> 
>> 2012-07-05 09:33:48 ERROR nova.manager [-] Error during ComputeManager.update_available_resource: Domain not found: no domain with matching id 2
>> 2012-07-05 09:33:48 TRACE nova.manager Traceback (most recent call last):
>> 2012-07-05 09:33:48 TRACE nova.manager   File "/usr/lib/python2.7/dist-packages/nova/manager.py", line 155, in periodic_tasks
>> 2012-07-05 09:33:48 TRACE nova.manager     task(self, context)
>> 2012-07-05 09:33:48 TRACE nova.manager   File "/usr/lib/python2.7/dist-packages/nova/compute/manager.py", line 2409, in update_available_resource
>> 2012-07-05 09:33:48 TRACE nova.manager     self.driver.update_available_resource(context, self.host)
>> 2012-07-05 09:33:48 TRACE nova.manager   File "/usr/lib/python2.7/dist-packages/nova/virt/libvirt/connection.py", line 1936, in update_available_resource
>> 2012-07-05 09:33:48 TRACE nova.manager     'vcpus_used': self.get_vcpu_used(),
>> 2012-07-05 09:33:48 TRACE nova.manager   File "/usr/lib/python2.7/dist-packages/nova/virt/libvirt/connection.py", line 1743, in get_vcpu_used
>> 2012-07-05 09:33:48 TRACE nova.manager     dom = self._conn.lookupByID(dom_id)
>> 2012-07-05 09:33:48 TRACE nova.manager   File "/usr/lib/python2.7/dist-packages/libvirt.py", line 2363, in lookupByID
>> 2012-07-05 09:33:48 TRACE nova.manager     if ret is None:raise libvirtError('virDomainLookupByID() failed', conn=self)
>> 2012-07-05 09:33:48 TRACE nova.manager libvirtError: Domain not found: no domain with matching id 2
>> 
>> Any ideas on how to solve this?
>> 
>> Regards,
>> Leander
>> 
>> _______________________________________________
>> Mailing list: https://launchpad.net/~openstack
>> Post to     : openstack at lists.launchpad.net
>> Unsubscribe : https://launchpad.net/~openstack
>> More help   : https://help.launchpad.net/ListHelp
>> 
>> 
>> 
>> 
>> _______________________________________________
>> Mailing list: https://launchpad.net/~openstack
>> Post to     : openstack at lists.launchpad.net
>> Unsubscribe : https://launchpad.net/~openstack
>> More help   : https://help.launchpad.net/ListHelp
> 
> 

-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.openstack.org/pipermail/openstack/attachments/20120710/1b25db64/attachment.html>
-------------- next part --------------
A non-text attachment was scrubbed...
Name: NUAGECO-LOGO-Fblan_petit.jpg
Type: image/jpeg
Size: 10122 bytes
Desc: not available
URL: <http://lists.openstack.org/pipermail/openstack/attachments/20120710/1b25db64/attachment.jpg>


More information about the Openstack mailing list