[Openstack] [OpenStack][Nova] Live Migration Error

Sébastien Han han.sebastien at gmail.com
Tue Jul 10 09:49:44 UTC 2012


I forgot to ask, did you enable the vnc console?

If so, with which parameters?


On Tue, Jul 10, 2012 at 11:48 AM, Sébastien Han <han.sebastien at gmail.com>wrote:

> Hi!
>
> Usually you get:
>
> 2012-07-09 13:58:27.179+0000: 10227: warning : qemuDomainObjTaint:1134 :
>> Domain id=2 name='instance-00000002'
>> uuid=57aca8a6-d062-4a08-8d87-e4d11d259ac7 is tainted: high-privileges
>
>
> when you change permission in libvirt (root I presumed) which is not
> necessary.
>
> 2012-07-10 09:26:02.244+0000: 10229: error : virExecWithHook:328 : Cannot
>> find 'pm-is-supported' in path: No such file or directory
>
>
> This error is harmless and can be easily solved by installing the
> following package:
>
> sudo apt-get install pm-utils -y
>
>
> Do you have something in the nova-scheduler logs?
>
> Cheers!
>
> On Tue, Jul 10, 2012 at 11:29 AM, Leander Bessa Beernaert <
> leanderbb at gmail.com> wrote:
>
>> Well i've checked the libvirt logs on both nodes and i found these two
>> lines:
>>
>> 2012-07-09 13:58:27.179+0000: 10227: warning : qemuDomainObjTaint:1134 :
>>> Domain id=2 name='instance-00000002'
>>> uuid=57aca8a6-d062-4a08-8d87-e4d11d259ac7 is tainted: high-privileges
>>> 2012-07-09 13:58:27.736+0000: 10226: error : qemuMonitorIORead:513 :
>>> Unable to read from monitor: Connection reset by peer
>>
>>
>> The log is alos filled with the message below, it gets repeated over and
>> over and over.
>>
>> 2012-07-10 09:26:02.244+0000: 10229: error : virExecWithHook:328 : Cannot
>>> find 'pm-is-supported' in path: No such file or directory
>>> 2012-07-10 09:26:02.244+0000: 10229: warning : qemuCapsInit:856 : Failed
>>> to get host power management capabilities
>>
>>
>> On Tue, Jul 10, 2012 at 8:16 AM, Razique Mahroua <
>> razique.mahroua at gmail.com> wrote:
>>
>>> Hi Leander,
>>> try to check libvirtd.log files,
>>> is the instance still running on the first node while you are lauching
>>> the migration process ?
>>>
>>> Razique
>>>  *Nuage & Co - Razique Mahroua** *
>>> razique.mahroua at gmail.com
>>>
>>>
>>> Le 9 juil. 2012 à 16:09, Leander Bessa Beernaert a écrit :
>>>
>>> Ok, so i've updated to the test packages from
>>>
>>> The migration still fails, but i see no errors in the logs. I'm trying
>>> to migrate a VM with the m1.tiny flavor from one machine to another. Their
>>> hardware are identical and they have more than enough resources to support
>>> the m1.tiny flavor:
>>>
>>> cloud35         (total)                         4    3867     186
>>>> cloud35         (used_now)                      0     312       5
>>>> cloud35         (used_max)                      0       0       0
>>>
>>>
>>> These are the logs from the origin compute node:
>>> http://paste.openstack.org/show/19319/  and  the destination compute
>>> node: http://paste.openstack.org/show/19318/ . The scheduler's log has
>>> no visible errors or stack traces.
>>>
>>> I'm still using nfsv4.
>>>
>>> Any ideas?
>>>
>>>
>>> On Fri, Jul 6, 2012 at 7:57 PM, Leander Bessa Beernaert <
>>> leanderbb at gmail.com> wrote:
>>>
>>>> Thanks for the tip, it's a better than nothing :)
>>>>
>>>> Regards,
>>>> Leander
>>>>
>>>> On Fri, Jul 6, 2012 at 6:32 PM, Mandar Vaze / मंदार वझे <
>>>> mandarvaze at gmail.com> wrote:
>>>>
>>>>> Not sure if you are able to debug this, but a while ago there was a
>>>>> bug where instance.id was passed where instance.uuid was expected.
>>>>> This used to cause some problem.
>>>>> It looks like you are using distribution package rather than devstack
>>>>> installation, so it is likely that the issue is now fixed. Can you try
>>>>> latest packages (and/or try devstack if you can)
>>>>>
>>>>> I wish I could help more.
>>>>>
>>>>> -Mandar
>>>>>
>>>>>
>>>>> On Fri, Jul 6, 2012 at 3:26 PM, Leander Bessa Beernaert <
>>>>> leanderbb at gmail.com> wrote:
>>>>>
>>>>>> Hello,
>>>>>>
>>>>>> I've recently setup a system to test out the live migration feature.
>>>>>> So far i've been able to launch the instances with the shared nfs folder.
>>>>>> However, when i run the live-migration command i encounter this error in
>>>>>> the destination compute node:
>>>>>>
>>>>>> 2012-07-05 09:33:48 ERROR nova.manager [-] Error during
>>>>>>> ComputeManager.update_available_resource: Domain not found: no domain with
>>>>>>> matching id 2
>>>>>>> 2012-07-05 09:33:48 TRACE nova.manager Traceback (most recent call
>>>>>>> last):
>>>>>>> 2012-07-05 09:33:48 TRACE nova.manager   File
>>>>>>> "/usr/lib/python2.7/dist-packages/nova/manager.py", line 155, in
>>>>>>> periodic_tasks
>>>>>>> 2012-07-05 09:33:48 TRACE nova.manager     task(self, context)
>>>>>>> 2012-07-05 09:33:48 TRACE nova.manager   File
>>>>>>> "/usr/lib/python2.7/dist-packages/nova/compute/manager.py", line 2409, in
>>>>>>> update_available_resource
>>>>>>> 2012-07-05 09:33:48 TRACE nova.manager
>>>>>>> self.driver.update_available_resource(context, self.host)
>>>>>>> 2012-07-05 09:33:48 TRACE nova.manager   File
>>>>>>> "/usr/lib/python2.7/dist-packages/nova/virt/libvirt/connection.py", line
>>>>>>> 1936, in update_available_resource
>>>>>>> 2012-07-05 09:33:48 TRACE nova.manager     'vcpus_used':
>>>>>>> self.get_vcpu_used(),
>>>>>>> 2012-07-05 09:33:48 TRACE nova.manager   File
>>>>>>> "/usr/lib/python2.7/dist-packages/nova/virt/libvirt/connection.py", line
>>>>>>> 1743, in get_vcpu_used
>>>>>>> 2012-07-05 09:33:48 TRACE nova.manager     dom =
>>>>>>> self._conn.lookupByID(dom_id)
>>>>>>> 2012-07-05 09:33:48 TRACE nova.manager   File
>>>>>>> "/usr/lib/python2.7/dist-packages/libvirt.py", line 2363, in lookupByID
>>>>>>> 2012-07-05 09:33:48 TRACE nova.manager     if ret is None:raise
>>>>>>> libvirtError('virDomainLookupByID() failed', conn=self)
>>>>>>> 2012-07-05 09:33:48 TRACE nova.manager libvirtError: Domain not
>>>>>>> found: no domain with matching id 2
>>>>>>
>>>>>>
>>>>>> Any ideas on how to solve this?
>>>>>>
>>>>>> Regards,
>>>>>> Leander
>>>>>>
>>>>>> _______________________________________________
>>>>>> Mailing list: https://launchpad.net/~openstack
>>>>>> Post to     : openstack at lists.launchpad.net
>>>>>> Unsubscribe : https://launchpad.net/~openstack
>>>>>> More help   : https://help.launchpad.net/ListHelp
>>>>>>
>>>>>>
>>>>>
>>>>
>>> _______________________________________________
>>> Mailing list: https://launchpad.net/~openstack
>>> Post to     : openstack at lists.launchpad.net
>>> Unsubscribe : https://launchpad.net/~openstack
>>> More help   : https://help.launchpad.net/ListHelp
>>>
>>>
>>>
>>
>> _______________________________________________
>> Mailing list: https://launchpad.net/~openstack
>> Post to     : openstack at lists.launchpad.net
>> Unsubscribe : https://launchpad.net/~openstack
>> More help   : https://help.launchpad.net/ListHelp
>>
>>
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.openstack.org/pipermail/openstack/attachments/20120710/1748b6a0/attachment.html>


More information about the Openstack mailing list