[Openstack] [OpenStack][Nova] Live Migration Error

Leander Bessa Beernaert leanderbb at gmail.com
Tue Jul 10 09:52:45 UTC 2012


If i don't run libvirt with root, it can't write to the shared folder. It's
the only way i've been able to get this to work. :S

Below follows the configuration of one of the compute nodes. 10.0.1.1 is
the controller and 10.0.1.2 is the compute node.

novncproxy_base_url=http://10.0.1.1:6080/vnc_auto.html
> xvpvncproxy_base_url=http://10.0.1.1:6081/console
> vncserver_proxyclient_address=10.0.1.2
> vncserver_listen=10.0.1.2


On Tue, Jul 10, 2012 at 10:49 AM, Sébastien Han <han.sebastien at gmail.com>wrote:

> I forgot to ask, did you enable the vnc console?
>
> If so, with which parameters?
>
>
> On Tue, Jul 10, 2012 at 11:48 AM, Sébastien Han <han.sebastien at gmail.com>wrote:
>
>> Hi!
>>
>> Usually you get:
>>
>> 2012-07-09 13:58:27.179+0000: 10227: warning : qemuDomainObjTaint:1134 :
>>> Domain id=2 name='instance-00000002'
>>> uuid=57aca8a6-d062-4a08-8d87-e4d11d259ac7 is tainted: high-privileges
>>
>>
>> when you change permission in libvirt (root I presumed) which is not
>> necessary.
>>
>> 2012-07-10 09:26:02.244+0000: 10229: error : virExecWithHook:328 : Cannot
>>> find 'pm-is-supported' in path: No such file or directory
>>
>>
>> This error is harmless and can be easily solved by installing the
>> following package:
>>
>> sudo apt-get install pm-utils -y
>>
>>
>> Do you have something in the nova-scheduler logs?
>>
>> Cheers!
>>
>> On Tue, Jul 10, 2012 at 11:29 AM, Leander Bessa Beernaert <
>> leanderbb at gmail.com> wrote:
>>
>>> Well i've checked the libvirt logs on both nodes and i found these two
>>> lines:
>>>
>>> 2012-07-09 13:58:27.179+0000: 10227: warning : qemuDomainObjTaint:1134 :
>>>> Domain id=2 name='instance-00000002'
>>>> uuid=57aca8a6-d062-4a08-8d87-e4d11d259ac7 is tainted: high-privileges
>>>> 2012-07-09 13:58:27.736+0000: 10226: error : qemuMonitorIORead:513 :
>>>> Unable to read from monitor: Connection reset by peer
>>>
>>>
>>> The log is alos filled with the message below, it gets repeated over and
>>> over and over.
>>>
>>> 2012-07-10 09:26:02.244+0000: 10229: error : virExecWithHook:328 :
>>>> Cannot find 'pm-is-supported' in path: No such file or directory
>>>> 2012-07-10 09:26:02.244+0000: 10229: warning : qemuCapsInit:856 :
>>>> Failed to get host power management capabilities
>>>
>>>
>>> On Tue, Jul 10, 2012 at 8:16 AM, Razique Mahroua <
>>> razique.mahroua at gmail.com> wrote:
>>>
>>>> Hi Leander,
>>>> try to check libvirtd.log files,
>>>> is the instance still running on the first node while you are lauching
>>>> the migration process ?
>>>>
>>>> Razique
>>>>  *Nuage & Co - Razique Mahroua** *
>>>> razique.mahroua at gmail.com
>>>>
>>>>
>>>> Le 9 juil. 2012 à 16:09, Leander Bessa Beernaert a écrit :
>>>>
>>>> Ok, so i've updated to the test packages from
>>>>
>>>> The migration still fails, but i see no errors in the logs. I'm trying
>>>> to migrate a VM with the m1.tiny flavor from one machine to another. Their
>>>> hardware are identical and they have more than enough resources to support
>>>> the m1.tiny flavor:
>>>>
>>>> cloud35         (total)                         4    3867     186
>>>>> cloud35         (used_now)                      0     312       5
>>>>> cloud35         (used_max)                      0       0       0
>>>>
>>>>
>>>> These are the logs from the origin compute node:
>>>> http://paste.openstack.org/show/19319/  and  the destination compute
>>>> node: http://paste.openstack.org/show/19318/ . The scheduler's log has
>>>> no visible errors or stack traces.
>>>>
>>>> I'm still using nfsv4.
>>>>
>>>> Any ideas?
>>>>
>>>>
>>>> On Fri, Jul 6, 2012 at 7:57 PM, Leander Bessa Beernaert <
>>>> leanderbb at gmail.com> wrote:
>>>>
>>>>> Thanks for the tip, it's a better than nothing :)
>>>>>
>>>>> Regards,
>>>>> Leander
>>>>>
>>>>> On Fri, Jul 6, 2012 at 6:32 PM, Mandar Vaze / मंदार वझे <
>>>>> mandarvaze at gmail.com> wrote:
>>>>>
>>>>>> Not sure if you are able to debug this, but a while ago there was a
>>>>>> bug where instance.id was passed where instance.uuid was expected.
>>>>>> This used to cause some problem.
>>>>>> It looks like you are using distribution package rather than devstack
>>>>>> installation, so it is likely that the issue is now fixed. Can you try
>>>>>> latest packages (and/or try devstack if you can)
>>>>>>
>>>>>> I wish I could help more.
>>>>>>
>>>>>> -Mandar
>>>>>>
>>>>>>
>>>>>> On Fri, Jul 6, 2012 at 3:26 PM, Leander Bessa Beernaert <
>>>>>> leanderbb at gmail.com> wrote:
>>>>>>
>>>>>>> Hello,
>>>>>>>
>>>>>>> I've recently setup a system to test out the live migration feature.
>>>>>>> So far i've been able to launch the instances with the shared nfs folder.
>>>>>>> However, when i run the live-migration command i encounter this error in
>>>>>>> the destination compute node:
>>>>>>>
>>>>>>> 2012-07-05 09:33:48 ERROR nova.manager [-] Error during
>>>>>>>> ComputeManager.update_available_resource: Domain not found: no domain with
>>>>>>>> matching id 2
>>>>>>>> 2012-07-05 09:33:48 TRACE nova.manager Traceback (most recent call
>>>>>>>> last):
>>>>>>>> 2012-07-05 09:33:48 TRACE nova.manager   File
>>>>>>>> "/usr/lib/python2.7/dist-packages/nova/manager.py", line 155, in
>>>>>>>> periodic_tasks
>>>>>>>> 2012-07-05 09:33:48 TRACE nova.manager     task(self, context)
>>>>>>>> 2012-07-05 09:33:48 TRACE nova.manager   File
>>>>>>>> "/usr/lib/python2.7/dist-packages/nova/compute/manager.py", line 2409, in
>>>>>>>> update_available_resource
>>>>>>>> 2012-07-05 09:33:48 TRACE nova.manager
>>>>>>>> self.driver.update_available_resource(context, self.host)
>>>>>>>> 2012-07-05 09:33:48 TRACE nova.manager   File
>>>>>>>> "/usr/lib/python2.7/dist-packages/nova/virt/libvirt/connection.py", line
>>>>>>>> 1936, in update_available_resource
>>>>>>>> 2012-07-05 09:33:48 TRACE nova.manager     'vcpus_used':
>>>>>>>> self.get_vcpu_used(),
>>>>>>>> 2012-07-05 09:33:48 TRACE nova.manager   File
>>>>>>>> "/usr/lib/python2.7/dist-packages/nova/virt/libvirt/connection.py", line
>>>>>>>> 1743, in get_vcpu_used
>>>>>>>> 2012-07-05 09:33:48 TRACE nova.manager     dom =
>>>>>>>> self._conn.lookupByID(dom_id)
>>>>>>>> 2012-07-05 09:33:48 TRACE nova.manager   File
>>>>>>>> "/usr/lib/python2.7/dist-packages/libvirt.py", line 2363, in lookupByID
>>>>>>>> 2012-07-05 09:33:48 TRACE nova.manager     if ret is None:raise
>>>>>>>> libvirtError('virDomainLookupByID() failed', conn=self)
>>>>>>>> 2012-07-05 09:33:48 TRACE nova.manager libvirtError: Domain not
>>>>>>>> found: no domain with matching id 2
>>>>>>>
>>>>>>>
>>>>>>> Any ideas on how to solve this?
>>>>>>>
>>>>>>> Regards,
>>>>>>> Leander
>>>>>>>
>>>>>>> _______________________________________________
>>>>>>> Mailing list: https://launchpad.net/~openstack
>>>>>>> Post to     : openstack at lists.launchpad.net
>>>>>>> Unsubscribe : https://launchpad.net/~openstack
>>>>>>> More help   : https://help.launchpad.net/ListHelp
>>>>>>>
>>>>>>>
>>>>>>
>>>>>
>>>> _______________________________________________
>>>> Mailing list: https://launchpad.net/~openstack
>>>> Post to     : openstack at lists.launchpad.net
>>>> Unsubscribe : https://launchpad.net/~openstack
>>>> More help   : https://help.launchpad.net/ListHelp
>>>>
>>>>
>>>>
>>>
>>> _______________________________________________
>>> Mailing list: https://launchpad.net/~openstack
>>> Post to     : openstack at lists.launchpad.net
>>> Unsubscribe : https://launchpad.net/~openstack
>>> More help   : https://help.launchpad.net/ListHelp
>>>
>>>
>>
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.openstack.org/pipermail/openstack/attachments/20120710/edcbbca4/attachment.html>


More information about the Openstack mailing list