[Openstack] SSH permissions (was: Re: Error during live migration)

Guilherme Russi luisguilherme.cr at gmail.com
Mon Aug 19 20:27:19 UTC 2013


Hello again Razique, I was trying to migrate through the terminal and I'm
getting this line from my controller node:

nova live-migration f18b6d2b-f685-4476-a7c0-e735db5113cf 192.168.3.3
ERROR: Compute service of 192.168.3.3 is unavailable at this time. (HTTP
400) (Request-ID: req-454960fb-6652-46ed-9080-6d23b384cbb3)

Haven't found anything yet. Any ideas?

Regards.

Guilherme.


2013/8/19 Guilherme Russi <luisguilherme.cr at gmail.com>

> I'm migrating it through the dashboard with the admin user. Is it wrong?
>
> Thank you.
>
> Guilherme.
>
>
> 2013/8/19 Razique Mahroua <razique.mahroua at gmail.com>
>
>> quick question, how do you migrate the instance ? and does it belong to
>> the tenant?
>>
>>
>> Le 19 août 2013 à 15:36, Guilherme Russi <luisguilherme.cr at gmail.com> a
>> écrit :
>>
>> The VM looks to be there:
>>
>> du -h instances/
>> 4,0K instances/keys
>> 1,9G instances/_base
>> 0 instances/locks
>> 310M instances/edca1460-9c97-475e-997b-3266b235e797
>> 260M instances/0fe59b28-04a3-41bb-9926-b7d139a70548
>> 323M instances/aeda1c56-72ba-40a7-9857-01a622540505
>> 7,0M instances/f4af662d-ae55-4346-82cf-a997192706b5
>> 264M instances/483f98e3-8ef5-43e2-8c3a-def55abdabcd
>> 268M instances/eeb81e32-32ce-49fa-8249-afc572f26233
>> 268M instances/c0224f0d-d7e8-4621-8c1a-af68740a3267
>> 188M instances/72ec37a3-b209-4729-b628-005fdcea5a3c
>> 12M instances/2c7c123b-456a-4c70-a991-80efee541324
>> 11M instances/c9e1c5ed-a108-4196-bfbc-24495e2e71bd
>> 6,5M instances/96d8f2de-a72d-40b5-b083-1f5e511f91ce
>> 531M instances/1b35a4a6-e158-48d3-a804-44b05ec171d0
>> 529M instances/1a2b6146-6ef0-4d76-a603-e4715b4dfb62
>> 514M instances/f6cb9f81-9281-419f-9e20-adaa6a07cea3
>> 2,7G instances/75e479cf-37a4-4511-b2a5-d4f3027702d4_resize
>> 2,4G instances/75e479cf-37a4-4511-b2a5-d4f3027702d4
>> 11G instances/
>>
>>
>> It's still running, no Error messagem from dashboard (it is showing
>> Status: Resize/Migrate - Task: Resizing or Migrating - Power State:
>> Running). The logs are not showing that permission error anymore. Should
>> I wait to see if it will migrate or returns another error status? One thing
>> is true, the machine is really slow now :P
>>
>> Regards.
>>
>> Guilherme.
>>
>>
>> 2013/8/19 Razique Mahroua <razique.mahroua at gmail.com>
>>
>>> Looks like the file doesn't exist.
>>> Can you spawn a new instance and try to migrate it again?
>>> thanks
>>>
>>> *Razique Mahroua** - **Nuage & Co*
>>> razique.mahroua at gmail.com
>>> Tel : +33 9 72 37 94 15
>>>
>>> <NUAGECO-LOGO-Fblan_petit.jpg>
>>>
>>> Le 19 août 2013 à 13:41, Guilherme Russi <luisguilherme.cr at gmail.com> a
>>> écrit :
>>>
>>> Hello Peter,
>>>
>>>  Thank you for your help too :)
>>>  I've changed the permission and now I'm getting another error when I
>>> try to migrate:
>>>
>>> cat /var/log/nova/nova-compute.log | grep -E ERROR
>>> 2013-08-19 08:17:19.990 12667 ERROR nova.manager [-] Error during
>>> ComputeManager._run_image_cache_manager_pass: Unexpected error while
>>> running command.
>>> 2013-08-19 08:23:40.309 12667 ERROR nova.network.quantumv2 [-]
>>> _get_auth_token() failed
>>> 2013-08-19 08:24:27.013 12667 ERROR nova.virt.libvirt.driver [-] Getting
>>> disk size of instance-000000bc: [Errno 2] No such file or directory:
>>> '/var/lib/nova/instances/75e479cf-37a4-4511-b2a5-d4f3027702d4/disk'
>>>
>>>
>>> cat /var/log/nova/nova-compute.log | grep -E TRACE
>>> 2013-08-19 08:17:19.990 12667 TRACE nova.manager Traceback (most recent
>>> call last):
>>> 2013-08-19 08:17:19.990 12667 TRACE nova.manager   File
>>> "/usr/lib/python2.7/dist-packages/nova/manager.py", line 241, in
>>> periodic_tasks
>>> 2013-08-19 08:17:19.990 12667 TRACE nova.manager     task(self, context)
>>> 2013-08-19 08:17:19.990 12667 TRACE nova.manager   File
>>> "/usr/lib/python2.7/dist-packages/nova/compute/manager.py", line 4020, in
>>> _run_image_cache_manager_pass
>>> 2013-08-19 08:17:19.990 12667 TRACE nova.manager
>>> self.driver.manage_image_cache(context, filtered_instances)
>>> 2013-08-19 08:17:19.990 12667 TRACE nova.manager   File
>>> "/usr/lib/python2.7/dist-packages/nova/virt/libvirt/driver.py", line 3401,
>>> in manage_image_cache
>>> 2013-08-19 08:17:19.990 12667 TRACE nova.manager
>>> self.image_cache_manager.verify_base_images(context, all_instances)
>>> 2013-08-19 08:17:19.990 12667 TRACE nova.manager   File
>>> "/usr/lib/python2.7/dist-packages/nova/virt/libvirt/imagecache.py", line
>>> 588, in verify_base_images
>>> 2013-08-19 08:17:19.990 12667 TRACE nova.manager
>>> self._handle_base_image(img, base_file)
>>> 2013-08-19 08:17:19.990 12667 TRACE nova.manager   File
>>> "/usr/lib/python2.7/dist-packages/nova/virt/libvirt/imagecache.py", line
>>> 553, in _handle_base_image
>>> 2013-08-19 08:17:19.990 12667 TRACE nova.manager
>>> virtutils.chown(base_file, os.getuid())
>>> 2013-08-19 08:17:19.990 12667 TRACE nova.manager   File
>>> "/usr/lib/python2.7/dist-packages/nova/virt/libvirt/utils.py", line 449, in
>>> chown
>>> 2013-08-19 08:17:19.990 12667 TRACE nova.manager     execute('chown',
>>> owner, path, run_as_root=True)
>>> 2013-08-19 08:17:19.990 12667 TRACE nova.manager   File
>>> "/usr/lib/python2.7/dist-packages/nova/virt/libvirt/utils.py", line 47, in
>>> execute
>>> 2013-08-19 08:17:19.990 12667 TRACE nova.manager     return
>>> utils.execute(*args, **kwargs)
>>> 2013-08-19 08:17:19.990 12667 TRACE nova.manager   File
>>> "/usr/lib/python2.7/dist-packages/nova/utils.py", line 239, in execute
>>> 2013-08-19 08:17:19.990 12667 TRACE nova.manager     cmd=' '.join(cmd))
>>> 2013-08-19 08:17:19.990 12667 TRACE nova.manager ProcessExecutionError:
>>> Unexpected error while running command.
>>> 2013-08-19 08:17:19.990 12667 TRACE nova.manager Command: sudo
>>> nova-rootwrap /etc/nova/rootwrap.conf chown 123
>>> /var/lib/nova/instances/_base/b0d90eb60f048f6de16c3efbc0eb4c16cbac77a6
>>> 2013-08-19 08:17:19.990 12667 TRACE nova.manager Exit code: 1
>>> 2013-08-19 08:17:19.990 12667 TRACE nova.manager Stdout: ''
>>> 2013-08-19 08:17:19.990 12667 TRACE nova.manager Stderr: 'Traceback
>>> (most recent call last):\n  File "/usr/bin/nova-rootwrap", line 108, in
>>> <module>\n    os.getlogin(), pwd.getpwuid(os.getuid())[0],\nOSError: [Errno
>>> 22] Invalid argument\n'
>>> 2013-08-19 08:17:19.990 12667 TRACE nova.manager
>>> 2013-08-19 08:23:40.309 12667 TRACE nova.network.quantumv2 Traceback
>>> (most recent call last):
>>> 2013-08-19 08:23:40.309 12667 TRACE nova.network.quantumv2   File
>>> "/usr/lib/python2.7/dist-packages/nova/network/quantumv2/__init__.py", line
>>> 40, in _get_auth_token
>>> 2013-08-19 08:23:40.309 12667 TRACE nova.network.quantumv2
>>> httpclient.authenticate()
>>> 2013-08-19 08:23:40.309 12667 TRACE nova.network.quantumv2   File
>>> "/usr/lib/python2.7/dist-packages/quantumclient/client.py", line 198, in
>>> authenticate
>>> 2013-08-19 08:23:40.309 12667 TRACE nova.network.quantumv2     raise
>>> exceptions.Unauthorized(message=body)
>>> 2013-08-19 08:23:40.309 12667 TRACE nova.network.quantumv2 Unauthorized:
>>> Request Timeout
>>> 2013-08-19 08:23:40.309 12667 TRACE nova.network.quantumv2
>>>
>>> Any ideas?
>>>
>>> Regards.
>>>
>>> Guilherme.
>>>
>>>
>>> 2013/8/19 Peter Funk <pf at artcom-gmbh.de>
>>>
>>>> Guilherme Russi wrote 16.08.2013 um 09:53:
>>>> >  Here is my directory configs:
>>>> >
>>>> > drwxrwxrwx  2 nova nova     76 Ago 15 14:54 .ssh
>>>> >
>>>> > I made chmod a+xwr at the folder at all computers, I made it because I
>>>> > couldn't find the reason I was no allowed to migrate, can it be like
>>>> that?
>>>>
>>>> For SSH to work the permissions of $HOME/.ssh should look like this:
>>>>   drwx------  2 nova nova     76 Ago 15 14:54 .ssh
>>>>
>>>> Use:
>>>>   chmod -R go-rwx .ssh
>>>>
>>>> Regards, Peter Funk
>>>> --
>>>> Peter Funk, home: ✉Oldenburger Str.86, D-27777 Ganderkesee
>>>> mobile:+49-179-640-8878 phone:+49-421-20419-0 <
>>>> http://www.artcom-gmbh.de/>
>>>> office: ArtCom GmbH, ✉Haferwende 2, D-28357 Bremen, Germany
>>>>
>>>> _______________________________________________
>>>> Mailing list:
>>>> http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack
>>>> Post to     : openstack at lists.openstack.org
>>>> Unsubscribe :
>>>> http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack
>>>>
>>>
>>> _______________________________________________
>>> Mailing list:
>>> http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack
>>> Post to     : openstack at lists.openstack.org
>>> Unsubscribe :
>>> http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack
>>>
>>>
>>>
>>
>>
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.openstack.org/pipermail/openstack/attachments/20130819/9facd44b/attachment.html>


More information about the Openstack mailing list