[Openstack] SSH permissions (was: Re: Error during live migration)
Guilherme Russi
luisguilherme.cr at gmail.com
Tue Aug 20 13:06:06 UTC 2013
Hello Razique, I'm trying to run this command: nova live-migration
f18b6d2b-f685-4476-a7c0-e735db5113cf 192.168.3.3
Where f18b6d2b-f685-4476-a7c0-e735db5113cf is the ID from my new VM and
192.168.3.3 is the destiny compute node.
I'm executing this command from my controller node and I have all IPs and
machine names at my /etc/hosts, but I'm getting this error:
ERROR: Compute service of 192.168.3.3 is unavailable at this time. (HTTP
400) (Request-ID: req-117a0eea-14e8-4db2-a3d0-cbb14e438f85)
It's almost there (I guess), hope you all can help me :)
And Julie, thank you for saying that, now I'm only trying to migrate
through the terminal.
Regards.
Guilherme.
2013/8/20 Razique Mahroua <razique.mahroua at gmail.com>
> It's getting weird really…
> now can you run a live migration from the command line?
>
> Le 20 août 2013 à 00:20, Guilherme Russi <luisguilherme.cr at gmail.com> a
> écrit :
>
> Yes:
>
> nova-manage service list
> Binary Host Zone
> Status State Updated_At
> nova-cert hemera internal
> enabled :-) 2013-08-19 22:20:14
> nova-scheduler hemera internal
> enabled :-) 2013-08-19 22:20:14
> nova-consoleauth hemera internal
> enabled :-) 2013-08-19 22:20:17
> nova-conductor hemera internal
> enabled :-) 2013-08-19 22:20:14
> nova-compute caos nova
> enabled :-) 2013-08-19 22:20:17
> nova-compute tiresias nova
> enabled :-) 2013-08-19 22:20:17
>
>
>
> 2013/8/19 Razique Mahroua <razique.mahroua at gmail.com>
>
>> Can you run nova-manage service list ?
>>
>> *Razique Mahroua** - **Nuage & Co*
>> razique.mahroua at gmail.com
>> Tel : +33 9 72 37 94 15
>>
>> <NUAGECO-LOGO-Fblan_petit.jpg>
>>
>> Le 19 août 2013 à 22:27, Guilherme Russi <luisguilherme.cr at gmail.com> a
>> écrit :
>>
>> Hello again Razique, I was trying to migrate through the terminal and I'm
>> getting this line from my controller node:
>>
>> nova live-migration f18b6d2b-f685-4476-a7c0-e735db5113cf 192.168.3.3
>> ERROR: Compute service of 192.168.3.3 is unavailable at this time. (HTTP
>> 400) (Request-ID: req-454960fb-6652-46ed-9080-6d23b384cbb3)
>>
>> Haven't found anything yet. Any ideas?
>>
>> Regards.
>>
>> Guilherme.
>>
>>
>> 2013/8/19 Guilherme Russi <luisguilherme.cr at gmail.com>
>>
>>> I'm migrating it through the dashboard with the admin user. Is it wrong?
>>>
>>> Thank you.
>>>
>>> Guilherme.
>>>
>>>
>>> 2013/8/19 Razique Mahroua <razique.mahroua at gmail.com>
>>>
>>>> quick question, how do you migrate the instance ? and does it belong to
>>>> the tenant?
>>>>
>>>>
>>>> Le 19 août 2013 à 15:36, Guilherme Russi <luisguilherme.cr at gmail.com>
>>>> a écrit :
>>>>
>>>> The VM looks to be there:
>>>>
>>>> du -h instances/
>>>> 4,0K instances/keys
>>>> 1,9G instances/_base
>>>> 0 instances/locks
>>>> 310M instances/edca1460-9c97-475e-997b-3266b235e797
>>>> 260M instances/0fe59b28-04a3-41bb-9926-b7d139a70548
>>>> 323M instances/aeda1c56-72ba-40a7-9857-01a622540505
>>>> 7,0M instances/f4af662d-ae55-4346-82cf-a997192706b5
>>>> 264M instances/483f98e3-8ef5-43e2-8c3a-def55abdabcd
>>>> 268M instances/eeb81e32-32ce-49fa-8249-afc572f26233
>>>> 268M instances/c0224f0d-d7e8-4621-8c1a-af68740a3267
>>>> 188M instances/72ec37a3-b209-4729-b628-005fdcea5a3c
>>>> 12M instances/2c7c123b-456a-4c70-a991-80efee541324
>>>> 11M instances/c9e1c5ed-a108-4196-bfbc-24495e2e71bd
>>>> 6,5M instances/96d8f2de-a72d-40b5-b083-1f5e511f91ce
>>>> 531M instances/1b35a4a6-e158-48d3-a804-44b05ec171d0
>>>> 529M instances/1a2b6146-6ef0-4d76-a603-e4715b4dfb62
>>>> 514M instances/f6cb9f81-9281-419f-9e20-adaa6a07cea3
>>>> 2,7G instances/75e479cf-37a4-4511-b2a5-d4f3027702d4_resize
>>>> 2,4G instances/75e479cf-37a4-4511-b2a5-d4f3027702d4
>>>> 11G instances/
>>>>
>>>>
>>>> It's still running, no Error messagem from dashboard (it is showing
>>>> Status: Resize/Migrate - Task: Resizing or Migrating - Power State:
>>>> Running). The logs are not showing that permission error anymore.
>>>> Should I wait to see if it will migrate or returns another error status?
>>>> One thing is true, the machine is really slow now :P
>>>>
>>>> Regards.
>>>>
>>>> Guilherme.
>>>>
>>>>
>>>> 2013/8/19 Razique Mahroua <razique.mahroua at gmail.com>
>>>>
>>>>> Looks like the file doesn't exist.
>>>>> Can you spawn a new instance and try to migrate it again?
>>>>> thanks
>>>>>
>>>>> *Razique Mahroua** - **Nuage & Co*
>>>>> razique.mahroua at gmail.com
>>>>> Tel : +33 9 72 37 94 15
>>>>>
>>>>> <NUAGECO-LOGO-Fblan_petit.jpg>
>>>>>
>>>>> Le 19 août 2013 à 13:41, Guilherme Russi <luisguilherme.cr at gmail.com>
>>>>> a écrit :
>>>>>
>>>>> Hello Peter,
>>>>>
>>>>> Thank you for your help too :)
>>>>> I've changed the permission and now I'm getting another error when I
>>>>> try to migrate:
>>>>>
>>>>> cat /var/log/nova/nova-compute.log | grep -E ERROR
>>>>> 2013-08-19 08:17:19.990 12667 ERROR nova.manager [-] Error during
>>>>> ComputeManager._run_image_cache_manager_pass: Unexpected error while
>>>>> running command.
>>>>> 2013-08-19 08:23:40.309 12667 ERROR nova.network.quantumv2 [-]
>>>>> _get_auth_token() failed
>>>>> 2013-08-19 08:24:27.013 12667 ERROR nova.virt.libvirt.driver [-]
>>>>> Getting disk size of instance-000000bc: [Errno 2] No such file or
>>>>> directory:
>>>>> '/var/lib/nova/instances/75e479cf-37a4-4511-b2a5-d4f3027702d4/disk'
>>>>>
>>>>>
>>>>> cat /var/log/nova/nova-compute.log | grep -E TRACE
>>>>> 2013-08-19 08:17:19.990 12667 TRACE nova.manager Traceback (most
>>>>> recent call last):
>>>>> 2013-08-19 08:17:19.990 12667 TRACE nova.manager File
>>>>> "/usr/lib/python2.7/dist-packages/nova/manager.py", line 241, in
>>>>> periodic_tasks
>>>>> 2013-08-19 08:17:19.990 12667 TRACE nova.manager task(self,
>>>>> context)
>>>>> 2013-08-19 08:17:19.990 12667 TRACE nova.manager File
>>>>> "/usr/lib/python2.7/dist-packages/nova/compute/manager.py", line 4020, in
>>>>> _run_image_cache_manager_pass
>>>>> 2013-08-19 08:17:19.990 12667 TRACE nova.manager
>>>>> self.driver.manage_image_cache(context, filtered_instances)
>>>>> 2013-08-19 08:17:19.990 12667 TRACE nova.manager File
>>>>> "/usr/lib/python2.7/dist-packages/nova/virt/libvirt/driver.py", line 3401,
>>>>> in manage_image_cache
>>>>> 2013-08-19 08:17:19.990 12667 TRACE nova.manager
>>>>> self.image_cache_manager.verify_base_images(context, all_instances)
>>>>> 2013-08-19 08:17:19.990 12667 TRACE nova.manager File
>>>>> "/usr/lib/python2.7/dist-packages/nova/virt/libvirt/imagecache.py", line
>>>>> 588, in verify_base_images
>>>>> 2013-08-19 08:17:19.990 12667 TRACE nova.manager
>>>>> self._handle_base_image(img, base_file)
>>>>> 2013-08-19 08:17:19.990 12667 TRACE nova.manager File
>>>>> "/usr/lib/python2.7/dist-packages/nova/virt/libvirt/imagecache.py", line
>>>>> 553, in _handle_base_image
>>>>> 2013-08-19 08:17:19.990 12667 TRACE nova.manager
>>>>> virtutils.chown(base_file, os.getuid())
>>>>> 2013-08-19 08:17:19.990 12667 TRACE nova.manager File
>>>>> "/usr/lib/python2.7/dist-packages/nova/virt/libvirt/utils.py", line 449, in
>>>>> chown
>>>>> 2013-08-19 08:17:19.990 12667 TRACE nova.manager execute('chown',
>>>>> owner, path, run_as_root=True)
>>>>> 2013-08-19 08:17:19.990 12667 TRACE nova.manager File
>>>>> "/usr/lib/python2.7/dist-packages/nova/virt/libvirt/utils.py", line 47, in
>>>>> execute
>>>>> 2013-08-19 08:17:19.990 12667 TRACE nova.manager return
>>>>> utils.execute(*args, **kwargs)
>>>>> 2013-08-19 08:17:19.990 12667 TRACE nova.manager File
>>>>> "/usr/lib/python2.7/dist-packages/nova/utils.py", line 239, in execute
>>>>> 2013-08-19 08:17:19.990 12667 TRACE nova.manager cmd='
>>>>> '.join(cmd))
>>>>> 2013-08-19 08:17:19.990 12667 TRACE nova.manager
>>>>> ProcessExecutionError: Unexpected error while running command.
>>>>> 2013-08-19 08:17:19.990 12667 TRACE nova.manager Command: sudo
>>>>> nova-rootwrap /etc/nova/rootwrap.conf chown 123
>>>>> /var/lib/nova/instances/_base/b0d90eb60f048f6de16c3efbc0eb4c16cbac77a6
>>>>> 2013-08-19 08:17:19.990 12667 TRACE nova.manager Exit code: 1
>>>>> 2013-08-19 08:17:19.990 12667 TRACE nova.manager Stdout: ''
>>>>> 2013-08-19 08:17:19.990 12667 TRACE nova.manager Stderr: 'Traceback
>>>>> (most recent call last):\n File "/usr/bin/nova-rootwrap", line 108, in
>>>>> <module>\n os.getlogin(), pwd.getpwuid(os.getuid())[0],\nOSError: [Errno
>>>>> 22] Invalid argument\n'
>>>>> 2013-08-19 08:17:19.990 12667 TRACE nova.manager
>>>>> 2013-08-19 08:23:40.309 12667 TRACE nova.network.quantumv2 Traceback
>>>>> (most recent call last):
>>>>> 2013-08-19 08:23:40.309 12667 TRACE nova.network.quantumv2 File
>>>>> "/usr/lib/python2.7/dist-packages/nova/network/quantumv2/__init__.py", line
>>>>> 40, in _get_auth_token
>>>>> 2013-08-19 08:23:40.309 12667 TRACE nova.network.quantumv2
>>>>> httpclient.authenticate()
>>>>> 2013-08-19 08:23:40.309 12667 TRACE nova.network.quantumv2 File
>>>>> "/usr/lib/python2.7/dist-packages/quantumclient/client.py", line 198, in
>>>>> authenticate
>>>>> 2013-08-19 08:23:40.309 12667 TRACE nova.network.quantumv2 raise
>>>>> exceptions.Unauthorized(message=body)
>>>>> 2013-08-19 08:23:40.309 12667 TRACE nova.network.quantumv2
>>>>> Unauthorized: Request Timeout
>>>>> 2013-08-19 08:23:40.309 12667 TRACE nova.network.quantumv2
>>>>>
>>>>> Any ideas?
>>>>>
>>>>> Regards.
>>>>>
>>>>> Guilherme.
>>>>>
>>>>>
>>>>> 2013/8/19 Peter Funk <pf at artcom-gmbh.de>
>>>>>
>>>>>> Guilherme Russi wrote 16.08.2013 um 09:53:
>>>>>> > Here is my directory configs:
>>>>>> >
>>>>>> > drwxrwxrwx 2 nova nova 76 Ago 15 14:54 .ssh
>>>>>> >
>>>>>> > I made chmod a+xwr at the folder at all computers, I made it
>>>>>> because I
>>>>>> > couldn't find the reason I was no allowed to migrate, can it be
>>>>>> like that?
>>>>>>
>>>>>> For SSH to work the permissions of $HOME/.ssh should look like this:
>>>>>> drwx------ 2 nova nova 76 Ago 15 14:54 .ssh
>>>>>>
>>>>>> Use:
>>>>>> chmod -R go-rwx .ssh
>>>>>>
>>>>>> Regards, Peter Funk
>>>>>> --
>>>>>> Peter Funk, home: ✉Oldenburger Str.86, D-27777 Ganderkesee
>>>>>> mobile:+49-179-640-8878 phone:+49-421-20419-0 <
>>>>>> http://www.artcom-gmbh.de/>
>>>>>> office: ArtCom GmbH, ✉Haferwende 2, D-28357 Bremen, Germany
>>>>>>
>>>>>> _______________________________________________
>>>>>> Mailing list:
>>>>>> http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack
>>>>>> Post to : openstack at lists.openstack.org
>>>>>> Unsubscribe :
>>>>>> http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack
>>>>>>
>>>>>
>>>>> _______________________________________________
>>>>> Mailing list:
>>>>> http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack
>>>>> Post to : openstack at lists.openstack.org
>>>>> Unsubscribe :
>>>>> http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack
>>>>>
>>>>>
>>>>>
>>>>
>>>>
>>>
>>
>>
>
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.openstack.org/pipermail/openstack/attachments/20130820/7373ea44/attachment.html>
More information about the Openstack
mailing list