[Openstack] SSH permissions (was: Re: Error during live migration)
Joe Breu
joseph.breu at rackspace.com
Tue Aug 20 15:33:22 UTC 2013
You need to use the hostname of the compute node, not the IP address for migration. In your case this would be either caos or tiersias
On Aug 20, 2013, at 8:06 AM, Guilherme Russi wrote:
Hello Razique, I'm trying to run this command: nova live-migration f18b6d2b-f685-4476-a7c0-e735db5113cf 192.168.3.3
Where f18b6d2b-f685-4476-a7c0-e735db5113cf is the ID from my new VM and 192.168.3.3 is the destiny compute node.
I'm executing this command from my controller node and I have all IPs and machine names at my /etc/hosts, but I'm getting this error:
ERROR: Compute service of 192.168.3.3 is unavailable at this time. (HTTP 400) (Request-ID: req-117a0eea-14e8-4db2-a3d0-cbb14e438f85)
It's almost there (I guess), hope you all can help me :)
And Julie, thank you for saying that, now I'm only trying to migrate through the terminal.
Regards.
Guilherme.
2013/8/20 Razique Mahroua <razique.mahroua at gmail.com<mailto:razique.mahroua at gmail.com>>
It's getting weird really…
now can you run a live migration from the command line?
Le 20 août 2013 à 00:20, Guilherme Russi <luisguilherme.cr at gmail.com<mailto:luisguilherme.cr at gmail.com>> a écrit :
Yes:
nova-manage service list
Binary Host Zone Status State Updated_At
nova-cert hemera internal enabled :-) 2013-08-19 22:20:14
nova-scheduler hemera internal enabled :-) 2013-08-19 22:20:14
nova-consoleauth hemera internal enabled :-) 2013-08-19 22:20:17
nova-conductor hemera internal enabled :-) 2013-08-19 22:20:14
nova-compute caos nova enabled :-) 2013-08-19 22:20:17
nova-compute tiresias nova enabled :-) 2013-08-19 22:20:17
2013/8/19 Razique Mahroua <razique.mahroua at gmail.com<mailto:razique.mahroua at gmail.com>>
Can you run nova-manage service list ?
Razique Mahroua - Nuage & Co
razique.mahroua at gmail.com<mailto:razique.mahroua at gmail.com>
Tel : +33 9 72 37 94 15<tel:%2B33%209%2072%2037%2094%2015>
<NUAGECO-LOGO-Fblan_petit.jpg>
Le 19 août 2013 à 22:27, Guilherme Russi <luisguilherme.cr at gmail.com<mailto:luisguilherme.cr at gmail.com>> a écrit :
Hello again Razique, I was trying to migrate through the terminal and I'm getting this line from my controller node:
nova live-migration f18b6d2b-f685-4476-a7c0-e735db5113cf 192.168.3.3
ERROR: Compute service of 192.168.3.3 is unavailable at this time. (HTTP 400) (Request-ID: req-454960fb-6652-46ed-9080-6d23b384cbb3)
Haven't found anything yet. Any ideas?
Regards.
Guilherme.
2013/8/19 Guilherme Russi <luisguilherme.cr at gmail.com<mailto:luisguilherme.cr at gmail.com>>
I'm migrating it through the dashboard with the admin user. Is it wrong?
Thank you.
Guilherme.
2013/8/19 Razique Mahroua <razique.mahroua at gmail.com<mailto:razique.mahroua at gmail.com>>
quick question, how do you migrate the instance ? and does it belong to the tenant?
Le 19 août 2013 à 15:36, Guilherme Russi <luisguilherme.cr at gmail.com<mailto:luisguilherme.cr at gmail.com>> a écrit :
The VM looks to be there:
du -h instances/
4,0K instances/keys
1,9G instances/_base
0 instances/locks
310M instances/edca1460-9c97-475e-997b-3266b235e797
260M instances/0fe59b28-04a3-41bb-9926-b7d139a70548
323M instances/aeda1c56-72ba-40a7-9857-01a622540505
7,0M instances/f4af662d-ae55-4346-82cf-a997192706b5
264M instances/483f98e3-8ef5-43e2-8c3a-def55abdabcd
268M instances/eeb81e32-32ce-49fa-8249-afc572f26233
268M instances/c0224f0d-d7e8-4621-8c1a-af68740a3267
188M instances/72ec37a3-b209-4729-b628-005fdcea5a3c
12M instances/2c7c123b-456a-4c70-a991-80efee541324
11M instances/c9e1c5ed-a108-4196-bfbc-24495e2e71bd
6,5M instances/96d8f2de-a72d-40b5-b083-1f5e511f91ce
531M instances/1b35a4a6-e158-48d3-a804-44b05ec171d0
529M instances/1a2b6146-6ef0-4d76-a603-e4715b4dfb62
514M instances/f6cb9f81-9281-419f-9e20-adaa6a07cea3
2,7G instances/75e479cf-37a4-4511-b2a5-d4f3027702d4_resize
2,4G instances/75e479cf-37a4-4511-b2a5-d4f3027702d4
11G instances/
It's still running, no Error messagem from dashboard (it is showing Status: Resize/Migrate - Task: Resizing or Migrating - Power State: Running). The logs are not showing that permission error anymore. Should I wait to see if it will migrate or returns another error status? One thing is true, the machine is really slow now :P
Regards.
Guilherme.
2013/8/19 Razique Mahroua <razique.mahroua at gmail.com<mailto:razique.mahroua at gmail.com>>
Looks like the file doesn't exist.
Can you spawn a new instance and try to migrate it again?
thanks
Razique Mahroua - Nuage & Co
razique.mahroua at gmail.com<mailto:razique.mahroua at gmail.com>
Tel : +33 9 72 37 94 15<tel:%2B33%209%2072%2037%2094%2015>
<NUAGECO-LOGO-Fblan_petit.jpg>
Le 19 août 2013 à 13:41, Guilherme Russi <luisguilherme.cr at gmail.com<mailto:luisguilherme.cr at gmail.com>> a écrit :
Hello Peter,
Thank you for your help too :)
I've changed the permission and now I'm getting another error when I try to migrate:
cat /var/log/nova/nova-compute.log | grep -E ERROR
2013-08-19 08:17:19.990 12667<tel:19.990%2012667> ERROR nova.manager [-] Error during ComputeManager._run_image_cache_manager_pass: Unexpected error while running command.
2013-08-19 08:23:40.309 12667 ERROR nova.network.quantumv2 [-] _get_auth_token() failed
2013-08-19 08:24:27.013 12667 ERROR nova.virt.libvirt.driver [-] Getting disk size of instance-000000bc: [Errno 2] No such file or directory: '/var/lib/nova/instances/75e479cf-37a4-4511-b2a5-d4f3027702d4/disk'
cat /var/log/nova/nova-compute.log | grep -E TRACE
2013-08-19 08:17:19.990 12667<tel:19.990%2012667> TRACE nova.manager Traceback (most recent call last):
2013-08-19 08:17:19.990 12667<tel:19.990%2012667> TRACE nova.manager File "/usr/lib/python2.7/dist-packages/nova/manager.py", line 241, in periodic_tasks
2013-08-19 08:17:19.990 12667<tel:19.990%2012667> TRACE nova.manager task(self, context)
2013-08-19 08:17:19.990 12667<tel:19.990%2012667> TRACE nova.manager File "/usr/lib/python2.7/dist-packages/nova/compute/manager.py", line 4020, in _run_image_cache_manager_pass
2013-08-19 08:17:19.990 12667<tel:19.990%2012667> TRACE nova.manager self.driver.manage_image_cache(context, filtered_instances)
2013-08-19 08:17:19.990 12667<tel:19.990%2012667> TRACE nova.manager File "/usr/lib/python2.7/dist-packages/nova/virt/libvirt/driver.py", line 3401, in manage_image_cache
2013-08-19 08:17:19.990 12667<tel:19.990%2012667> TRACE nova.manager self.image_cache_manager.verify_base_images(context, all_instances)
2013-08-19 08:17:19.990 12667<tel:19.990%2012667> TRACE nova.manager File "/usr/lib/python2.7/dist-packages/nova/virt/libvirt/imagecache.py", line 588, in verify_base_images
2013-08-19 08:17:19.990 12667<tel:19.990%2012667> TRACE nova.manager self._handle_base_image(img, base_file)
2013-08-19 08:17:19.990 12667<tel:19.990%2012667> TRACE nova.manager File "/usr/lib/python2.7/dist-packages/nova/virt/libvirt/imagecache.py", line 553, in _handle_base_image
2013-08-19 08:17:19.990 12667<tel:19.990%2012667> TRACE nova.manager virtutils.chown(base_file, os.getuid())
2013-08-19 08:17:19.990 12667<tel:19.990%2012667> TRACE nova.manager File "/usr/lib/python2.7/dist-packages/nova/virt/libvirt/utils.py", line 449, in chown
2013-08-19 08:17:19.990 12667<tel:19.990%2012667> TRACE nova.manager execute('chown', owner, path, run_as_root=True)
2013-08-19 08:17:19.990 12667<tel:19.990%2012667> TRACE nova.manager File "/usr/lib/python2.7/dist-packages/nova/virt/libvirt/utils.py", line 47, in execute
2013-08-19 08:17:19.990 12667<tel:19.990%2012667> TRACE nova.manager return utils.execute(*args, **kwargs)
2013-08-19 08:17:19.990 12667<tel:19.990%2012667> TRACE nova.manager File "/usr/lib/python2.7/dist-packages/nova/utils.py", line 239, in execute
2013-08-19 08:17:19.990 12667<tel:19.990%2012667> TRACE nova.manager cmd=' '.join(cmd))
2013-08-19 08:17:19.990 12667<tel:19.990%2012667> TRACE nova.manager ProcessExecutionError: Unexpected error while running command.
2013-08-19 08:17:19.990 12667<tel:19.990%2012667> TRACE nova.manager Command: sudo nova-rootwrap /etc/nova/rootwrap.conf chown 123 /var/lib/nova/instances/_base/b0d90eb60f048f6de16c3efbc0eb4c16cbac77a6
2013-08-19 08:17:19.990 12667<tel:19.990%2012667> TRACE nova.manager Exit code: 1
2013-08-19 08:17:19.990 12667<tel:19.990%2012667> TRACE nova.manager Stdout: ''
2013-08-19 08:17:19.990 12667<tel:19.990%2012667> TRACE nova.manager Stderr: 'Traceback (most recent call last):\n File "/usr/bin/nova-rootwrap", line 108, in <module>\n os.getlogin(), pwd.getpwuid(os.getuid())[0],\nOSError: [Errno 22] Invalid argument\n'
2013-08-19 08:17:19.990 12667<tel:19.990%2012667> TRACE nova.manager
2013-08-19 08:23:40.309 12667 TRACE nova.network.quantumv2 Traceback (most recent call last):
2013-08-19 08:23:40.309 12667 TRACE nova.network.quantumv2 File "/usr/lib/python2.7/dist-packages/nova/network/quantumv2/__init__.py", line 40, in _get_auth_token
2013-08-19 08:23:40.309 12667 TRACE nova.network.quantumv2 httpclient.authenticate()
2013-08-19 08:23:40.309 12667 TRACE nova.network.quantumv2 File "/usr/lib/python2.7/dist-packages/quantumclient/client.py", line 198, in authenticate
2013-08-19 08:23:40.309 12667 TRACE nova.network.quantumv2 raise exceptions.Unauthorized(message=body)
2013-08-19 08:23:40.309 12667 TRACE nova.network.quantumv2 Unauthorized: Request Timeout
2013-08-19 08:23:40.309 12667 TRACE nova.network.quantumv2
Any ideas?
Regards.
Guilherme.
2013/8/19 Peter Funk <pf at artcom-gmbh.de<mailto:pf at artcom-gmbh.de>>
Guilherme Russi wrote 16.08.2013 um 09:53:
> Here is my directory configs:
>
> drwxrwxrwx 2 nova nova 76 Ago 15 14:54 .ssh
>
> I made chmod a+xwr at the folder at all computers, I made it because I
> couldn't find the reason I was no allowed to migrate, can it be like that?
For SSH to work the permissions of $HOME/.ssh should look like this:
drwx------ 2 nova nova 76 Ago 15 14:54 .ssh
Use:
chmod -R go-rwx .ssh
Regards, Peter Funk
--
Peter Funk, home: ✉Oldenburger Str.86, D-27777 Ganderkesee
mobile:+49-179-640-8878<tel:%2B49-179-640-8878> phone:+49-421-20419-0<tel:%2B49-421-20419-0> <http://www.artcom-gmbh.de/>
office: ArtCom GmbH, ✉Haferwende 2, D-28357 Bremen, Germany
_______________________________________________
Mailing list: http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack
Post to : openstack at lists.openstack.org<mailto:openstack at lists.openstack.org>
Unsubscribe : http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack
_______________________________________________
Mailing list: http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack
Post to : openstack at lists.openstack.org<mailto:openstack at lists.openstack.org>
Unsubscribe : http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack
_______________________________________________
Mailing list: http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack
Post to : openstack at lists.openstack.org<mailto:openstack at lists.openstack.org>
Unsubscribe : http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.openstack.org/pipermail/openstack/attachments/20130820/b885be02/attachment.html>
More information about the Openstack
mailing list