[Openstack] SSH permissions (was: Re: Error during live migration)

Joe Breu joseph.breu at rackspace.com
Tue Aug 20 15:33:22 UTC 2013


You need to use the hostname of the compute node, not the IP address for migration.  In your case this would be either caos or tiersias


On Aug 20, 2013, at 8:06 AM, Guilherme Russi wrote:

Hello Razique, I'm trying to run this command: nova live-migration f18b6d2b-f685-4476-a7c0-e735db5113cf 192.168.3.3

Where f18b6d2b-f685-4476-a7c0-e735db5113cf is the ID from my new VM and 192.168.3.3 is the destiny compute node.
I'm executing this command from my controller node and I have all IPs and machine names at my /etc/hosts, but I'm getting this error:

ERROR: Compute service of 192.168.3.3 is unavailable at this time. (HTTP 400) (Request-ID: req-117a0eea-14e8-4db2-a3d0-cbb14e438f85)

It's almost there (I guess), hope you all can help me :)

And Julie, thank you for saying that, now I'm only trying to migrate through the terminal.

Regards.

Guilherme.


2013/8/20 Razique Mahroua <razique.mahroua at gmail.com<mailto:razique.mahroua at gmail.com>>
It's getting weird really…
now can you run a live migration from the command line?

Le 20 août 2013 à 00:20, Guilherme Russi <luisguilherme.cr at gmail.com<mailto:luisguilherme.cr at gmail.com>> a écrit :

Yes:

nova-manage service list
Binary           Host                                 Zone             Status     State Updated_At
nova-cert        hemera                               internal         enabled    :-)   2013-08-19 22:20:14
nova-scheduler   hemera                               internal         enabled    :-)   2013-08-19 22:20:14
nova-consoleauth hemera                               internal         enabled    :-)   2013-08-19 22:20:17
nova-conductor   hemera                               internal         enabled    :-)   2013-08-19 22:20:14
nova-compute     caos                                 nova             enabled    :-)   2013-08-19 22:20:17
nova-compute     tiresias                             nova             enabled    :-)   2013-08-19 22:20:17



2013/8/19 Razique Mahroua <razique.mahroua at gmail.com<mailto:razique.mahroua at gmail.com>>
Can you run nova-manage service list ?

Razique Mahroua - Nuage & Co
razique.mahroua at gmail.com<mailto:razique.mahroua at gmail.com>
Tel : +33 9 72 37 94 15<tel:%2B33%209%2072%2037%2094%2015>

<NUAGECO-LOGO-Fblan_petit.jpg>

Le 19 août 2013 à 22:27, Guilherme Russi <luisguilherme.cr at gmail.com<mailto:luisguilherme.cr at gmail.com>> a écrit :

Hello again Razique, I was trying to migrate through the terminal and I'm getting this line from my controller node:

nova live-migration f18b6d2b-f685-4476-a7c0-e735db5113cf 192.168.3.3
ERROR: Compute service of 192.168.3.3 is unavailable at this time. (HTTP 400) (Request-ID: req-454960fb-6652-46ed-9080-6d23b384cbb3)

Haven't found anything yet. Any ideas?

Regards.

Guilherme.


2013/8/19 Guilherme Russi <luisguilherme.cr at gmail.com<mailto:luisguilherme.cr at gmail.com>>
I'm migrating it through the dashboard with the admin user. Is it wrong?

Thank you.

Guilherme.


2013/8/19 Razique Mahroua <razique.mahroua at gmail.com<mailto:razique.mahroua at gmail.com>>
quick question, how do you migrate the instance ? and does it belong to the tenant?


Le 19 août 2013 à 15:36, Guilherme Russi <luisguilherme.cr at gmail.com<mailto:luisguilherme.cr at gmail.com>> a écrit :

The VM looks to be there:

du -h instances/
4,0K instances/keys
1,9G instances/_base
0 instances/locks
310M instances/edca1460-9c97-475e-997b-3266b235e797
260M instances/0fe59b28-04a3-41bb-9926-b7d139a70548
323M instances/aeda1c56-72ba-40a7-9857-01a622540505
7,0M instances/f4af662d-ae55-4346-82cf-a997192706b5
264M instances/483f98e3-8ef5-43e2-8c3a-def55abdabcd
268M instances/eeb81e32-32ce-49fa-8249-afc572f26233
268M instances/c0224f0d-d7e8-4621-8c1a-af68740a3267
188M instances/72ec37a3-b209-4729-b628-005fdcea5a3c
12M instances/2c7c123b-456a-4c70-a991-80efee541324
11M instances/c9e1c5ed-a108-4196-bfbc-24495e2e71bd
6,5M instances/96d8f2de-a72d-40b5-b083-1f5e511f91ce
531M instances/1b35a4a6-e158-48d3-a804-44b05ec171d0
529M instances/1a2b6146-6ef0-4d76-a603-e4715b4dfb62
514M instances/f6cb9f81-9281-419f-9e20-adaa6a07cea3
2,7G instances/75e479cf-37a4-4511-b2a5-d4f3027702d4_resize
2,4G instances/75e479cf-37a4-4511-b2a5-d4f3027702d4
11G instances/


It's still running, no Error messagem from dashboard (it is showing Status: Resize/Migrate - Task: Resizing or Migrating - Power State: Running). The logs are not showing that permission error anymore. Should I wait to see if it will migrate or returns another error status? One thing is true, the machine is really slow now :P

Regards.

Guilherme.


2013/8/19 Razique Mahroua <razique.mahroua at gmail.com<mailto:razique.mahroua at gmail.com>>
Looks like the file doesn't exist.
Can you spawn a new instance and try to migrate it again?
thanks

Razique Mahroua - Nuage & Co
razique.mahroua at gmail.com<mailto:razique.mahroua at gmail.com>
Tel : +33 9 72 37 94 15<tel:%2B33%209%2072%2037%2094%2015>

<NUAGECO-LOGO-Fblan_petit.jpg>

Le 19 août 2013 à 13:41, Guilherme Russi <luisguilherme.cr at gmail.com<mailto:luisguilherme.cr at gmail.com>> a écrit :

Hello Peter,

 Thank you for your help too :)
 I've changed the permission and now I'm getting another error when I try to migrate:

cat /var/log/nova/nova-compute.log | grep -E ERROR
2013-08-19 08:17:19.990 12667<tel:19.990%2012667> ERROR nova.manager [-] Error during ComputeManager._run_image_cache_manager_pass: Unexpected error while running command.
2013-08-19 08:23:40.309 12667 ERROR nova.network.quantumv2 [-] _get_auth_token() failed
2013-08-19 08:24:27.013 12667 ERROR nova.virt.libvirt.driver [-] Getting disk size of instance-000000bc: [Errno 2] No such file or directory: '/var/lib/nova/instances/75e479cf-37a4-4511-b2a5-d4f3027702d4/disk'


cat /var/log/nova/nova-compute.log | grep -E TRACE
2013-08-19 08:17:19.990 12667<tel:19.990%2012667> TRACE nova.manager Traceback (most recent call last):
2013-08-19 08:17:19.990 12667<tel:19.990%2012667> TRACE nova.manager   File "/usr/lib/python2.7/dist-packages/nova/manager.py", line 241, in periodic_tasks
2013-08-19 08:17:19.990 12667<tel:19.990%2012667> TRACE nova.manager     task(self, context)
2013-08-19 08:17:19.990 12667<tel:19.990%2012667> TRACE nova.manager   File "/usr/lib/python2.7/dist-packages/nova/compute/manager.py", line 4020, in _run_image_cache_manager_pass
2013-08-19 08:17:19.990 12667<tel:19.990%2012667> TRACE nova.manager     self.driver.manage_image_cache(context, filtered_instances)
2013-08-19 08:17:19.990 12667<tel:19.990%2012667> TRACE nova.manager   File "/usr/lib/python2.7/dist-packages/nova/virt/libvirt/driver.py", line 3401, in manage_image_cache
2013-08-19 08:17:19.990 12667<tel:19.990%2012667> TRACE nova.manager     self.image_cache_manager.verify_base_images(context, all_instances)
2013-08-19 08:17:19.990 12667<tel:19.990%2012667> TRACE nova.manager   File "/usr/lib/python2.7/dist-packages/nova/virt/libvirt/imagecache.py", line 588, in verify_base_images
2013-08-19 08:17:19.990 12667<tel:19.990%2012667> TRACE nova.manager     self._handle_base_image(img, base_file)
2013-08-19 08:17:19.990 12667<tel:19.990%2012667> TRACE nova.manager   File "/usr/lib/python2.7/dist-packages/nova/virt/libvirt/imagecache.py", line 553, in _handle_base_image
2013-08-19 08:17:19.990 12667<tel:19.990%2012667> TRACE nova.manager     virtutils.chown(base_file, os.getuid())
2013-08-19 08:17:19.990 12667<tel:19.990%2012667> TRACE nova.manager   File "/usr/lib/python2.7/dist-packages/nova/virt/libvirt/utils.py", line 449, in chown
2013-08-19 08:17:19.990 12667<tel:19.990%2012667> TRACE nova.manager     execute('chown', owner, path, run_as_root=True)
2013-08-19 08:17:19.990 12667<tel:19.990%2012667> TRACE nova.manager   File "/usr/lib/python2.7/dist-packages/nova/virt/libvirt/utils.py", line 47, in execute
2013-08-19 08:17:19.990 12667<tel:19.990%2012667> TRACE nova.manager     return utils.execute(*args, **kwargs)
2013-08-19 08:17:19.990 12667<tel:19.990%2012667> TRACE nova.manager   File "/usr/lib/python2.7/dist-packages/nova/utils.py", line 239, in execute
2013-08-19 08:17:19.990 12667<tel:19.990%2012667> TRACE nova.manager     cmd=' '.join(cmd))
2013-08-19 08:17:19.990 12667<tel:19.990%2012667> TRACE nova.manager ProcessExecutionError: Unexpected error while running command.
2013-08-19 08:17:19.990 12667<tel:19.990%2012667> TRACE nova.manager Command: sudo nova-rootwrap /etc/nova/rootwrap.conf chown 123 /var/lib/nova/instances/_base/b0d90eb60f048f6de16c3efbc0eb4c16cbac77a6
2013-08-19 08:17:19.990 12667<tel:19.990%2012667> TRACE nova.manager Exit code: 1
2013-08-19 08:17:19.990 12667<tel:19.990%2012667> TRACE nova.manager Stdout: ''
2013-08-19 08:17:19.990 12667<tel:19.990%2012667> TRACE nova.manager Stderr: 'Traceback (most recent call last):\n  File "/usr/bin/nova-rootwrap", line 108, in <module>\n    os.getlogin(), pwd.getpwuid(os.getuid())[0],\nOSError: [Errno 22] Invalid argument\n'
2013-08-19 08:17:19.990 12667<tel:19.990%2012667> TRACE nova.manager
2013-08-19 08:23:40.309 12667 TRACE nova.network.quantumv2 Traceback (most recent call last):
2013-08-19 08:23:40.309 12667 TRACE nova.network.quantumv2   File "/usr/lib/python2.7/dist-packages/nova/network/quantumv2/__init__.py", line 40, in _get_auth_token
2013-08-19 08:23:40.309 12667 TRACE nova.network.quantumv2     httpclient.authenticate()
2013-08-19 08:23:40.309 12667 TRACE nova.network.quantumv2   File "/usr/lib/python2.7/dist-packages/quantumclient/client.py", line 198, in authenticate
2013-08-19 08:23:40.309 12667 TRACE nova.network.quantumv2     raise exceptions.Unauthorized(message=body)
2013-08-19 08:23:40.309 12667 TRACE nova.network.quantumv2 Unauthorized: Request Timeout
2013-08-19 08:23:40.309 12667 TRACE nova.network.quantumv2

Any ideas?

Regards.

Guilherme.


2013/8/19 Peter Funk <pf at artcom-gmbh.de<mailto:pf at artcom-gmbh.de>>
Guilherme Russi wrote 16.08.2013 um 09:53:
>  Here is my directory configs:
>
> drwxrwxrwx  2 nova nova     76 Ago 15 14:54 .ssh
>
> I made chmod a+xwr at the folder at all computers, I made it because I
> couldn't find the reason I was no allowed to migrate, can it be like that?

For SSH to work the permissions of $HOME/.ssh should look like this:
  drwx------  2 nova nova     76 Ago 15 14:54 .ssh

Use:
  chmod -R go-rwx .ssh

Regards, Peter Funk
--
Peter Funk, home: ✉Oldenburger Str.86, D-27777 Ganderkesee
mobile:+49-179-640-8878<tel:%2B49-179-640-8878> phone:+49-421-20419-0<tel:%2B49-421-20419-0> <http://www.artcom-gmbh.de/>
office: ArtCom GmbH, ✉Haferwende 2, D-28357 Bremen, Germany

_______________________________________________
Mailing list: http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack
Post to     : openstack at lists.openstack.org<mailto:openstack at lists.openstack.org>
Unsubscribe : http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack

_______________________________________________
Mailing list: http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack
Post to     : openstack at lists.openstack.org<mailto:openstack at lists.openstack.org>
Unsubscribe : http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack









_______________________________________________
Mailing list: http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack
Post to     : openstack at lists.openstack.org<mailto:openstack at lists.openstack.org>
Unsubscribe : http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack

-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.openstack.org/pipermail/openstack/attachments/20130820/b885be02/attachment.html>


More information about the Openstack mailing list