[Openstack] FW: FW: FW: nova-compute goes to XXX but is alive
raghavendra.lad at accenture.com
raghavendra.lad at accenture.com
Thu Dec 4 05:41:28 UTC 2014
Hi Guillermo,
The nova database if you have tried multiple times and made changes it would not sync and hence the service shows down.
I have tested myself. However when you drop nova database, recreate, restart services and then sync it worked for me with : ) ‘s.
Regards,
Raghavendra Lad
From: Guillermo Alvarado [mailto:guillermoalvarado89 at gmail.com]
Sent: Thursday, December 04, 2014 11:06 AM
To: Lad, Raghavendra
Cc: openstack at lists.openstack.org
Subject: Re: FW: FW: [Openstack] nova-compute goes to XXX but is alive
The XXX were to not put my real passwords, I can see :-) each I run nova-restart, even I have VMs created in these compute nodes, but for some reason after 30 minutes :-) leave and arrive XXX always happens the same....
Why a db sync will help? Sorry for being incredulous....
2014-12-03 23:29 GMT-06:00 <raghavendra.lad at accenture.com<mailto:raghavendra.lad at accenture.com>>:
Hi Guillermo,
The steps provided should work. You can type rabbit user, password and rabbit port in your controller configuration.
If the nova database gives the xxx again then you may have to drop the nova database and create again.
After which run the db sync and restart the nova services that should do the trick and you can see ☺ : )
Please wait for some time and check it should all be going great..!
Regards,
Raghavendra Lad
From: Guillermo Alvarado [mailto:guillermoalvarado89 at gmail.com<mailto:guillermoalvarado89 at gmail.com>]
Sent: Thursday, December 04, 2014 10:54 AM
To: Lad, Raghavendra
Cc: openstack at lists.openstack.org<mailto:openstack at lists.openstack.org>
Subject: Re: FW: [Openstack] nova-compute goes to XXX but is alive
Hi, thanks for your answer. This installations was running two weeks and today start the problems, service list says :-) but after 30 mins says XXX
Do you have another idea?
Thanks
2014-12-03 23:11 GMT-06:00 <raghavendra.lad at accenture.com<mailto:raghavendra.lad at accenture.com>>:
Hi Guillermo,
Please check your controller and compute nova.conf files also make sure you have run the nova-manage db sync on controller and then restart all services. You can try running the service nova-compute restart on compute node and check.
The nova service-list should work fine.
You may remove the vif
vif_plugging_is_fatal=false
vif_plugging_timeout=0
in the controller nova.conf file.
Cheers,
Raghavendra Lad
From: Guillermo Alvarado [mailto:guillermoalvarado89 at gmail.com<mailto:guillermoalvarado89 at gmail.com>]
Sent: Thursday, December 04, 2014 10:07 AM
To: openstack at lists.openstack.org<mailto:openstack at lists.openstack.org>
Subject: [Openstack] nova-compute goes to XXX but is alive
Hello there,
I am having a lot of problems with nova compute, after 30 minutes all my computes report XXX when I execute nova-manage service list
So I restart rabbitmq and conductor in the controller node, libvrit-bin and nova-compute in all the compute hosts and everithing return OK but 30 minutes after all the computes goes XXX again.
I noticed that nova-compute, libvirt-bin are ok in thoose hosts and the rabbitmq server is alive but I still have XXX witth nova-manage servce list.
I am using ubuntu 14.04 and Icehouse release. I noticed too this in the rabbitmq log:
=WARNING REPORT==== 3-Dec-2014::21:58:51 ===
closing AMQP connection <0.17955.6> (172.18.0.13:40959<http://172.18.0.13:40959> -> 172.18.0.10:5672<http://172.18.0.10:5672>):
connection_closed_abruptly
=WARNING REPORT==== 3-Dec-2014::21:58:51 ===
closing AMQP connection <0.19007.6> (172.18.0.13:40961<http://172.18.0.13:40961> -> 172.18.0.10:5672<http://172.18.0.10:5672>):
connection_closed_abruptly
=INFO REPORT==== 3-Dec-2014::22:04:02 ===
accepting AMQP connection <0.19867.7> (172.18.0.15:53656<http://172.18.0.15:53656> -> 172.18.0.10:5672<http://172.18.0.10:5672>)
=INFO REPORT==== 3-Dec-2014::22:06:31 ===
accepting AMQP connection <0.23039.7> (172.18.0.13:41162<http://172.18.0.13:41162> -> 172.18.0.10:5672<http://172.18.0.10:5672>)
=INFO REPORT==== 3-Dec-2014::22:06:31 ===
accepting AMQP connection <0.23049.7> (172.18.0.13:41163<http://172.18.0.13:41163> -> 172.18.0.10:5672<http://172.18.0.10:5672>)
=INFO REPORT==== 3-Dec-2014::22:06:32 ===
closing AMQP connection <0.23049.7> (172.18.0.13:41163<http://172.18.0.13:41163> -> 172.18.0.10:5672<http://172.18.0.10:5672>)
=WARNING REPORT==== 3-Dec-2014::22:06:32 ===
closing AMQP connection <0.23039.7> (172.18.0.13:41162<http://172.18.0.13:41162> -> 172.18.0.10:5672<http://172.18.0.10:5672>):
connection_closed_abruptly
=INFO REPORT==== 3-Dec-2014::22:11:14 ===
accepting AMQP connection <0.30069.7> (172.18.0.13:41177<http://172.18.0.13:41177> -> 172.18.0.10:5672<http://172.18.0.10:5672>)
=INFO REPORT==== 3-Dec-2014::22:11:14 ===
accepting AMQP connection <0.30079.7> (172.18.0.13:41178<http://172.18.0.13:41178> -> 172.18.0.10:5672<http://172.18.0.10:5672>)
=INFO REPORT==== 3-Dec-2014::22:11:20 ===
accepting AMQP connection <0.30236.7> (172.18.0.13:41179<http://172.18.0.13:41179> -> 172.18.0.10:5672<http://172.18.0.10:5672>)
=INFO REPORT==== 3-Dec-2014::22:19:19 ===
accepting AMQP connection <0.7975.8> (172.18.0.10:39881<http://172.18.0.10:39881> -> 172.18.0.10:5672<http://172.18.0.10:5672>)
=INFO REPORT==== 3-Dec-2014::22:19:20 ===
accepting AMQP connection <0.8132.8> (172.18.0.10:39893<http://172.18.0.10:39893> -> 172.18.0.10:5672<http://172.18.0.10:5672>)
=INFO REPORT==== 3-Dec-2014::22:19:22 ===
accepting AMQP connection <0.8342.8> (172.18.0.13:41275<http://172.18.0.13:41275> -> 172.18.0.10:5672<http://172.18.0.10:5672>)
=INFO REPORT==== 3-Dec-2014::22:19:48 ===
accepting AMQP connection <0.9155.8> (172.18.0.10:39976<http://172.18.0.10:39976> -> 172.18.0.10:5672<http://172.18.0.10:5672>)
=INFO REPORT==== 3-Dec-2014::22:19:48 ===
accepting AMQP connection <0.9165.8> (172.18.0.10:39977<http://172.18.0.10:39977> -> 172.18.0.10:5672<http://172.18.0.10:5672>)
=WARNING REPORT==== 3-Dec-2014::22:19:56 ===
closing AMQP connection <0.9155.8> (172.18.0.10:39976<http://172.18.0.10:39976> -> 172.18.0.10:5672<http://172.18.0.10:5672>):
connection_closed_abruptly
=WARNING REPORT==== 3-Dec-2014::22:19:56 ===
closing AMQP connection <0.9165.8> (172.18.0.10:39977<http://172.18.0.10:39977> -> 172.18.0.10:5672<http://172.18.0.10:5672>):
connection_closed_abruptly
=INFO REPORT==== 3-Dec-2014::22:21:18 ===
accepting AMQP connection <0.10063.8> (172.18.0.10:40151<http://172.18.0.10:40151> -> 172.18.0.10:5672<http://172.18.0.10:5672>)
=INFO REPORT==== 3-Dec-2014::22:21:18 ===
accepting AMQP connection <0.10073.8> (172.18.0.10:40152<http://172.18.0.10:40152> -> 172.18.0.10:5672<http://172.18.0.10:5672>)
=WARNING REPORT==== 3-Dec-2014::22:24:20 ===
closing AMQP connection <0.10073.8> (172.18.0.10:40152<http://172.18.0.10:40152> -> 172.18.0.10:5672<http://172.18.0.10:5672>):
connection_closed_abruptly
=WARNING REPORT==== 3-Dec-2014::22:24:20 ===
closing AMQP connection <0.10063.8> (172.18.0.10:40151<http://172.18.0.10:40151> -> 172.18.0.10:5672<http://172.18.0.10:5672>):
connection_closed_abruptly
172.18.0.10 is the IP of my controller node
and this in the libvirt log:
2014-12-03 16:33:48.121+0000: 53769: error : virNetSocketReadWire:1454 : End of file while reading data: Input/output error
2014-12-03 17:27:14.557+0000: 53769: error : virNetSocketReadWire:1454 : End of file while reading data: Input/output error
2014-12-03 20:16:46.270+0000: 53769: error : virNetSocketReadWire:1454 : End of file while reading data: Input/output error
2014-12-03 21:22:18.255+0000: 53769: error : virNetSocketReadWire:1454 : End of file while reading data: Input/output error
2014-12-03 21:53:07.884+0000: 53769: error : virNetSocketReadWire:1454 : End of file while reading data: Input/output error
2014-12-03 23:10:49.062+0000: 53769: error : virNetSocketReadWire:1454 : End of file while reading data: Input/output error
2014-12-04 00:22:13.703+0000: 53769: error : virNetSocketReadWire:1454 : End of file while reading data: Input/output error
2014-12-04 00:39:57.967+0000: 53769: error : virNetSocketReadWire:1454 : End of file
This is my nova-conf:
[DEFAULT]
dhcpbridge_flagfile=/etc/nova/nova.conf
dhcpbridge=/usr/bin/nova-dhcpbridge
logdir=/var/log/nova
state_path=/var/lib/nova
lock_path=/var/lock/nova
force_dhcp_release=True
iscsi_helper=tgtadm
libvirt_use_virtio_for_bridges=True
connection_type=libvirt
root_helper=sudo nova-rootwrap /etc/nova/rootwrap.conf
verbose=True
ec2_private_dns_show_ip=True
api_paste_config=/etc/nova/api-paste.ini
volumes_path=/var/lib/nova/volumes
enabled_apis=ec2,osapi_compute,metadata
auth_strategy = keystone
rpc_backend = rabbit
rabbit_host = 172.18.0.10
my_ip = 172.18.0.23
vnc_enabled = True
vncserver_listen = 0.0.0.0
vncserver_proxyclient_address = 172.18.0.16
novncproxy_base_url = http://PUBLIC_IP:6080/vnc_auto.html
glance_host = 172.18.0.10
vif_plugging_is_fatal=false
vif_plugging_timeout=0
network_api_class = nova.network.neutronv2.api.API
neutron_url = http://172.18.0.10:9696
neutron_auth_strategy = keystone
neutron_admin_tenant_name = service
neutron_admin_username = neutron
neutron_admin_password = XXX
neutron_admin_auth_url = http://172.18.0.10:35357/v2.0
linuxnet_interface_driver = nova.network.linux_net.LinuxOVSInterfaceDriver
firewall_driver = nova.virt.firewall.NoopFirewallDriver
security_group_api = neutron
[database]
connection = mysql://nova:XXX@172.18.0.10/nova<http://nova:XXX@172.18.0.10/nova>
[keystone_authtoken]
auth_uri = http://172.18.0.10:5000
auth_host = 172.18.0.10
auth_port = 35357
auth_protocol = http
admin_tenant_name = service
admin_user = nova
admin_password = XXX
Any help would be appreciated.
Thanks in advance!
~GA
________________________________
This message is for the designated recipient only and may contain privileged, proprietary, or otherwise confidential information. If you have received it in error, please notify the sender immediately and delete the original. Any other use of the e-mail by you is prohibited. Where allowed by local law, electronic communications with Accenture and its affiliates, including e-mail and instant messaging (including content), may be scanned by our systems for the purposes of information security and assessment of internal compliance with Accenture policy.
______________________________________________________________________________________
www.accenture.com<http://www.accenture.com>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.openstack.org/pipermail/openstack/attachments/20141204/0aa4f2cc/attachment.html>
More information about the Openstack
mailing list