[Openstack] Cant ping private or floating IP

Chathura M. Sarathchandra Magurawalage 77.chathura at gmail.com
Sat Feb 16 21:45:10 UTC 2013


Thanks Ransy,

I have created a tar file with the configuration and log files in it.
Please download it using the following URL. I have pasted the output of the
commands below.

https://www.dropbox.com/s/qyfcsn50060y304/confilesnlogs.tar

*Controller node:*
*root at controller:~# keystone endpoint-list*
+----------------------------------+-----------+-------------------------------------------------+-------------------------------------------------+--------------------------------------------+Controller
node
|                id                |   region  |
 publicurl                    |                   internalurl
    |                  adminurl                  |
+----------------------------------+-----------+-------------------------------------------------+-------------------------------------------------+--------------------------------------------+
| 2c9a1cb0fe8247d9b7716432cf459fe5 | RegionOne |
http://192.168.2.225:8774/v2/$(tenant_id)s   |
http://192.168.2.225:8774/v2/$(tenant_id)s   |
http://192.168.2.225:8774/v2/$(tenant_id)s |
| 2d306903ed3342a8aaaac7c5680c116f | RegionOne |
http://192.168.2.225:9696/           |
http://192.168.2.225:9696/          |
http://192.168.2.225:9696/         |
| 3848114f120f42bf819bc2443b28ac9e | RegionOne |
http://192.168.2.225:8080/v1/AUTH_$(tenant_id)s |
http://192.168.2.225:8080/v1/AUTH_$(tenant_id)s |
http://192.168.2.225:8080/v1        |
| 4955173b8d9e4d33ae4a5b29dc12c74d | RegionOne |
http://192.168.2.225:8776/v1/$(tenant_id)s   |
http://192.168.2.225:8776/v1/$(tenant_id)s   |
http://192.168.2.225:8776/v1/$(tenant_id)s |
| d313aa76bf854dde94f33a49a9f0c8ac | RegionOne |
http://192.168.2.225:9292/v2          |
http://192.168.2.225:9292/v2          |
http://192.168.2.225:9292/v2       |
| e5aa4ecf3cbe4dd5aba9b204c74fee6a | RegionOne |
http://192.168.2.225:5000/v2.0         |
http://192.168.2.225:5000/v2.0         |
http://192.168.2.225:35357/v2.0       |
| fba6f790e3b444c890d114f13cd32b37 | RegionOne |
http://192.168.2.225:8773/services/Cloud    |
http://192.168.2.225:8773/services/Cloud    |
http://192.168.2.225:8773/services/Admin  |
+----------------------------------+-----------+-------------------------------------------------+-------------------------------------------------+--------------------------------------------+

*root at controller:~# ip link show*
1: lo: <LOOPBACK,UP,LOWER_UP> mtu 16436 qdisc noqueue state UNKNOWN
    link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00
2: eth0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc mq state UP qlen
1000
    link/ether d4:ae:52:bb:aa:20 brd ff:ff:ff:ff:ff:ff
3: eth1: <BROADCAST,MULTICAST> mtu 1500 qdisc noop state DOWN qlen 1000
    link/ether d4:ae:52:bb:aa:21 brd ff:ff:ff:ff:ff:ff
4: eth0.2 at eth0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc noqueue
state UP
    link/ether d4:ae:52:bb:aa:20 brd ff:ff:ff:ff:ff:ff
5: br-int: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc noqueue state
UNKNOWN
    link/ether ba:7a:e9:dc:2b:41 brd ff:ff:ff:ff:ff:ff
7: br-ex: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc noqueue state
UNKNOWN
    link/ether 9a:41:c8:8a:9e:49 brd ff:ff:ff:ff:ff:ff
8: tapf71b5b86-5c: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc noqueue
state UNKNOWN
    link/ether 2a:44:a3:d1:7d:f3 brd ff:ff:ff:ff:ff:ff
9: qr-4d088f3a-78: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc noqueue
state UNKNOWN
    link/ether ca:5b:8d:4d:6d:fb brd ff:ff:ff:ff:ff:ff
10: qg-6f8374cb-cb: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc
noqueue state UNKNOWN
    link/ether 0e:7f:dd:3a:80:bc brd ff:ff:ff:ff:ff:ff
27: br-tun: <BROADCAST,MULTICAST> mtu 1500 qdisc noop state DOWN
    link/ether 8a:cf:ec:7c:15:40 brd ff:ff:ff:ff:ff:ff

*root at controller:~# ip route show*
default via 192.168.2.253 dev eth0.2
default via 192.168.2.253 dev eth0.2  metric 100
10.5.5.0/24 dev tapf71b5b86-5c  proto kernel  scope link  src 10.5.5.2
10.5.5.0/24 dev qr-4d088f3a-78  proto kernel  scope link  src 10.5.5.1
10.10.10.0/24 dev eth0  proto kernel  scope link  src 10.10.10.1
192.168.2.0/24 dev eth0.2  proto kernel  scope link  src 192.168.2.225
192.168.2.0/24 dev qg-6f8374cb-cb  proto kernel  scope link  src
192.168.2.151
192.168.2.0/24 dev br-ex  proto kernel  scope link  src 192.168.2.225

*$ ip netns show (Did not return anything)*

*root at controller:~# ovs-vsctl show*
a566afae-d7a8-42a9-aefe-8b0f2f7054a3
    Bridge br-tun
        Port "gre-4"
            Interface "gre-4"
                type: gre
                options: {in_key=flow, out_key=flow,
remote_ip="10.10.10.12"}
        Port "gre-3"
            Interface "gre-3"
                type: gre
                options: {in_key=flow, out_key=flow, remote_ip="127.0.0.1"}
        Port patch-int
            Interface patch-int
                type: patch
                options: {peer=patch-tun}
        Port br-tun
            Interface br-tun
                type: internal
        Port "gre-1"
            Interface "gre-1"
                type: gre
                options: {in_key=flow, out_key=flow, remote_ip="10.0.0.3"}
    Bridge br-ex
        Port br-ex
            Interface br-ex
                type: internal
        Port "qg-6f8374cb-cb"
            Interface "qg-6f8374cb-cb"
                type: internal
        Port "br0"
            Interface "br0"
    Bridge br-int
        Port br-int
            Interface br-int
                type: internal
        Port "tapf71b5b86-5c"
            tag: 1
            Interface "tapf71b5b86-5c"
                type: internal
        Port patch-tun
            Interface patch-tun
                type: patch
                options: {peer=patch-int}
        Port "qr-4d088f3a-78"
            tag: 1
            Interface "qr-4d088f3a-78"
                type: internal
    ovs_version: "1.4.0+build0"


*Compute node:*

*root at cronus:~# ip link show*
1: lo: <LOOPBACK,UP,LOWER_UP> mtu 16436 qdisc noqueue state UNKNOWN
    link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00
2: eth0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc mq state UP qlen
1000
    link/ether d4:ae:52:bb:a1:9d brd ff:ff:ff:ff:ff:ff
3: eth1: <BROADCAST,MULTICAST> mtu 1500 qdisc noop state DOWN qlen 1000
    link/ether d4:ae:52:bb:a1:9e brd ff:ff:ff:ff:ff:ff
4: eth0.2 at eth0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc noqueue
state UP
    link/ether d4:ae:52:bb:a1:9d brd ff:ff:ff:ff:ff:ff
5: br-int: <BROADCAST,MULTICAST> mtu 1500 qdisc noop state DOWN
    link/ether ae:9b:43:09:af:40 brd ff:ff:ff:ff:ff:ff
9: qbr256f5ed2-43: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc noqueue
state UP
    link/ether c6:c0:df:64:c6:99 brd ff:ff:ff:ff:ff:ff
10: qvo256f5ed2-43: <BROADCAST,MULTICAST,PROMISC,UP,LOWER_UP> mtu 1500
qdisc pfifo_fast state UP qlen 1000
    link/ether 76:25:8b:fd:90:3b brd ff:ff:ff:ff:ff:ff
11: qvb256f5ed2-43: <BROADCAST,MULTICAST,PROMISC,UP,LOWER_UP> mtu 1500
qdisc pfifo_fast master qbr256f5ed2-43 state UP qlen 1000
    link/ether c6:c0:df:64:c6:99 brd ff:ff:ff:ff:ff:ff
13: br-tun: <BROADCAST,MULTICAST> mtu 1500 qdisc noop state DOWN
    link/ether be:8c:30:78:35:48 brd ff:ff:ff:ff:ff:ff
15: vnet0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc pfifo_fast
master qbr256f5ed2-43 state UNKNOWN qlen 500
    link/ether fe:16:3e:57:ec:ff brd ff:ff:ff:ff:ff:ff

*root at cronus:~# ip route show*
default via 192.168.2.253 dev eth0.2  metric 100
10.10.10.0/24 dev eth0  proto kernel  scope link  src 10.10.10.12
192.168.2.0/24 dev eth0.2  proto kernel  scope link  src 192.168.2.234

*root at cronus:~# ovs-vsctl show*
d85bc334-6d64-4a13-b851-d56b18ff1549
    Bridge br-int
        Port "qvo0e743b01-89"
            tag: 4095
            Interface "qvo0e743b01-89"
        Port "qvo256f5ed2-43"
            tag: 1
            Interface "qvo256f5ed2-43"
        Port patch-tun
            Interface patch-tun
                type: patch
                options: {peer=patch-int}
        Port br-int
            Interface br-int
                type: internal
        Port "qvoee3d4131-2a"
            tag: 4095
            Interface "qvoee3d4131-2a"
        Port "qvocbc816bd-3d"
            tag: 4095
            Interface "qvocbc816bd-3d"
    Bridge br-tun
        Port br-tun
            Interface br-tun
                type: internal
        Port "gre-2"
            Interface "gre-2"
                type: gre
                options: {in_key=flow, out_key=flow, remote_ip="10.10.10.1"}
        Port "gre-1"
            Interface "gre-1"
                type: gre
                options: {in_key=flow, out_key=flow, remote_ip="10.0.0.3"}
        Port patch-int
            Interface patch-int
                type: patch
                options: {peer=patch-tun}
        Port "gre-3"
            Interface "gre-3"
                type: gre
                options: {in_key=flow, out_key=flow, remote_ip="127.0.0.1"}
    ovs_version: "1.4.0+build0"


Thanks I appreciate your help.

On 16 February 2013 16:49, Jean-Baptiste RANSY <
jean-baptiste.ransy at alyseo.com> wrote:

>  Please provide files listed bellow :
>
> Controller Node :
> /etc/nova/nova.conf
> /etc/nova/api-paste.ini
> /etc/quantum/l3_agent.ini
> /etc/quantum/quantum.conf
> /etc/quantum/dhcp_agent.ini
> /etc/quantum/plugins/openvswitch/ovs_quantum_plugin.ini
> /etc/quantum/api-paste.ini
> /var/log/nova/*.log
> /var/log/quantum/*.log
>
> Compute Node :
> /etc/nova/nova.conf
> /etc/nova/nova-compute.conf
> /etc/nova/api-paste.ini
> /etc/quantum/quantum.conf
> /etc/quantum/plugins/openvswitch/ovs_quantum_plugin.ini
> /var/log/nova/*.log
> /var/log/quantum/*.log
>
> Plus, complete output of the following commands :
>
> Controller Node :
> $ keystone endpoint-list
> $ ip link show
> $ ip route show
> $ ip netns show
> $ ovs-vsctl show
>
> Compute Node :
> $ ip link show
> $ ip route show
> $ ovs-vsctl show
>
> Regards,
>
> Jean-Baptiste RANSY
>
>
>
> On 02/16/2013 05:32 PM, Chathura M. Sarathchandra Magurawalage wrote:
>
> Hello Jean,
>
>  Thanks for your reply.
>
>  I followed the instructions in
> http://docs.openstack.org/folsom/basic-install/content/basic-install_network.html.
> And my Controller and the Network-node is installed in the same physical
> node.
>
>  I am using Folsom but without Network namespaces.
>
>  But in the website you have provided it states that "If you run both L3
> + DHCP services on the same node, you should enable namespaces to avoid
> conflicts with routes :"
>
>  But currently quantum-dhcp-agent and quantum-l3-agent are running in the
> same node?
>
>  Additionally the control node serves as a DHCP server for the local
> network ( Don't know if that would make and difference)
>
>  Any idea what the problem could be?
>
>
> On 16 February 2013 16:21, Jean-Baptiste RANSY <
> jean-baptiste.ransy at alyseo.com> wrote:
>
>>  Hello Chathura,
>>
>> Are you using Folsom with Network Namespaces ?
>>
>> If yes, have a look here :
>> http://docs.openstack.org/folsom/openstack-network/admin/content/ch_limitations.html
>>
>>
>> Regards,
>>
>> Jean-Baptsite RANSY
>>
>>
>>
>> On 02/16/2013 05:01 PM, Chathura M. Sarathchandra Magurawalage wrote:
>>
>>  Hello guys,
>>
>>  The problem still exists. Any ideas?
>>
>>  Thanks
>>
>>   On 15 February 2013 14:37, Sylvain Bauza <sylvain.bauza at digimind.com>wrote:
>>
>>> Metadata API allows to fetch SSH credentials when booting (pubkey I
>>> mean).
>>> If a VM is unable to reach metadata service, then it won't be able to
>>> get its public key, so you won't be able to connect, unless you
>>> specifically go thru a Password authentication (provided password auth is
>>> enabled in /etc/ssh/sshd_config, which is not the case with Ubuntu cloud
>>> archive).
>>> There is also a side effect, the boot process is longer as the instance
>>> is waiting for the curl timeout (60sec.) to finish booting up.
>>>
>>> Re: Quantum, the metadata API is actually DNAT'd from Network node to
>>> the Nova-api node (here 172.16.0.1 as internal management IP) :
>>> Chain quantum-l3-agent-PREROUTING (1 references)
>>>
>>> target     prot opt source               destination
>>>  DNAT       tcp  --  0.0.0.0/0            169.254.169.254      tcp
>>> dpt:80 to:172.16.0.1:8775
>>>
>>>
>>> Anyway, the first step is to :
>>> 1. grab the console.log
>>> 2. access thru VNC to the desired instance
>>>
>>> Troubleshooting will be easier once that done.
>>>
>>> -Sylvain
>>>
>>>
>>>
>>> Le 15/02/2013 14:24, Chathura M. Sarathchandra Magurawalage a écrit :
>>>
>>>>  Hello Guys,
>>>>
>>>> Not sure if this is the right port but these are the results:
>>>>
>>>>  *Compute node:*
>>>>
>>>>
>>>> root at computenode:~# netstat -an | grep 8775
>>>>  tcp        0      0 0.0.0.0:8775 <http://0.0.0.0:8775>  0.0.0.0:*
>>>>           LISTEN
>>>>
>>>> *Controller: *
>>>>
>>>>
>>>> root at controller:~# netstat -an | grep 8775
>>>>  tcp        0      0 0.0.0.0:8775 <http://0.0.0.0:8775>  0.0.0.0:*
>>>>           LISTEN
>>>>
>>>> *Additionally I cant curl 169.254.169.254 from the compute node. I am
>>>> not sure if this is related to not being able to PING the VM.*
>>>>
>>>>
>>>> curl -v http://169.254.169.254
>>>> * About to connect() to 169.254.169.254 port 80 (#0)
>>>> *   Trying 169.254.169.254...
>>>>
>>>> Thanks for your help
>>>>
>>>>
>>>>
>>>> -----------------------------------------------------------------------------------------------------------------------------
>>>> Chathura Madhusanka Sarathchandra Magurawalage.
>>>> 1NW.2.1, Desk 2
>>>> School of Computer Science and Electronic Engineering
>>>> University Of Essex
>>>> United Kingdom.
>>>>
>>>>  Email: csarata at essex.ac.uk <mailto:csarata at essex.ac.uk>
>>>>           chathura.sarathchandra at gmail.com <mailto:
>>>> 77.chathura at gmail.com>
>>>> 77.chathura at gmail.com <mailto:77.chathura at gmail.com>
>>>>
>>>>
>>>>
>>>> On 15 February 2013 11:03, Anil Vishnoi <vishnoianil at gmail.com <mailto:
>>>> vishnoianil at gmail.com>> wrote:
>>>>
>>>>     If you are using ubuntu cloud image then the only way to log-in is
>>>>     to do ssh with the public key. For that you have to create ssh key
>>>>     pair and download the ssh key. You can create this ssh pair using
>>>>     horizon/cli.
>>>>
>>>>
>>>>     On Fri, Feb 15, 2013 at 4:27 PM, Sylvain Bauza
>>>>      <sylvain.bauza at digimind.com <mailto:sylvain.bauza at digimind.com>>
>>>>
>>>>     wrote:
>>>>
>>>>
>>>>         Le 15/02/2013 11:42, Chathura M. Sarathchandra Magurawalage a
>>>>         écrit :
>>>>
>>>>
>>>>             How can I log into the VM from VNC? What are the
>>>> credentials?
>>>>
>>>>
>>>>         You have multiple ways to get VNC access. The easiest one is
>>>>         thru Horizon. Other can be looking at the KVM command-line for
>>>>         the desired instance (on the compute node) and check the vnc
>>>>         port in use (assuming KVM as hypervisor).
>>>>         This is basic knowledge of Nova.
>>>>
>>>>
>>>>
>>>>             nova-api-metadata is running fine in the compute node.
>>>>
>>>>
>>>>         Make sure the metadata port is avaible thanks to telnet or
>>>>         netstat, nova-api can be running without listening on metadata
>>>>         port.
>>>>
>>>>
>>>>
>>>>
>>>>         _______________________________________________
>>>>         Mailing list: https://launchpad.net/~openstack
>>>>          <https://launchpad.net/%7Eopenstack>
>>>>         Post to     : openstack at lists.launchpad.net
>>>>         <mailto:openstack at lists.launchpad.net>
>>>>         Unsubscribe : https://launchpad.net/~openstack
>>>>         <https://launchpad.net/%7Eopenstack>
>>>>
>>>>         More help   : https://help.launchpad.net/ListHelp
>>>>
>>>>
>>>>
>>>>
>>>>     --     Thanks & Regards
>>>>     --Anil Kumar Vishnoi
>>>>
>>>>
>>>>
>>>
>>
>>
>>  _______________________________________________
>> Mailing list: https://launchpad.net/~openstack
>>
>> Post to     : openstack at lists.launchpad.net
>> Unsubscribe : https://launchpad.net/~openstack
>> More help   : https://help.launchpad.net/ListHelp
>>
>>
>>
>
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.openstack.org/pipermail/openstack/attachments/20130216/9f90a65a/attachment.html>


More information about the Openstack mailing list