[Openstack] Cant ping private or floating IP

Chathura M. Sarathchandra Magurawalage 77.chathura at gmail.com
Sun Feb 17 02:18:56 UTC 2013


I did that. But no luck.

You should be able to curl 169.254.169.254 from compute node, which I cant
at the moment.

I have got the bridge set up in the l3_agent.ini

Every time I reset the controller the same ip was set back on the interface.

root at controller:~# ifconfig -a
br-ex     Link encap:Ethernet  HWaddr 9a:41:c8:8a:9e:49
          UP BROADCAST RUNNING MULTICAST  MTU:1500  Metric:1
          RX packets:6 errors:0 dropped:0 overruns:0 frame:0
          TX packets:6 errors:0 dropped:0 overruns:0 carrier:0
          collisions:0 txqueuelen:0
          RX bytes:468 (468.0 B)  TX bytes:468 (468.0 B)

br-int    Link encap:Ethernet  HWaddr ba:7a:e9:dc:2b:41
          inet6 addr: fe80::b87a:e9ff:fedc:2b41/64 Scope:Link
          UP BROADCAST RUNNING MULTICAST  MTU:1500  Metric:1
          RX packets:26722 errors:0 dropped:0 overruns:0 frame:0
          TX packets:6 errors:0 dropped:0 overruns:0 carrier:0
          collisions:0 txqueuelen:0
          RX bytes:3170358 (3.1 MB)  TX bytes:468 (468.0 B)

br-tun    Link encap:Ethernet  HWaddr 62:76:89:16:5a:47
          BROADCAST MULTICAST  MTU:1500  Metric:1
          RX packets:71 errors:0 dropped:0 overruns:0 frame:0
          TX packets:0 errors:0 dropped:0 overruns:0 carrier:0
          collisions:0 txqueuelen:0
          RX bytes:6760 (6.7 KB)  TX bytes:0 (0.0 B)

eth0      Link encap:Ethernet  HWaddr d4:ae:52:bb:aa:20
          inet addr:10.10.10.1  Bcast:10.10.10.255  Mask:255.255.255.0
          inet6 addr: fe80::d6ae:52ff:febb:aa20/64 Scope:Link
          UP BROADCAST RUNNING MULTICAST  MTU:1500  Metric:1
          RX packets:3472579 errors:0 dropped:0 overruns:0 frame:0
          TX packets:1742233 errors:0 dropped:0 overruns:0 carrier:0
          collisions:0 txqueuelen:1000
          RX bytes:760812279 (760.8 MB)  TX bytes:839533492 (839.5 MB)
          Interrupt:16 Memory:c0000000-c0012800

eth1      Link encap:Ethernet  HWaddr d4:ae:52:bb:aa:21
          BROADCAST MULTICAST  MTU:1500  Metric:1
          RX packets:0 errors:0 dropped:0 overruns:0 frame:0
          TX packets:0 errors:0 dropped:0 overruns:0 carrier:0
          collisions:0 txqueuelen:1000
          RX bytes:0 (0.0 B)  TX bytes:0 (0.0 B)
          Interrupt:17 Memory:c2000000-c2012800

eth0.2    Link encap:Ethernet  HWaddr d4:ae:52:bb:aa:20
          inet addr:192.168.2.225  Bcast:192.168.2.255  Mask:255.255.255.0
          inet6 addr: fe80::d6ae:52ff:febb:aa20/64 Scope:Link
          UP BROADCAST RUNNING MULTICAST  MTU:1500  Metric:1
          RX packets:2328304 errors:0 dropped:0 overruns:0 frame:0
          TX packets:755931 errors:0 dropped:0 overruns:0 carrier:0
          collisions:0 txqueuelen:0
          RX bytes:402847437 (402.8 MB)  TX bytes:497850618 (497.8 MB)

lo        Link encap:Local Loopback
          inet addr:127.0.0.1  Mask:255.0.0.0
          inet6 addr: ::1/128 Scope:Host
          UP LOOPBACK RUNNING  MTU:16436  Metric:1
          RX packets:10366873 errors:0 dropped:0 overruns:0 frame:0
          TX packets:10366873 errors:0 dropped:0 overruns:0 carrier:0
          collisions:0 txqueuelen:0
          RX bytes:3863071771 (3.8 GB)  TX bytes:3863071771 (3.8 GB)

qg-6f8374cb-cb Link encap:Ethernet  HWaddr 0e:7f:dd:3a:80:bc
          inet addr:192.168.2.151  Bcast:192.168.2.255  Mask:255.255.255.0
          inet6 addr: fe80::c7f:ddff:fe3a:80bc/64 Scope:Link
          UP BROADCAST RUNNING MULTICAST  MTU:1500  Metric:1
          RX packets:6 errors:0 dropped:0 overruns:0 frame:0
          TX packets:6 errors:0 dropped:0 overruns:0 carrier:0
          collisions:0 txqueuelen:0
          RX bytes:468 (468.0 B)  TX bytes:468 (468.0 B)

qr-4d088f3a-78 Link encap:Ethernet  HWaddr ca:5b:8d:4d:6d:fb
          inet addr:10.5.5.1  Bcast:10.5.5.255  Mask:255.255.255.0
          inet6 addr: fe80::c85b:8dff:fe4d:6dfb/64 Scope:Link
          UP BROADCAST RUNNING MULTICAST  MTU:1500  Metric:1
          RX packets:23657 errors:0 dropped:0 overruns:0 frame:0
          TX packets:10274 errors:0 dropped:0 overruns:0 carrier:0
          collisions:0 txqueuelen:0
          RX bytes:2891452 (2.8 MB)  TX bytes:565416 (565.4 KB)

tapf71b5b86-5c Link encap:Ethernet  HWaddr 2a:44:a3:d1:7d:f3
          inet addr:10.5.5.2  Bcast:10.5.5.255  Mask:255.255.255.0
          inet6 addr: fe80::2844:a3ff:fed1:7df3/64 Scope:Link
          UP BROADCAST RUNNING MULTICAST  MTU:1500  Metric:1
          RX packets:20310 errors:0 dropped:0 overruns:0 frame:0
          TX packets:27538 errors:0 dropped:0 overruns:0 carrier:0
          collisions:0 txqueuelen:0
          RX bytes:2278652 (2.2 MB)  TX bytes:2921102 (2.9 MB)


On 17 February 2013 01:52, Jean-Baptiste RANSY <
jean-baptiste.ransy at alyseo.com> wrote:

>  # ip addr del 192.168.2.225/24 dev br-ex
>
> you must configure br-ex in /etc/quantum/l3_agent.ini
> (external_network_bridge = br-ex), but it should not affect the current
> problem
>
> JB
>
>
> On 02/17/2013 02:47 AM, Chathura M. Sarathchandra Magurawalage wrote:
>
> Do you mean br-ex interface and eth0.2? well the br-ex was created by
> openstack and eth0.2 is the network that I get access to outside from.
>
>  Thanks.
>
>
>  On 17 February 2013 01:33, Jean-Baptiste RANSY <
> jean-baptiste.ransy at alyseo.com> wrote:
>
>>  You have two interfaces with the same ip address
>>
>> The problem must be here now.
>>
>> JB
>>
>>
>>  *root at controller:~# ip addr show*
>> 4: eth0.2 at eth0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc noqueue
>> state UP
>>     link/ether d4:ae:52:bb:aa:20 brd ff:ff:ff:ff:ff:ff
>>     inet 192.168.2.225/24 brd 192.168.2.255 scope global eth0.2
>>     inet6 fe80::d6ae:52ff:febb:aa20/64 scope link
>>        valid_lft forever preferred_lft forever
>> 7: br-ex: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc noqueue state
>> UNKNOWN
>>      link/ether 9a:41:c8:8a:9e:49 brd ff:ff:ff:ff:ff:ff
>>     inet 192.168.2.225/24 scope global br-ex
>>
>>
>>
>> On 02/17/2013 02:15 AM, Chathura M. Sarathchandra Magurawalage wrote:
>>
>> Hello JB,
>>
>>  I changed the IP and restarted the quantum-l3-agent but still no luck :(
>>
>>  Thanks.
>>
>>
>> -----------------------------------------------------------------------------------------------------------------------------
>> Chathura Madhusanka Sarathchandra Magurawalage.
>> 1NW.2.1, Desk 2
>> School of Computer Science and Electronic Engineering
>> University Of Essex
>>  United Kingdom.
>>
>>  Email: csarata at essex.ac.uk
>>           chathura.sarathchandra at gmail.com <77.chathura at gmail.com>
>>            77.chathura at gmail.com
>>
>>
>> On 17 February 2013 00:47, Jean-Baptiste RANSY <
>> jean-baptiste.ransy at alyseo.com> wrote:
>>
>>>  and restart quantum-l3-agent :)
>>>
>>> JB
>>>
>>>
>>> On 02/17/2013 01:46 AM, Jean-Baptiste RANSY wrote:
>>>
>>> Found !
>>>
>>> On the controller node you must change the metadata_ip in
>>> /etc/quantum/l3_agent.ini
>>>
>>> This params is used to create the nat rule quantum-l3-agent-PREROUTING
>>>
>>> Just replace 127.0.0.1 by 192.168.2.225 and that should be ok.
>>>
>>> JB
>>>
>>>
>>> On 02/17/2013 01:04 AM, Jean-Baptiste RANSY wrote:
>>>
>>> Hi, Chathura
>>>
>>> The compute node log file /var/log/nova/nova-api.log is too light (maybe
>>> logrotate :p)
>>>
>>> Please, clear nova-api.log, restart nova-api service, start a new
>>> instance and wait cloud-init fail to retrieve metadata.
>>>
>>> Thx,
>>>
>>> JB
>>>
>>>
>>> On 02/16/2013 11:35 PM, Chathura M. Sarathchandra Magurawalage wrote:
>>>
>>>  Thanks for that.
>>> *
>>> *
>>> *root at controller:~# ip addr show*
>>> 1: lo: <LOOPBACK,UP,LOWER_UP> mtu 16436 qdisc noqueue state UNKNOWN
>>>     link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00
>>>     inet 127.0.0.1/8 scope host lo
>>>     inet6 ::1/128 scope host
>>>        valid_lft forever preferred_lft forever
>>> 2: eth0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc mq state UP
>>> qlen 1000
>>>     link/ether d4:ae:52:bb:aa:20 brd ff:ff:ff:ff:ff:ff
>>>     inet 10.10.10.1/24 brd 10.10.10.255 scope global eth0
>>>     inet6 fe80::d6ae:52ff:febb:aa20/64 scope link
>>>        valid_lft forever preferred_lft forever
>>> 3: eth1: <BROADCAST,MULTICAST> mtu 1500 qdisc noop state DOWN qlen 1000
>>>     link/ether d4:ae:52:bb:aa:21 brd ff:ff:ff:ff:ff:ff
>>> 4: eth0.2 at eth0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc
>>> noqueue state UP
>>>     link/ether d4:ae:52:bb:aa:20 brd ff:ff:ff:ff:ff:ff
>>>     inet 192.168.2.225/24 brd 192.168.2.255 scope global eth0.2
>>>     inet6 fe80::d6ae:52ff:febb:aa20/64 scope link
>>>        valid_lft forever preferred_lft forever
>>> 5: br-int: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc noqueue
>>> state UNKNOWN
>>>     link/ether ba:7a:e9:dc:2b:41 brd ff:ff:ff:ff:ff:ff
>>>     inet6 fe80::b87a:e9ff:fedc:2b41/64 scope link
>>>        valid_lft forever preferred_lft forever
>>> 7: br-ex: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc noqueue state
>>> UNKNOWN
>>>      link/ether 9a:41:c8:8a:9e:49 brd ff:ff:ff:ff:ff:ff
>>>     inet 192.168.2.225/24 scope global br-ex
>>> 8: tapf71b5b86-5c: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc
>>> noqueue state UNKNOWN
>>>     link/ether 2a:44:a3:d1:7d:f3 brd ff:ff:ff:ff:ff:ff
>>>     inet 10.5.5.2/24 brd 10.5.5.255 scope global tapf71b5b86-5c
>>>     inet6 fe80::2844:a3ff:fed1:7df3/64 scope link
>>>        valid_lft forever preferred_lft forever
>>> 9: qr-4d088f3a-78: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc
>>> noqueue state UNKNOWN
>>>     link/ether ca:5b:8d:4d:6d:fb brd ff:ff:ff:ff:ff:ff
>>>     inet 10.5.5.1/24 brd 10.5.5.255 scope global qr-4d088f3a-78
>>>     inet6 fe80::c85b:8dff:fe4d:6dfb/64 scope link
>>>        valid_lft forever preferred_lft forever
>>> 10: qg-6f8374cb-cb: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc
>>> noqueue state UNKNOWN
>>>     link/ether 0e:7f:dd:3a:80:bc brd ff:ff:ff:ff:ff:ff
>>>     inet 192.168.2.151/24 brd 192.168.2.255 scope global qg-6f8374cb-cb
>>>     inet6 fe80::c7f:ddff:fe3a:80bc/64 scope link
>>>        valid_lft forever preferred_lft forever
>>> 27: br-tun: <BROADCAST,MULTICAST> mtu 1500 qdisc noop state DOWN
>>>     link/ether 8a:cf:ec:7c:15:40 brd ff:ff:ff:ff:ff:ff
>>>
>>>  *cat /proc/sys/net/ipv4/ip_forward*
>>> 1
>>>
>>>  *root at computenode:~# ip addr show*
>>> 1: lo: <LOOPBACK,UP,LOWER_UP> mtu 16436 qdisc noqueue state UNKNOWN
>>>     link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00
>>>     inet 127.0.0.1/8 scope host lo
>>>     inet6 ::1/128 scope host
>>>        valid_lft forever preferred_lft forever
>>> 2: eth0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc mq state UP
>>> qlen 1000
>>>     link/ether d4:ae:52:bb:a1:9d brd ff:ff:ff:ff:ff:ff
>>>     inet 10.10.10.12/24 brd 10.10.10.255 scope global eth0
>>>     inet6 fe80::d6ae:52ff:febb:a19d/64 scope link
>>>        valid_lft forever preferred_lft forever
>>> 3: eth1: <BROADCAST,MULTICAST> mtu 1500 qdisc noop state DOWN qlen 1000
>>>     link/ether d4:ae:52:bb:a1:9e brd ff:ff:ff:ff:ff:ff
>>> 4: eth0.2 at eth0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc
>>> noqueue state UP
>>>     link/ether d4:ae:52:bb:a1:9d brd ff:ff:ff:ff:ff:ff
>>>     inet 192.168.2.234/24 brd 192.168.2.255 scope global eth0.2
>>>     inet6 fe80::d6ae:52ff:febb:a19d/64 scope link
>>>        valid_lft forever preferred_lft forever
>>> 5: br-int: <BROADCAST,MULTICAST> mtu 1500 qdisc noop state DOWN
>>>     link/ether ae:9b:43:09:af:40 brd ff:ff:ff:ff:ff:ff
>>> 9: qbr256f5ed2-43: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc
>>> noqueue state UP
>>>     link/ether c6:c0:df:64:c6:99 brd ff:ff:ff:ff:ff:ff
>>>     inet6 fe80::20e8:b9ff:fe6c:6f55/64 scope link
>>>        valid_lft forever preferred_lft forever
>>> 10: qvo256f5ed2-43: <BROADCAST,MULTICAST,PROMISC,UP,LOWER_UP> mtu 1500
>>> qdisc pfifo_fast state UP qlen 1000
>>>     link/ether 76:25:8b:fd:90:3b brd ff:ff:ff:ff:ff:ff
>>>     inet6 fe80::7425:8bff:fefd:903b/64 scope link
>>>        valid_lft forever preferred_lft forever
>>> 11: qvb256f5ed2-43: <BROADCAST,MULTICAST,PROMISC,UP,LOWER_UP> mtu 1500
>>> qdisc pfifo_fast master qbr256f5ed2-43 state UP qlen 1000
>>>     link/ether c6:c0:df:64:c6:99 brd ff:ff:ff:ff:ff:ff
>>>     inet6 fe80::c4c0:dfff:fe64:c699/64 scope link
>>>        valid_lft forever preferred_lft forever
>>> 13: br-tun: <BROADCAST,MULTICAST> mtu 1500 qdisc noop state DOWN
>>>     link/ether be:8c:30:78:35:48 brd ff:ff:ff:ff:ff:ff
>>> 15: vnet0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc pfifo_fast
>>> master qbr256f5ed2-43 state UNKNOWN qlen 500
>>>     link/ether fe:16:3e:57:ec:ff brd ff:ff:ff:ff:ff:ff
>>>     inet6 fe80::fc16:3eff:fe57:ecff/64 scope link
>>>        valid_lft forever preferred_lft forever
>>>
>>>  btw cronus is my compute node and I have renamed it to computenode to
>>> understand it better.
>>>
>>> On 16 February 2013 22:11, Jean-Baptiste RANSY <
>>> jean-baptiste.ransy at alyseo.com> wrote:
>>>
>>>>  Download in progress
>>>>
>>>> Can you send me the output of those commands i forgot :
>>>>
>>>> Controller Node:
>>>> $ ip addr show
>>>> $ cat /proc/sys/net/ipv4/ip_forward
>>>>
>>>> Compute Node:
>>>> $ ip addr show
>>>>
>>>>
>>>> JB
>>>>
>>>>
>>>>
>>>> On 02/16/2013 10:45 PM, Chathura M. Sarathchandra Magurawalage wrote:
>>>>
>>>>  Thanks Ransy,
>>>>
>>>>  I have created a tar file with the configuration and log files in it.
>>>> Please download it using the following URL. I have pasted the output of the
>>>> commands below.
>>>>
>>>>  https://www.dropbox.com/s/qyfcsn50060y304/confilesnlogs.tar
>>>>
>>>>  *Controller node:*
>>>> *root at controller:~# keystone endpoint-list*
>>>> +----------------------------------+-----------+-------------------------------------------------+-------------------------------------------------+--------------------------------------------+Controller
>>>> node
>>>> |                id                |   region  |
>>>>  publicurl                    |                   internalurl
>>>>     |                  adminurl                  |
>>>>
>>>> +----------------------------------+-----------+-------------------------------------------------+-------------------------------------------------+--------------------------------------------+
>>>> | 2c9a1cb0fe8247d9b7716432cf459fe5 | RegionOne |
>>>> http://192.168.2.225:8774/v2/$(tenant_id)s   |
>>>> http://192.168.2.225:8774/v2/$(tenant_id)s   |
>>>> http://192.168.2.225:8774/v2/$(tenant_id)s |
>>>> | 2d306903ed3342a8aaaac7c5680c116f | RegionOne |
>>>> http://192.168.2.225:9696/           |
>>>> http://192.168.2.225:9696/           |
>>>> http://192.168.2.225:9696/         |
>>>> | 3848114f120f42bf819bc2443b28ac9e | RegionOne |
>>>> http://192.168.2.225:8080/v1/AUTH_$(tenant_id)s |
>>>> http://192.168.2.225:8080/v1/AUTH_$(tenant_id)s |
>>>> http://192.168.2.225:8080/v1        |
>>>> | 4955173b8d9e4d33ae4a5b29dc12c74d | RegionOne |
>>>> http://192.168.2.225:8776/v1/$(tenant_id)s   |
>>>> http://192.168.2.225:8776/v1/$(tenant_id)s   |
>>>> http://192.168.2.225:8776/v1/$(tenant_id)s |
>>>> | d313aa76bf854dde94f33a49a9f0c8ac | RegionOne |
>>>> http://192.168.2.225:9292/v2          |
>>>> http://192.168.2.225:9292/v2          |
>>>> http://192.168.2.225:9292/v2        |
>>>> | e5aa4ecf3cbe4dd5aba9b204c74fee6a | RegionOne |
>>>> http://192.168.2.225:5000/v2.0         |
>>>> http://192.168.2.225:5000/v2.0         |
>>>> http://192.168.2.225:35357/v2.0       |
>>>> | fba6f790e3b444c890d114f13cd32b37 | RegionOne |
>>>> http://192.168.2.225:8773/services/Cloud    |
>>>> http://192.168.2.225:8773/services/Cloud    |
>>>> http://192.168.2.225:8773/services/Admin  |
>>>>
>>>> +----------------------------------+-----------+-------------------------------------------------+-------------------------------------------------+--------------------------------------------+
>>>>
>>>>  *root at controller:~# ip link show*
>>>> 1: lo: <LOOPBACK,UP,LOWER_UP> mtu 16436 qdisc noqueue state UNKNOWN
>>>>     link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00
>>>> 2: eth0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc mq state UP
>>>> qlen 1000
>>>>     link/ether d4:ae:52:bb:aa:20 brd ff:ff:ff:ff:ff:ff
>>>> 3: eth1: <BROADCAST,MULTICAST> mtu 1500 qdisc noop state DOWN qlen 1000
>>>>     link/ether d4:ae:52:bb:aa:21 brd ff:ff:ff:ff:ff:ff
>>>>  4: eth0.2 at eth0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc
>>>> noqueue state UP
>>>>     link/ether d4:ae:52:bb:aa:20 brd ff:ff:ff:ff:ff:ff
>>>> 5: br-int: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc noqueue
>>>> state UNKNOWN
>>>>     link/ether ba:7a:e9:dc:2b:41 brd ff:ff:ff:ff:ff:ff
>>>> 7: br-ex: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc noqueue
>>>> state UNKNOWN
>>>>     link/ether 9a:41:c8:8a:9e:49 brd ff:ff:ff:ff:ff:ff
>>>> 8: tapf71b5b86-5c: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc
>>>> noqueue state UNKNOWN
>>>>     link/ether 2a:44:a3:d1:7d:f3 brd ff:ff:ff:ff:ff:ff
>>>> 9: qr-4d088f3a-78: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc
>>>> noqueue state UNKNOWN
>>>>     link/ether ca:5b:8d:4d:6d:fb brd ff:ff:ff:ff:ff:ff
>>>> 10: qg-6f8374cb-cb: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc
>>>> noqueue state UNKNOWN
>>>>     link/ether 0e:7f:dd:3a:80:bc brd ff:ff:ff:ff:ff:ff
>>>> 27: br-tun: <BROADCAST,MULTICAST> mtu 1500 qdisc noop state DOWN
>>>>     link/ether 8a:cf:ec:7c:15:40 brd ff:ff:ff:ff:ff:ff
>>>>
>>>>  *root at controller:~# ip route show*
>>>> default via 192.168.2.253 dev eth0.2
>>>> default via 192.168.2.253 dev eth0.2  metric 100
>>>> 10.5.5.0/24 dev tapf71b5b86-5c  proto kernel  scope link  src 10.5.5.2
>>>> 10.5.5.0/24 dev qr-4d088f3a-78  proto kernel  scope link  src 10.5.5.1
>>>> 10.10.10.0/24 dev eth0  proto kernel  scope link  src 10.10.10.1
>>>> 192.168.2.0/24 dev eth0.2  proto kernel  scope link  src 192.168.2.225
>>>> 192.168.2.0/24 dev qg-6f8374cb-cb  proto kernel  scope link  src
>>>> 192.168.2.151
>>>> 192.168.2.0/24 dev br-ex  proto kernel  scope link  src 192.168.2.225
>>>>
>>>> *$ ip netns show (Did not return anything)*
>>>>
>>>>  *root at controller:~# ovs-vsctl show*
>>>> a566afae-d7a8-42a9-aefe-8b0f2f7054a3
>>>>     Bridge br-tun
>>>>         Port "gre-4"
>>>>             Interface "gre-4"
>>>>                 type: gre
>>>>                 options: {in_key=flow, out_key=flow,
>>>> remote_ip="10.10.10.12"}
>>>>         Port "gre-3"
>>>>             Interface "gre-3"
>>>>                  type: gre
>>>>                 options: {in_key=flow, out_key=flow,
>>>> remote_ip="127.0.0.1"}
>>>>         Port patch-int
>>>>             Interface patch-int
>>>>                 type: patch
>>>>                 options: {peer=patch-tun}
>>>>         Port br-tun
>>>>              Interface br-tun
>>>>                 type: internal
>>>>         Port "gre-1"
>>>>             Interface "gre-1"
>>>>                 type: gre
>>>>                  options: {in_key=flow, out_key=flow,
>>>> remote_ip="10.0.0.3"}
>>>>     Bridge br-ex
>>>>          Port br-ex
>>>>             Interface br-ex
>>>>                 type: internal
>>>>         Port "qg-6f8374cb-cb"
>>>>             Interface "qg-6f8374cb-cb"
>>>>                 type: internal
>>>>         Port "br0"
>>>>              Interface "br0"
>>>>     Bridge br-int
>>>>         Port br-int
>>>>             Interface br-int
>>>>                 type: internal
>>>>          Port "tapf71b5b86-5c"
>>>>             tag: 1
>>>>             Interface "tapf71b5b86-5c"
>>>>                 type: internal
>>>>         Port patch-tun
>>>>             Interface patch-tun
>>>>                 type: patch
>>>>                 options: {peer=patch-int}
>>>>          Port "qr-4d088f3a-78"
>>>>             tag: 1
>>>>             Interface "qr-4d088f3a-78"
>>>>                 type: internal
>>>>     ovs_version: "1.4.0+build0"
>>>>
>>>>
>>>>  *Compute node:*
>>>>
>>>>  *root at cronus:~# ip link show*
>>>> 1: lo: <LOOPBACK,UP,LOWER_UP> mtu 16436 qdisc noqueue state UNKNOWN
>>>>      link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00
>>>> 2: eth0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc mq state UP
>>>> qlen 1000
>>>>     link/ether d4:ae:52:bb:a1:9d brd ff:ff:ff:ff:ff:ff
>>>> 3: eth1: <BROADCAST,MULTICAST> mtu 1500 qdisc noop state DOWN qlen 1000
>>>>     link/ether d4:ae:52:bb:a1:9e brd ff:ff:ff:ff:ff:ff
>>>> 4: eth0.2 at eth0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc
>>>> noqueue state UP
>>>>     link/ether d4:ae:52:bb:a1:9d brd ff:ff:ff:ff:ff:ff
>>>> 5: br-int: <BROADCAST,MULTICAST> mtu 1500 qdisc noop state DOWN
>>>>     link/ether ae:9b:43:09:af:40 brd ff:ff:ff:ff:ff:ff
>>>> 9: qbr256f5ed2-43: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc
>>>> noqueue state UP
>>>>     link/ether c6:c0:df:64:c6:99 brd ff:ff:ff:ff:ff:ff
>>>> 10: qvo256f5ed2-43: <BROADCAST,MULTICAST,PROMISC,UP,LOWER_UP> mtu 1500
>>>> qdisc pfifo_fast state UP qlen 1000
>>>>     link/ether 76:25:8b:fd:90:3b brd ff:ff:ff:ff:ff:ff
>>>> 11: qvb256f5ed2-43: <BROADCAST,MULTICAST,PROMISC,UP,LOWER_UP> mtu 1500
>>>> qdisc pfifo_fast master qbr256f5ed2-43 state UP qlen 1000
>>>>     link/ether c6:c0:df:64:c6:99 brd ff:ff:ff:ff:ff:ff
>>>> 13: br-tun: <BROADCAST,MULTICAST> mtu 1500 qdisc noop state DOWN
>>>>     link/ether be:8c:30:78:35:48 brd ff:ff:ff:ff:ff:ff
>>>> 15: vnet0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc pfifo_fast
>>>> master qbr256f5ed2-43 state UNKNOWN qlen 500
>>>>     link/ether fe:16:3e:57:ec:ff brd ff:ff:ff:ff:ff:ff
>>>>
>>>>   *root at cronus:~# ip route show*
>>>> default via 192.168.2.253 dev eth0.2  metric 100
>>>> 10.10.10.0/24 dev eth0  proto kernel  scope link  src 10.10.10.12
>>>> 192.168.2.0/24 dev eth0.2  proto kernel  scope link  src 192.168.2.234
>>>>
>>>>   *root at cronus:~# ovs-vsctl show*
>>>> d85bc334-6d64-4a13-b851-d56b18ff1549
>>>>     Bridge br-int
>>>>         Port "qvo0e743b01-89"
>>>>             tag: 4095
>>>>             Interface "qvo0e743b01-89"
>>>>         Port "qvo256f5ed2-43"
>>>>             tag: 1
>>>>             Interface "qvo256f5ed2-43"
>>>>         Port patch-tun
>>>>             Interface patch-tun
>>>>                 type: patch
>>>>                 options: {peer=patch-int}
>>>>         Port br-int
>>>>             Interface br-int
>>>>                 type: internal
>>>>         Port "qvoee3d4131-2a"
>>>>             tag: 4095
>>>>             Interface "qvoee3d4131-2a"
>>>>         Port "qvocbc816bd-3d"
>>>>             tag: 4095
>>>>             Interface "qvocbc816bd-3d"
>>>>     Bridge br-tun
>>>>         Port br-tun
>>>>             Interface br-tun
>>>>                 type: internal
>>>>         Port "gre-2"
>>>>             Interface "gre-2"
>>>>                 type: gre
>>>>                 options: {in_key=flow, out_key=flow,
>>>> remote_ip="10.10.10.1"}
>>>>         Port "gre-1"
>>>>             Interface "gre-1"
>>>>                 type: gre
>>>>                 options: {in_key=flow, out_key=flow,
>>>> remote_ip="10.0.0.3"}
>>>>         Port patch-int
>>>>             Interface patch-int
>>>>                 type: patch
>>>>                 options: {peer=patch-tun}
>>>>         Port "gre-3"
>>>>             Interface "gre-3"
>>>>                 type: gre
>>>>                 options: {in_key=flow, out_key=flow,
>>>> remote_ip="127.0.0.1"}
>>>>     ovs_version: "1.4.0+build0"
>>>>
>>>>
>>>>  Thanks I appreciate your help.
>>>>
>>>>   On 16 February 2013 16:49, Jean-Baptiste RANSY <
>>>> jean-baptiste.ransy at alyseo.com> wrote:
>>>>
>>>>>  Please provide files listed bellow :
>>>>>
>>>>> Controller Node :
>>>>> /etc/nova/nova.conf
>>>>> /etc/nova/api-paste.ini
>>>>> /etc/quantum/l3_agent.ini
>>>>> /etc/quantum/quantum.conf
>>>>> /etc/quantum/dhcp_agent.ini
>>>>> /etc/quantum/plugins/openvswitch/ovs_quantum_plugin.ini
>>>>> /etc/quantum/api-paste.ini
>>>>> /var/log/nova/*.log
>>>>> /var/log/quantum/*.log
>>>>>
>>>>> Compute Node :
>>>>> /etc/nova/nova.conf
>>>>> /etc/nova/nova-compute.conf
>>>>> /etc/nova/api-paste.ini
>>>>> /etc/quantum/quantum.conf
>>>>> /etc/quantum/plugins/openvswitch/ovs_quantum_plugin.ini
>>>>> /var/log/nova/*.log
>>>>> /var/log/quantum/*.log
>>>>>
>>>>> Plus, complete output of the following commands :
>>>>>
>>>>> Controller Node :
>>>>> $ keystone endpoint-list
>>>>> $ ip link show
>>>>> $ ip route show
>>>>> $ ip netns show
>>>>> $ ovs-vsctl show
>>>>>
>>>>> Compute Node :
>>>>> $ ip link show
>>>>> $ ip route show
>>>>> $ ovs-vsctl show
>>>>>
>>>>> Regards,
>>>>>
>>>>> Jean-Baptiste RANSY
>>>>>
>>>>>
>>>>>
>>>>> On 02/16/2013 05:32 PM, Chathura M. Sarathchandra Magurawalage wrote:
>>>>>
>>>>> Hello Jean,
>>>>>
>>>>>  Thanks for your reply.
>>>>>
>>>>>  I followed the instructions in
>>>>> http://docs.openstack.org/folsom/basic-install/content/basic-install_network.html.
>>>>> And my Controller and the Network-node is installed in the same physical
>>>>> node.
>>>>>
>>>>>  I am using Folsom but without Network namespaces.
>>>>>
>>>>>  But in the website you have provided it states that "If you run both
>>>>> L3 + DHCP services on the same node, you should enable namespaces to avoid
>>>>> conflicts with routes :"
>>>>>
>>>>>  But currently quantum-dhcp-agent and quantum-l3-agent are running in
>>>>> the same node?
>>>>>
>>>>>  Additionally the control node serves as a DHCP server for the local
>>>>> network ( Don't know if that would make and difference)
>>>>>
>>>>>  Any idea what the problem could be?
>>>>>
>>>>>
>>>>> On 16 February 2013 16:21, Jean-Baptiste RANSY <
>>>>> jean-baptiste.ransy at alyseo.com> wrote:
>>>>>
>>>>>>  Hello Chathura,
>>>>>>
>>>>>> Are you using Folsom with Network Namespaces ?
>>>>>>
>>>>>> If yes, have a look here :
>>>>>> http://docs.openstack.org/folsom/openstack-network/admin/content/ch_limitations.html
>>>>>>
>>>>>>
>>>>>> Regards,
>>>>>>
>>>>>> Jean-Baptsite RANSY
>>>>>>
>>>>>>
>>>>>>
>>>>>> On 02/16/2013 05:01 PM, Chathura M. Sarathchandra Magurawalage wrote:
>>>>>>
>>>>>>  Hello guys,
>>>>>>
>>>>>>  The problem still exists. Any ideas?
>>>>>>
>>>>>>  Thanks
>>>>>>
>>>>>>   On 15 February 2013 14:37, Sylvain Bauza <
>>>>>> sylvain.bauza at digimind.com> wrote:
>>>>>>
>>>>>>> Metadata API allows to fetch SSH credentials when booting (pubkey I
>>>>>>> mean).
>>>>>>> If a VM is unable to reach metadata service, then it won't be able
>>>>>>> to get its public key, so you won't be able to connect, unless you
>>>>>>> specifically go thru a Password authentication (provided password auth is
>>>>>>> enabled in /etc/ssh/sshd_config, which is not the case with Ubuntu cloud
>>>>>>> archive).
>>>>>>> There is also a side effect, the boot process is longer as the
>>>>>>> instance is waiting for the curl timeout (60sec.) to finish booting up.
>>>>>>>
>>>>>>> Re: Quantum, the metadata API is actually DNAT'd from Network node
>>>>>>> to the Nova-api node (here 172.16.0.1 as internal management IP) :
>>>>>>> Chain quantum-l3-agent-PREROUTING (1 references)
>>>>>>>
>>>>>>> target     prot opt source               destination
>>>>>>>  DNAT       tcp  --  0.0.0.0/0            169.254.169.254      tcp
>>>>>>> dpt:80 to:172.16.0.1:8775
>>>>>>>
>>>>>>>
>>>>>>> Anyway, the first step is to :
>>>>>>> 1. grab the console.log
>>>>>>> 2. access thru VNC to the desired instance
>>>>>>>
>>>>>>> Troubleshooting will be easier once that done.
>>>>>>>
>>>>>>> -Sylvain
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>> Le 15/02/2013 14:24, Chathura M. Sarathchandra Magurawalage a écrit :
>>>>>>>
>>>>>>>>  Hello Guys,
>>>>>>>>
>>>>>>>> Not sure if this is the right port but these are the results:
>>>>>>>>
>>>>>>>>  *Compute node:*
>>>>>>>>
>>>>>>>>
>>>>>>>> root at computenode:~# netstat -an | grep 8775
>>>>>>>>  tcp        0      0 0.0.0.0:8775 <http://0.0.0.0:8775>  0.0.0.0:*
>>>>>>>>               LISTEN
>>>>>>>>
>>>>>>>> *Controller: *
>>>>>>>>
>>>>>>>>
>>>>>>>> root at controller:~# netstat -an | grep 8775
>>>>>>>>  tcp        0      0 0.0.0.0:8775 <http://0.0.0.0:8775>  0.0.0.0:*
>>>>>>>>               LISTEN
>>>>>>>>
>>>>>>>> *Additionally I cant curl 169.254.169.254 from the compute node. I
>>>>>>>> am not sure if this is related to not being able to PING the VM.*
>>>>>>>>
>>>>>>>>
>>>>>>>> curl -v http://169.254.169.254
>>>>>>>> * About to connect() to 169.254.169.254 port 80 (#0)
>>>>>>>> *   Trying 169.254.169.254...
>>>>>>>>
>>>>>>>> Thanks for your help
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>> -----------------------------------------------------------------------------------------------------------------------------
>>>>>>>> Chathura Madhusanka Sarathchandra Magurawalage.
>>>>>>>> 1NW.2.1, Desk 2
>>>>>>>> School of Computer Science and Electronic Engineering
>>>>>>>> University Of Essex
>>>>>>>> United Kingdom.
>>>>>>>>
>>>>>>>>  Email: csarata at essex.ac.uk <mailto:csarata at essex.ac.uk>
>>>>>>>>           chathura.sarathchandra at gmail.com <mailto:
>>>>>>>> 77.chathura at gmail.com>
>>>>>>>> 77.chathura at gmail.com <mailto:77.chathura at gmail.com>
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>> On 15 February 2013 11:03, Anil Vishnoi <vishnoianil at gmail.com<mailto:
>>>>>>>> vishnoianil at gmail.com>> wrote:
>>>>>>>>
>>>>>>>>     If you are using ubuntu cloud image then the only way to log-in
>>>>>>>> is
>>>>>>>>     to do ssh with the public key. For that you have to create ssh
>>>>>>>> key
>>>>>>>>     pair and download the ssh key. You can create this ssh pair
>>>>>>>> using
>>>>>>>>     horizon/cli.
>>>>>>>>
>>>>>>>>
>>>>>>>>     On Fri, Feb 15, 2013 at 4:27 PM, Sylvain Bauza
>>>>>>>>      <sylvain.bauza at digimind.com <mailto:sylvain.bauza at digimind.com>>
>>>>>>>>
>>>>>>>>
>>>>>>>>     wrote:
>>>>>>>>
>>>>>>>>
>>>>>>>>         Le 15/02/2013 11:42, Chathura M. Sarathchandra Magurawalage
>>>>>>>> a
>>>>>>>>         écrit :
>>>>>>>>
>>>>>>>>
>>>>>>>>             How can I log into the VM from VNC? What are the
>>>>>>>> credentials?
>>>>>>>>
>>>>>>>>
>>>>>>>>         You have multiple ways to get VNC access. The easiest one is
>>>>>>>>         thru Horizon. Other can be looking at the KVM command-line
>>>>>>>> for
>>>>>>>>         the desired instance (on the compute node) and check the vnc
>>>>>>>>         port in use (assuming KVM as hypervisor).
>>>>>>>>         This is basic knowledge of Nova.
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>>             nova-api-metadata is running fine in the compute node.
>>>>>>>>
>>>>>>>>
>>>>>>>>         Make sure the metadata port is avaible thanks to telnet or
>>>>>>>>         netstat, nova-api can be running without listening on
>>>>>>>> metadata
>>>>>>>>         port.
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>>         _______________________________________________
>>>>>>>>         Mailing list: https://launchpad.net/~openstack
>>>>>>>>          <https://launchpad.net/%7Eopenstack>
>>>>>>>>         Post to     : openstack at lists.launchpad.net
>>>>>>>>         <mailto:openstack at lists.launchpad.net>
>>>>>>>>         Unsubscribe : https://launchpad.net/~openstack
>>>>>>>>         <https://launchpad.net/%7Eopenstack>
>>>>>>>>
>>>>>>>>         More help   : https://help.launchpad.net/ListHelp
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>>     --     Thanks & Regards
>>>>>>>>     --Anil Kumar Vishnoi
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>
>>>>>>
>>>>>>
>>>>>>  _______________________________________________
>>>>>> Mailing list: https://launchpad.net/~openstack
>>>>>>
>>>>>> Post to     : openstack at lists.launchpad.net
>>>>>> Unsubscribe : https://launchpad.net/~openstack
>>>>>> More help   : https://help.launchpad.net/ListHelp
>>>>>>
>>>>>>
>>>>>>
>>>>>
>>>>>
>>>>
>>>>
>>>
>>>
>>>
>>>
>>
>>
>
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.openstack.org/pipermail/openstack/attachments/20130217/4cc90090/attachment.html>


More information about the Openstack mailing list