← Back to team overview

openstack team mailing list archive

Re: Cant ping private or floating IP

 

You have two interfaces with the same ip address

The problem must be here now.

JB


> *root@controller:~# ip addr show*
> 4: eth0.2@eth0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc
> noqueue state UP 
>     link/ether d4:ae:52:bb:aa:20 brd ff:ff:ff:ff:ff:ff
>     inet 192.168.2.225/24 <http://192.168.2.225/24> brd 192.168.2.255
> scope global eth0.2
>     inet6 fe80::d6ae:52ff:febb:aa20/64 scope link 
>        valid_lft forever preferred_lft forever
> 7: br-ex: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc noqueue
> state UNKNOWN 
>     link/ether 9a:41:c8:8a:9e:49 brd ff:ff:ff:ff:ff:ff
>     inet 192.168.2.225/24 <http://192.168.2.225/24> scope global br-ex


On 02/17/2013 02:15 AM, Chathura M. Sarathchandra Magurawalage wrote:
> Hello JB,
>
> I changed the IP and restarted the quantum-l3-agent but still no luck :(
>
> Thanks.
>
> -----------------------------------------------------------------------------------------------------------------------------
> Chathura Madhusanka Sarathchandra Magurawalage.
> 1NW.2.1, Desk 2
> School of Computer Science and Electronic Engineering
> University Of Essex
> United Kingdom.
>
> Email: csarata@xxxxxxxxxxx <mailto:csarata@xxxxxxxxxxx>
>           chathura.sarathchandra@xxxxxxxxx <mailto:77.chathura@xxxxxxxxx>
>           77.chathura@xxxxxxxxx <mailto:77.chathura@xxxxxxxxx>
>
>
> On 17 February 2013 00:47, Jean-Baptiste RANSY
> <jean-baptiste.ransy@xxxxxxxxxx
> <mailto:jean-baptiste.ransy@xxxxxxxxxx>> wrote:
>
>     and restart quantum-l3-agent :)
>
>     JB
>
>
>     On 02/17/2013 01:46 AM, Jean-Baptiste RANSY wrote:
>>     Found !
>>
>>     On the controller node you must change the metadata_ip in
>>     /etc/quantum/l3_agent.ini
>>
>>     This params is used to create the nat rule
>>     quantum-l3-agent-PREROUTING
>>
>>     Just replace 127.0.0.1 by 192.168.2.225 and that should be ok.
>>
>>     JB
>>
>>
>>     On 02/17/2013 01:04 AM, Jean-Baptiste RANSY wrote:
>>>     Hi, Chathura
>>>
>>>     The compute node log file /var/log/nova/nova-api.log is too
>>>     light (maybe logrotate :p)
>>>
>>>     Please, clear nova-api.log, restart nova-api service, start a
>>>     new instance and wait cloud-init fail to retrieve metadata.
>>>
>>>     Thx,
>>>
>>>     JB
>>>
>>>
>>>     On 02/16/2013 11:35 PM, Chathura M. Sarathchandra Magurawalage
>>>     wrote:
>>>>     Thanks for that.
>>>>     *
>>>>     *
>>>>     *root@controller:~# ip addr show*
>>>>     1: lo: <LOOPBACK,UP,LOWER_UP> mtu 16436 qdisc noqueue state
>>>>     UNKNOWN 
>>>>         link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00
>>>>         inet 127.0.0.1/8 <http://127.0.0.1/8> scope host lo
>>>>         inet6 ::1/128 scope host 
>>>>            valid_lft forever preferred_lft forever
>>>>     2: eth0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc mq
>>>>     state UP qlen 1000
>>>>         link/ether d4:ae:52:bb:aa:20 brd ff:ff:ff:ff:ff:ff
>>>>         inet 10.10.10.1/24 <http://10.10.10.1/24> brd 10.10.10.255
>>>>     scope global eth0
>>>>         inet6 fe80::d6ae:52ff:febb:aa20/64 scope link 
>>>>            valid_lft forever preferred_lft forever
>>>>     3: eth1: <BROADCAST,MULTICAST> mtu 1500 qdisc noop state DOWN
>>>>     qlen 1000
>>>>         link/ether d4:ae:52:bb:aa:21 brd ff:ff:ff:ff:ff:ff
>>>>     4: eth0.2@eth0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500
>>>>     qdisc noqueue state UP 
>>>>         link/ether d4:ae:52:bb:aa:20 brd ff:ff:ff:ff:ff:ff
>>>>         inet 192.168.2.225/24 <http://192.168.2.225/24> brd
>>>>     192.168.2.255 scope global eth0.2
>>>>         inet6 fe80::d6ae:52ff:febb:aa20/64 scope link 
>>>>            valid_lft forever preferred_lft forever
>>>>     5: br-int: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc
>>>>     noqueue state UNKNOWN 
>>>>         link/ether ba:7a:e9:dc:2b:41 brd ff:ff:ff:ff:ff:ff
>>>>         inet6 fe80::b87a:e9ff:fedc:2b41/64 scope link 
>>>>            valid_lft forever preferred_lft forever
>>>>     7: br-ex: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc
>>>>     noqueue state UNKNOWN 
>>>>         link/ether 9a:41:c8:8a:9e:49 brd ff:ff:ff:ff:ff:ff
>>>>         inet 192.168.2.225/24 <http://192.168.2.225/24> scope
>>>>     global br-ex
>>>>     8: tapf71b5b86-5c: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500
>>>>     qdisc noqueue state UNKNOWN 
>>>>         link/ether 2a:44:a3:d1:7d:f3 brd ff:ff:ff:ff:ff:ff
>>>>         inet 10.5.5.2/24 <http://10.5.5.2/24> brd 10.5.5.255 scope
>>>>     global tapf71b5b86-5c
>>>>         inet6 fe80::2844:a3ff:fed1:7df3/64 scope link 
>>>>            valid_lft forever preferred_lft forever
>>>>     9: qr-4d088f3a-78: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500
>>>>     qdisc noqueue state UNKNOWN 
>>>>         link/ether ca:5b:8d:4d:6d:fb brd ff:ff:ff:ff:ff:ff
>>>>         inet 10.5.5.1/24 <http://10.5.5.1/24> brd 10.5.5.255 scope
>>>>     global qr-4d088f3a-78
>>>>         inet6 fe80::c85b:8dff:fe4d:6dfb/64 scope link 
>>>>            valid_lft forever preferred_lft forever
>>>>     10: qg-6f8374cb-cb: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500
>>>>     qdisc noqueue state UNKNOWN 
>>>>         link/ether 0e:7f:dd:3a:80:bc brd ff:ff:ff:ff:ff:ff
>>>>         inet 192.168.2.151/24 <http://192.168.2.151/24> brd
>>>>     192.168.2.255 scope global qg-6f8374cb-cb
>>>>         inet6 fe80::c7f:ddff:fe3a:80bc/64 scope link 
>>>>            valid_lft forever preferred_lft forever
>>>>     27: br-tun: <BROADCAST,MULTICAST> mtu 1500 qdisc noop state DOWN 
>>>>         link/ether 8a:cf:ec:7c:15:40 brd ff:ff:ff:ff:ff:ff
>>>>
>>>>     *cat /proc/sys/net/ipv4/ip_forward*
>>>>     1
>>>>
>>>>     *root@computenode:~# ip addr show*
>>>>     1: lo: <LOOPBACK,UP,LOWER_UP> mtu 16436 qdisc noqueue state
>>>>     UNKNOWN 
>>>>         link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00
>>>>         inet 127.0.0.1/8 <http://127.0.0.1/8> scope host lo
>>>>         inet6 ::1/128 scope host 
>>>>            valid_lft forever preferred_lft forever
>>>>     2: eth0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc mq
>>>>     state UP qlen 1000
>>>>         link/ether d4:ae:52:bb:a1:9d brd ff:ff:ff:ff:ff:ff
>>>>         inet 10.10.10.12/24 <http://10.10.10.12/24> brd
>>>>     10.10.10.255 scope global eth0
>>>>         inet6 fe80::d6ae:52ff:febb:a19d/64 scope link 
>>>>            valid_lft forever preferred_lft forever
>>>>     3: eth1: <BROADCAST,MULTICAST> mtu 1500 qdisc noop state DOWN
>>>>     qlen 1000
>>>>         link/ether d4:ae:52:bb:a1:9e brd ff:ff:ff:ff:ff:ff
>>>>     4: eth0.2@eth0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500
>>>>     qdisc noqueue state UP 
>>>>         link/ether d4:ae:52:bb:a1:9d brd ff:ff:ff:ff:ff:ff
>>>>         inet 192.168.2.234/24 <http://192.168.2.234/24> brd
>>>>     192.168.2.255 scope global eth0.2
>>>>         inet6 fe80::d6ae:52ff:febb:a19d/64 scope link 
>>>>            valid_lft forever preferred_lft forever
>>>>     5: br-int: <BROADCAST,MULTICAST> mtu 1500 qdisc noop state DOWN 
>>>>         link/ether ae:9b:43:09:af:40 brd ff:ff:ff:ff:ff:ff
>>>>     9: qbr256f5ed2-43: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500
>>>>     qdisc noqueue state UP 
>>>>         link/ether c6:c0:df:64:c6:99 brd ff:ff:ff:ff:ff:ff
>>>>         inet6 fe80::20e8:b9ff:fe6c:6f55/64 scope link 
>>>>            valid_lft forever preferred_lft forever
>>>>     10: qvo256f5ed2-43: <BROADCAST,MULTICAST,PROMISC,UP,LOWER_UP>
>>>>     mtu 1500 qdisc pfifo_fast state UP qlen 1000
>>>>         link/ether 76:25:8b:fd:90:3b brd ff:ff:ff:ff:ff:ff
>>>>         inet6 fe80::7425:8bff:fefd:903b/64 scope link 
>>>>            valid_lft forever preferred_lft forever
>>>>     11: qvb256f5ed2-43: <BROADCAST,MULTICAST,PROMISC,UP,LOWER_UP>
>>>>     mtu 1500 qdisc pfifo_fast master qbr256f5ed2-43 state UP qlen 1000
>>>>         link/ether c6:c0:df:64:c6:99 brd ff:ff:ff:ff:ff:ff
>>>>         inet6 fe80::c4c0:dfff:fe64:c699/64 scope link 
>>>>            valid_lft forever preferred_lft forever
>>>>     13: br-tun: <BROADCAST,MULTICAST> mtu 1500 qdisc noop state DOWN 
>>>>         link/ether be:8c:30:78:35:48 brd ff:ff:ff:ff:ff:ff
>>>>     15: vnet0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc
>>>>     pfifo_fast master qbr256f5ed2-43 state UNKNOWN qlen 500
>>>>         link/ether fe:16:3e:57:ec:ff brd ff:ff:ff:ff:ff:ff
>>>>         inet6 fe80::fc16:3eff:fe57:ecff/64 scope link 
>>>>            valid_lft forever preferred_lft forever
>>>>
>>>>     btw cronus is my compute node and I have renamed it to
>>>>     computenode to understand it better.
>>>>
>>>>     On 16 February 2013 22:11, Jean-Baptiste RANSY
>>>>     <jean-baptiste.ransy@xxxxxxxxxx
>>>>     <mailto:jean-baptiste.ransy@xxxxxxxxxx>> wrote:
>>>>
>>>>         Download in progress
>>>>
>>>>         Can you send me the output of those commands i forgot :
>>>>
>>>>         Controller Node:
>>>>         $ ip addr show
>>>>         $ cat /proc/sys/net/ipv4/ip_forward
>>>>
>>>>         Compute Node:
>>>>         $ ip addr show
>>>>
>>>>
>>>>         JB
>>>>
>>>>
>>>>
>>>>         On 02/16/2013 10:45 PM, Chathura M. Sarathchandra
>>>>         Magurawalage wrote:
>>>>>         Thanks Ransy,
>>>>>
>>>>>         I have created a tar file with the configuration and log
>>>>>         files in it. Please download it using the following URL. I
>>>>>         have pasted the output of the commands below.
>>>>>
>>>>>         https://www.dropbox.com/s/qyfcsn50060y304/confilesnlogs.tar
>>>>>
>>>>>         *Controller node:*
>>>>>         *root@controller:~# keystone endpoint-list*
>>>>>         +----------------------------------+-----------+-------------------------------------------------+-------------------------------------------------+--------------------------------------------+Controller
>>>>>         node
>>>>>         |                id                |   region  |          
>>>>>                  publicurl                    |                  
>>>>>         internalurl                   |                  adminurl
>>>>>                          |
>>>>>         +----------------------------------+-----------+-------------------------------------------------+-------------------------------------------------+--------------------------------------------+
>>>>>         | 2c9a1cb0fe8247d9b7716432cf459fe5 | RegionOne |  
>>>>>          http://192.168.2.225:8774/v2/$(tenant_id)s
>>>>>         <http://192.168.2.225:8774/v2/$%28tenant_id%29s>   |  
>>>>>          http://192.168.2.225:8774/v2/$(tenant_id)s
>>>>>         <http://192.168.2.225:8774/v2/$%28tenant_id%29s>   |
>>>>>         http://192.168.2.225:8774/v2/$(tenant_id)s
>>>>>         <http://192.168.2.225:8774/v2/$%28tenant_id%29s> |
>>>>>         | 2d306903ed3342a8aaaac7c5680c116f | RegionOne |          
>>>>>          http://192.168.2.225:9696/           |          
>>>>>          http://192.168.2.225:9696/           |        
>>>>>         http://192.168.2.225:9696/         |
>>>>>         | 3848114f120f42bf819bc2443b28ac9e | RegionOne |
>>>>>         http://192.168.2.225:8080/v1/AUTH_$(tenant_id)s
>>>>>         <http://192.168.2.225:8080/v1/AUTH_$%28tenant_id%29s> |
>>>>>         http://192.168.2.225:8080/v1/AUTH_$(tenant_id)s
>>>>>         <http://192.168.2.225:8080/v1/AUTH_$%28tenant_id%29s> |  
>>>>>              http://192.168.2.225:8080/v1        |
>>>>>         | 4955173b8d9e4d33ae4a5b29dc12c74d | RegionOne |  
>>>>>          http://192.168.2.225:8776/v1/$(tenant_id)s
>>>>>         <http://192.168.2.225:8776/v1/$%28tenant_id%29s>   |  
>>>>>          http://192.168.2.225:8776/v1/$(tenant_id)s
>>>>>         <http://192.168.2.225:8776/v1/$%28tenant_id%29s>   |
>>>>>         http://192.168.2.225:8776/v1/$(tenant_id)s
>>>>>         <http://192.168.2.225:8776/v1/$%28tenant_id%29s> |
>>>>>         | d313aa76bf854dde94f33a49a9f0c8ac | RegionOne |          
>>>>>         http://192.168.2.225:9292/v2          |          
>>>>>         http://192.168.2.225:9292/v2          |      
>>>>>          http://192.168.2.225:9292/v2        |
>>>>>         | e5aa4ecf3cbe4dd5aba9b204c74fee6a | RegionOne |        
>>>>>          http://192.168.2.225:5000/v2.0         |        
>>>>>          http://192.168.2.225:5000/v2.0         |    
>>>>>          http://192.168.2.225:35357/v2.0       |
>>>>>         | fba6f790e3b444c890d114f13cd32b37 | RegionOne |    
>>>>>         http://192.168.2.225:8773/services/Cloud    |    
>>>>>         http://192.168.2.225:8773/services/Cloud    |
>>>>>          http://192.168.2.225:8773/services/Admin  |
>>>>>         +----------------------------------+-----------+-------------------------------------------------+-------------------------------------------------+--------------------------------------------+
>>>>>
>>>>>         *root@controller:~# ip link show*
>>>>>         1: lo: <LOOPBACK,UP,LOWER_UP> mtu 16436 qdisc noqueue
>>>>>         state UNKNOWN 
>>>>>             link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00
>>>>>         2: eth0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc
>>>>>         mq state UP qlen 1000
>>>>>             link/ether d4:ae:52:bb:aa:20 brd ff:ff:ff:ff:ff:ff
>>>>>         3: eth1: <BROADCAST,MULTICAST> mtu 1500 qdisc noop state
>>>>>         DOWN qlen 1000
>>>>>             link/ether d4:ae:52:bb:aa:21 brd ff:ff:ff:ff:ff:ff
>>>>>         4: eth0.2@eth0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500
>>>>>         qdisc noqueue state UP 
>>>>>             link/ether d4:ae:52:bb:aa:20 brd ff:ff:ff:ff:ff:ff
>>>>>         5: br-int: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500
>>>>>         qdisc noqueue state UNKNOWN 
>>>>>             link/ether ba:7a:e9:dc:2b:41 brd ff:ff:ff:ff:ff:ff
>>>>>         7: br-ex: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc
>>>>>         noqueue state UNKNOWN 
>>>>>             link/ether 9a:41:c8:8a:9e:49 brd ff:ff:ff:ff:ff:ff
>>>>>         8: tapf71b5b86-5c: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu
>>>>>         1500 qdisc noqueue state UNKNOWN 
>>>>>             link/ether 2a:44:a3:d1:7d:f3 brd ff:ff:ff:ff:ff:ff
>>>>>         9: qr-4d088f3a-78: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu
>>>>>         1500 qdisc noqueue state UNKNOWN 
>>>>>             link/ether ca:5b:8d:4d:6d:fb brd ff:ff:ff:ff:ff:ff
>>>>>         10: qg-6f8374cb-cb: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu
>>>>>         1500 qdisc noqueue state UNKNOWN 
>>>>>             link/ether 0e:7f:dd:3a:80:bc brd ff:ff:ff:ff:ff:ff
>>>>>         27: br-tun: <BROADCAST,MULTICAST> mtu 1500 qdisc noop
>>>>>         state DOWN 
>>>>>             link/ether 8a:cf:ec:7c:15:40 brd ff:ff:ff:ff:ff:ff
>>>>>
>>>>>         *root@controller:~# ip route show*
>>>>>         default via 192.168.2.253 dev eth0.2 
>>>>>         default via 192.168.2.253 dev eth0.2  metric 100 
>>>>>         10.5.5.0/24 <http://10.5.5.0/24> dev tapf71b5b86-5c  proto
>>>>>         kernel  scope link  src 10.5.5.2 
>>>>>         10.5.5.0/24 <http://10.5.5.0/24> dev qr-4d088f3a-78  proto
>>>>>         kernel  scope link  src 10.5.5.1 
>>>>>         10.10.10.0/24 <http://10.10.10.0/24> dev eth0  proto
>>>>>         kernel  scope link  src 10.10.10.1 
>>>>>         192.168.2.0/24 <http://192.168.2.0/24> dev eth0.2  proto
>>>>>         kernel  scope link  src 192.168.2.225 
>>>>>         192.168.2.0/24 <http://192.168.2.0/24> dev qg-6f8374cb-cb
>>>>>          proto kernel  scope link  src 192.168.2.151 
>>>>>         192.168.2.0/24 <http://192.168.2.0/24> dev br-ex  proto
>>>>>         kernel  scope link  src 192.168.2.225
>>>>>
>>>>>         *$ ip netns show (Did not return anything)*
>>>>>
>>>>>         *root@controller:~# ovs-vsctl show*
>>>>>         a566afae-d7a8-42a9-aefe-8b0f2f7054a3
>>>>>             Bridge br-tun
>>>>>                 Port "gre-4"
>>>>>                     Interface "gre-4"
>>>>>                         type: gre
>>>>>                         options: {in_key=flow, out_key=flow,
>>>>>         remote_ip="10.10.10.12"}
>>>>>                 Port "gre-3"
>>>>>                     Interface "gre-3"
>>>>>                         type: gre
>>>>>                         options: {in_key=flow, out_key=flow,
>>>>>         remote_ip="127.0.0.1"}
>>>>>                 Port patch-int
>>>>>                     Interface patch-int
>>>>>                         type: patch
>>>>>                         options: {peer=patch-tun}
>>>>>                 Port br-tun
>>>>>                     Interface br-tun
>>>>>                         type: internal
>>>>>                 Port "gre-1"
>>>>>                     Interface "gre-1"
>>>>>                         type: gre
>>>>>                         options: {in_key=flow, out_key=flow,
>>>>>         remote_ip="10.0.0.3"}
>>>>>             Bridge br-ex
>>>>>                 Port br-ex
>>>>>                     Interface br-ex
>>>>>                         type: internal
>>>>>                 Port "qg-6f8374cb-cb"
>>>>>                     Interface "qg-6f8374cb-cb"
>>>>>                         type: internal
>>>>>                 Port "br0"
>>>>>                     Interface "br0"
>>>>>             Bridge br-int
>>>>>                 Port br-int
>>>>>                     Interface br-int
>>>>>                         type: internal
>>>>>                 Port "tapf71b5b86-5c"
>>>>>                     tag: 1
>>>>>                     Interface "tapf71b5b86-5c"
>>>>>                         type: internal
>>>>>                 Port patch-tun
>>>>>                     Interface patch-tun
>>>>>                         type: patch
>>>>>                         options: {peer=patch-int}
>>>>>                 Port "qr-4d088f3a-78"
>>>>>                     tag: 1
>>>>>                     Interface "qr-4d088f3a-78"
>>>>>                         type: internal
>>>>>             ovs_version: "1.4.0+build0"
>>>>>
>>>>>
>>>>>         *Compute node:*
>>>>>
>>>>>         *root@cronus:~# ip link show*
>>>>>         1: lo: <LOOPBACK,UP,LOWER_UP> mtu 16436 qdisc noqueue
>>>>>         state UNKNOWN 
>>>>>             link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00
>>>>>         2: eth0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc
>>>>>         mq state UP qlen 1000
>>>>>             link/ether d4:ae:52:bb:a1:9d brd ff:ff:ff:ff:ff:ff
>>>>>         3: eth1: <BROADCAST,MULTICAST> mtu 1500 qdisc noop state
>>>>>         DOWN qlen 1000
>>>>>             link/ether d4:ae:52:bb:a1:9e brd ff:ff:ff:ff:ff:ff
>>>>>         4: eth0.2@eth0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500
>>>>>         qdisc noqueue state UP 
>>>>>             link/ether d4:ae:52:bb:a1:9d brd ff:ff:ff:ff:ff:ff
>>>>>         5: br-int: <BROADCAST,MULTICAST> mtu 1500 qdisc noop state
>>>>>         DOWN 
>>>>>             link/ether ae:9b:43:09:af:40 brd ff:ff:ff:ff:ff:ff
>>>>>         9: qbr256f5ed2-43: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu
>>>>>         1500 qdisc noqueue state UP 
>>>>>             link/ether c6:c0:df:64:c6:99 brd ff:ff:ff:ff:ff:ff
>>>>>         10: qvo256f5ed2-43:
>>>>>         <BROADCAST,MULTICAST,PROMISC,UP,LOWER_UP> mtu 1500 qdisc
>>>>>         pfifo_fast state UP qlen 1000
>>>>>             link/ether 76:25:8b:fd:90:3b brd ff:ff:ff:ff:ff:ff
>>>>>         11: qvb256f5ed2-43:
>>>>>         <BROADCAST,MULTICAST,PROMISC,UP,LOWER_UP> mtu 1500 qdisc
>>>>>         pfifo_fast master qbr256f5ed2-43 state UP qlen 1000
>>>>>             link/ether c6:c0:df:64:c6:99 brd ff:ff:ff:ff:ff:ff
>>>>>         13: br-tun: <BROADCAST,MULTICAST> mtu 1500 qdisc noop
>>>>>         state DOWN 
>>>>>             link/ether be:8c:30:78:35:48 brd ff:ff:ff:ff:ff:ff
>>>>>         15: vnet0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500
>>>>>         qdisc pfifo_fast master qbr256f5ed2-43 state UNKNOWN qlen 500
>>>>>             link/ether fe:16:3e:57:ec:ff brd ff:ff:ff:ff:ff:ff
>>>>>
>>>>>         *root@cronus:~# ip route show*
>>>>>         default via 192.168.2.253 dev eth0.2  metric 100 
>>>>>         10.10.10.0/24 <http://10.10.10.0/24> dev eth0  proto
>>>>>         kernel  scope link  src 10.10.10.12 
>>>>>         192.168.2.0/24 <http://192.168.2.0/24> dev eth0.2  proto
>>>>>         kernel  scope link  src 192.168.2.234 
>>>>>
>>>>>         *root@cronus:~# ovs-vsctl show*
>>>>>         d85bc334-6d64-4a13-b851-d56b18ff1549
>>>>>             Bridge br-int
>>>>>                 Port "qvo0e743b01-89"
>>>>>                     tag: 4095
>>>>>                     Interface "qvo0e743b01-89"
>>>>>                 Port "qvo256f5ed2-43"
>>>>>                     tag: 1
>>>>>                     Interface "qvo256f5ed2-43"
>>>>>                 Port patch-tun
>>>>>                     Interface patch-tun
>>>>>                         type: patch
>>>>>                         options: {peer=patch-int}
>>>>>                 Port br-int
>>>>>                     Interface br-int
>>>>>                         type: internal
>>>>>                 Port "qvoee3d4131-2a"
>>>>>                     tag: 4095
>>>>>                     Interface "qvoee3d4131-2a"
>>>>>                 Port "qvocbc816bd-3d"
>>>>>                     tag: 4095
>>>>>                     Interface "qvocbc816bd-3d"
>>>>>             Bridge br-tun
>>>>>                 Port br-tun
>>>>>                     Interface br-tun
>>>>>                         type: internal
>>>>>                 Port "gre-2"
>>>>>                     Interface "gre-2"
>>>>>                         type: gre
>>>>>                         options: {in_key=flow, out_key=flow,
>>>>>         remote_ip="10.10.10.1"}
>>>>>                 Port "gre-1"
>>>>>                     Interface "gre-1"
>>>>>                         type: gre
>>>>>                         options: {in_key=flow, out_key=flow,
>>>>>         remote_ip="10.0.0.3"}
>>>>>                 Port patch-int
>>>>>                     Interface patch-int
>>>>>                         type: patch
>>>>>                         options: {peer=patch-tun}
>>>>>                 Port "gre-3"
>>>>>                     Interface "gre-3"
>>>>>                         type: gre
>>>>>                         options: {in_key=flow, out_key=flow,
>>>>>         remote_ip="127.0.0.1"}
>>>>>             ovs_version: "1.4.0+build0"
>>>>>
>>>>>
>>>>>         Thanks I appreciate your help.
>>>>>
>>>>>         On 16 February 2013 16:49, Jean-Baptiste RANSY
>>>>>         <jean-baptiste.ransy@xxxxxxxxxx
>>>>>         <mailto:jean-baptiste.ransy@xxxxxxxxxx>> wrote:
>>>>>
>>>>>             Please provide files listed bellow :
>>>>>
>>>>>             Controller Node :
>>>>>             /etc/nova/nova.conf
>>>>>             /etc/nova/api-paste.ini
>>>>>             /etc/quantum/l3_agent.ini
>>>>>             /etc/quantum/quantum.conf
>>>>>             /etc/quantum/dhcp_agent.ini
>>>>>             /etc/quantum/plugins/openvswitch/ovs_quantum_plugin.ini
>>>>>             /etc/quantum/api-paste.ini
>>>>>             /var/log/nova/*.log
>>>>>             /var/log/quantum/*.log
>>>>>
>>>>>             Compute Node :
>>>>>             /etc/nova/nova.conf
>>>>>             /etc/nova/nova-compute.conf
>>>>>             /etc/nova/api-paste.ini
>>>>>             /etc/quantum/quantum.conf
>>>>>             /etc/quantum/plugins/openvswitch/ovs_quantum_plugin.ini
>>>>>             /var/log/nova/*.log
>>>>>             /var/log/quantum/*.log
>>>>>
>>>>>             Plus, complete output of the following commands :
>>>>>
>>>>>             Controller Node :
>>>>>             $ keystone endpoint-list
>>>>>             $ ip link show
>>>>>             $ ip route show
>>>>>             $ ip netns show
>>>>>             $ ovs-vsctl show
>>>>>
>>>>>             Compute Node :
>>>>>             $ ip link show
>>>>>             $ ip route show
>>>>>             $ ovs-vsctl show
>>>>>
>>>>>             Regards,
>>>>>
>>>>>             Jean-Baptiste RANSY
>>>>>
>>>>>
>>>>>
>>>>>             On 02/16/2013 05:32 PM, Chathura M. Sarathchandra
>>>>>             Magurawalage wrote:
>>>>>>             Hello Jean,
>>>>>>
>>>>>>             Thanks for your reply.
>>>>>>
>>>>>>             I followed the instructions
>>>>>>             in http://docs.openstack.org/folsom/basic-install/content/basic-install_network.html.
>>>>>>             And my Controller and the Network-node is installed
>>>>>>             in the same physical node.
>>>>>>
>>>>>>             I am using Folsom but without Network namespaces. 
>>>>>>
>>>>>>             But in the website you have provided it states that
>>>>>>             "If you run both L3 + DHCP services on the same node,
>>>>>>             you should enable namespaces to avoid conflicts with
>>>>>>             routes :"
>>>>>>
>>>>>>             But currently quantum-dhcp-agent and quantum-l3-agent
>>>>>>             are running in the same node? 
>>>>>>
>>>>>>             Additionally the control node serves as a DHCP server
>>>>>>             for the local network ( Don't know if that would make
>>>>>>             and difference)
>>>>>>
>>>>>>             Any idea what the problem could be?
>>>>>>
>>>>>>
>>>>>>             On 16 February 2013 16:21, Jean-Baptiste RANSY
>>>>>>             <jean-baptiste.ransy@xxxxxxxxxx
>>>>>>             <mailto:jean-baptiste.ransy@xxxxxxxxxx>> wrote:
>>>>>>
>>>>>>                 Hello Chathura,
>>>>>>
>>>>>>                 Are you using Folsom with Network Namespaces ?
>>>>>>
>>>>>>                 If yes, have a look here :
>>>>>>                 http://docs.openstack.org/folsom/openstack-network/admin/content/ch_limitations.html
>>>>>>
>>>>>>
>>>>>>                 Regards,
>>>>>>
>>>>>>                 Jean-Baptsite RANSY
>>>>>>
>>>>>>
>>>>>>
>>>>>>                 On 02/16/2013 05:01 PM, Chathura M. Sarathchandra
>>>>>>                 Magurawalage wrote:
>>>>>>>                 Hello guys,
>>>>>>>
>>>>>>>                 The problem still exists. Any ideas?
>>>>>>>
>>>>>>>                 Thanks 
>>>>>>>
>>>>>>>                 On 15 February 2013 14:37, Sylvain Bauza
>>>>>>>                 <sylvain.bauza@xxxxxxxxxxxx
>>>>>>>                 <mailto:sylvain.bauza@xxxxxxxxxxxx>> wrote:
>>>>>>>
>>>>>>>                     Metadata API allows to fetch SSH credentials
>>>>>>>                     when booting (pubkey I mean).
>>>>>>>                     If a VM is unable to reach metadata service,
>>>>>>>                     then it won't be able to get its public key,
>>>>>>>                     so you won't be able to connect, unless you
>>>>>>>                     specifically go thru a Password
>>>>>>>                     authentication (provided password auth is
>>>>>>>                     enabled in /etc/ssh/sshd_config, which is
>>>>>>>                     not the case with Ubuntu cloud archive).
>>>>>>>                     There is also a side effect, the boot
>>>>>>>                     process is longer as the instance is waiting
>>>>>>>                     for the curl timeout (60sec.) to finish
>>>>>>>                     booting up.
>>>>>>>
>>>>>>>                     Re: Quantum, the metadata API is actually
>>>>>>>                     DNAT'd from Network node to the Nova-api
>>>>>>>                     node (here 172.16.0.1 as internal management
>>>>>>>                     IP) :
>>>>>>>                     Chain quantum-l3-agent-PREROUTING (1
>>>>>>>                     references)
>>>>>>>
>>>>>>>                     target     prot opt source              
>>>>>>>                     destination
>>>>>>>                     DNAT       tcp  --  0.0.0.0/0
>>>>>>>                     <http://0.0.0.0/0>          
>>>>>>>                      169.254.169.254      tcp dpt:80
>>>>>>>                     to:172.16.0.1:8775 <http://172.16.0.1:8775>
>>>>>>>
>>>>>>>
>>>>>>>                     Anyway, the first step is to :
>>>>>>>                     1. grab the console.log
>>>>>>>                     2. access thru VNC to the desired instance
>>>>>>>
>>>>>>>                     Troubleshooting will be easier once that done.
>>>>>>>
>>>>>>>                     -Sylvain
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>                     Le 15/02/2013 14:24, Chathura M.
>>>>>>>                     Sarathchandra Magurawalage a écrit :
>>>>>>>
>>>>>>>                         Hello Guys,
>>>>>>>
>>>>>>>                         Not sure if this is the right port but
>>>>>>>                         these are the results:
>>>>>>>
>>>>>>>                         *Compute node:*
>>>>>>>
>>>>>>>
>>>>>>>                         root@computenode:~# netstat -an | grep 8775
>>>>>>>                         tcp        0      0 0.0.0.0:8775
>>>>>>>                         <http://0.0.0.0:8775>
>>>>>>>                         <http://0.0.0.0:8775>  0.0.0.0:*        
>>>>>>>                               LISTEN
>>>>>>>
>>>>>>>                         *Controller: *
>>>>>>>
>>>>>>>
>>>>>>>                         root@controller:~# netstat -an | grep 8775
>>>>>>>                         tcp        0      0 0.0.0.0:8775
>>>>>>>                         <http://0.0.0.0:8775>
>>>>>>>                         <http://0.0.0.0:8775>  0.0.0.0:*        
>>>>>>>                               LISTEN
>>>>>>>
>>>>>>>                         *Additionally I cant curl
>>>>>>>                         169.254.169.254 from the compute node. I
>>>>>>>                         am not sure if this is related to not
>>>>>>>                         being able to PING the VM.*
>>>>>>>
>>>>>>>
>>>>>>>                         curl -v http://169.254.169.254
>>>>>>>                         * About to connect() to 169.254.169.254
>>>>>>>                         port 80 (#0)
>>>>>>>                         *   Trying 169.254.169.254...
>>>>>>>
>>>>>>>                         Thanks for your help
>>>>>>>
>>>>>>>
>>>>>>>                         -----------------------------------------------------------------------------------------------------------------------------
>>>>>>>                         Chathura Madhusanka Sarathchandra
>>>>>>>                         Magurawalage.
>>>>>>>                         1NW.2.1, Desk 2
>>>>>>>                         School of Computer Science and
>>>>>>>                         Electronic Engineering
>>>>>>>                         University Of Essex
>>>>>>>                         United Kingdom.
>>>>>>>
>>>>>>>                         Email: csarata@xxxxxxxxxxx
>>>>>>>                         <mailto:csarata@xxxxxxxxxxx>
>>>>>>>                         <mailto:csarata@xxxxxxxxxxx
>>>>>>>                         <mailto:csarata@xxxxxxxxxxx>>
>>>>>>>                                  
>>>>>>>                         chathura.sarathchandra@xxxxxxxxx
>>>>>>>                         <mailto:chathura.sarathchandra@xxxxxxxxx> <mailto:77.chathura@xxxxxxxxx
>>>>>>>                         <mailto:77.chathura@xxxxxxxxx>>
>>>>>>>                         77.chathura@xxxxxxxxx
>>>>>>>                         <mailto:77.chathura@xxxxxxxxx>
>>>>>>>                         <mailto:77.chathura@xxxxxxxxx
>>>>>>>                         <mailto:77.chathura@xxxxxxxxx>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>                         On 15 February 2013 11:03, Anil Vishnoi
>>>>>>>                         <vishnoianil@xxxxxxxxx
>>>>>>>                         <mailto:vishnoianil@xxxxxxxxx>
>>>>>>>                         <mailto:vishnoianil@xxxxxxxxx
>>>>>>>                         <mailto:vishnoianil@xxxxxxxxx>>> wrote:
>>>>>>>
>>>>>>>                             If you are using ubuntu cloud image
>>>>>>>                         then the only way to log-in is
>>>>>>>                             to do ssh with the public key. For
>>>>>>>                         that you have to create ssh key
>>>>>>>                             pair and download the ssh key. You
>>>>>>>                         can create this ssh pair using
>>>>>>>                             horizon/cli.
>>>>>>>
>>>>>>>
>>>>>>>                             On Fri, Feb 15, 2013 at 4:27 PM,
>>>>>>>                         Sylvain Bauza
>>>>>>>                             <sylvain.bauza@xxxxxxxxxxxx
>>>>>>>                         <mailto:sylvain.bauza@xxxxxxxxxxxx>
>>>>>>>                         <mailto:sylvain.bauza@xxxxxxxxxxxx
>>>>>>>                         <mailto:sylvain.bauza@xxxxxxxxxxxx>>>
>>>>>>>
>>>>>>>                             wrote:
>>>>>>>
>>>>>>>
>>>>>>>                                 Le 15/02/2013 11:42, Chathura M.
>>>>>>>                         Sarathchandra Magurawalage a
>>>>>>>                                 écrit :
>>>>>>>
>>>>>>>
>>>>>>>                                     How can I log into the VM
>>>>>>>                         from VNC? What are the credentials?
>>>>>>>
>>>>>>>
>>>>>>>                                 You have multiple ways to get
>>>>>>>                         VNC access. The easiest one is
>>>>>>>                                 thru Horizon. Other can be
>>>>>>>                         looking at the KVM command-line for
>>>>>>>                                 the desired instance (on the
>>>>>>>                         compute node) and check the vnc
>>>>>>>                                 port in use (assuming KVM as
>>>>>>>                         hypervisor).
>>>>>>>                                 This is basic knowledge of Nova.
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>                                     nova-api-metadata is running
>>>>>>>                         fine in the compute node.
>>>>>>>
>>>>>>>
>>>>>>>                                 Make sure the metadata port is
>>>>>>>                         avaible thanks to telnet or
>>>>>>>                                 netstat, nova-api can be running
>>>>>>>                         without listening on metadata
>>>>>>>                                 port.
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>                                
>>>>>>>                         _______________________________________________
>>>>>>>                                 Mailing list:
>>>>>>>                         https://launchpad.net/~openstack
>>>>>>>                         <https://launchpad.net/%7Eopenstack>
>>>>>>>                                 <https://launchpad.net/%7Eopenstack>
>>>>>>>                                 Post to     :
>>>>>>>                         openstack@xxxxxxxxxxxxxxxxxxx
>>>>>>>                         <mailto:openstack@xxxxxxxxxxxxxxxxxxx>
>>>>>>>                                
>>>>>>>                         <mailto:openstack@xxxxxxxxxxxxxxxxxxx
>>>>>>>                         <mailto:openstack@xxxxxxxxxxxxxxxxxxx>>
>>>>>>>                                 Unsubscribe :
>>>>>>>                         https://launchpad.net/~openstack
>>>>>>>                         <https://launchpad.net/%7Eopenstack>
>>>>>>>                                
>>>>>>>                         <https://launchpad.net/%7Eopenstack>
>>>>>>>
>>>>>>>                                 More help   :
>>>>>>>                         https://help.launchpad.net/ListHelp
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>                             --     Thanks & Regards
>>>>>>>                             --Anil Kumar Vishnoi
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>                 _______________________________________________
>>>>>>>                 Mailing list: https://launchpad.net/~openstack
>>>>>>>                 <https://launchpad.net/%7Eopenstack>
>>>>>>>                 Post to : openstack@xxxxxxxxxxxxxxxxxxx
>>>>>>>                 <mailto:openstack@xxxxxxxxxxxxxxxxxxx>
>>>>>>>                 Unsubscribe : https://launchpad.net/~openstack
>>>>>>>                 <https://launchpad.net/%7Eopenstack> More help :
>>>>>>>                 https://help.launchpad.net/ListHelp
>>>>>>
>>>>>>
>>>>>
>>>>>
>>>>
>>>>
>>>
>>
>
>


Follow ups

References