← Back to team overview

openstack team mailing list archive

Re: Cant ping private or floating IP

 

Download in progress

Can you send me the output of those commands i forgot :

Controller Node:
$ ip addr show
$ cat /proc/sys/net/ipv4/ip_forward

Compute Node:
$ ip addr show


JB


On 02/16/2013 10:45 PM, Chathura M. Sarathchandra Magurawalage wrote:
> Thanks Ransy,
>
> I have created a tar file with the configuration and log files in it.
> Please download it using the following URL. I have pasted the output
> of the commands below.
>
> https://www.dropbox.com/s/qyfcsn50060y304/confilesnlogs.tar
>
> *Controller node:*
> *root@controller:~# keystone endpoint-list*
> +----------------------------------+-----------+-------------------------------------------------+-------------------------------------------------+--------------------------------------------+Controller
> node
> |                id                |   region  |                  
>  publicurl                    |                   internalurl        
>           |                  adminurl                  |
> +----------------------------------+-----------+-------------------------------------------------+-------------------------------------------------+--------------------------------------------+
> | 2c9a1cb0fe8247d9b7716432cf459fe5 | RegionOne |  
>  http://192.168.2.225:8774/v2/$(tenant_id)s
> <http://192.168.2.225:8774/v2/$%28tenant_id%29s>   |  
>  http://192.168.2.225:8774/v2/$(tenant_id)s
> <http://192.168.2.225:8774/v2/$%28tenant_id%29s>   |
> http://192.168.2.225:8774/v2/$(tenant_id)s
> <http://192.168.2.225:8774/v2/$%28tenant_id%29s> |
> | 2d306903ed3342a8aaaac7c5680c116f | RegionOne |          
>  http://192.168.2.225:9696/           |          
>  http://192.168.2.225:9696/           |        
> http://192.168.2.225:9696/         |
> | 3848114f120f42bf819bc2443b28ac9e | RegionOne |
> http://192.168.2.225:8080/v1/AUTH_$(tenant_id)s
> <http://192.168.2.225:8080/v1/AUTH_$%28tenant_id%29s> |
> http://192.168.2.225:8080/v1/AUTH_$(tenant_id)s
> <http://192.168.2.225:8080/v1/AUTH_$%28tenant_id%29s> |      
>  http://192.168.2.225:8080/v1        |
> | 4955173b8d9e4d33ae4a5b29dc12c74d | RegionOne |  
>  http://192.168.2.225:8776/v1/$(tenant_id)s
> <http://192.168.2.225:8776/v1/$%28tenant_id%29s>   |  
>  http://192.168.2.225:8776/v1/$(tenant_id)s
> <http://192.168.2.225:8776/v1/$%28tenant_id%29s>   |
> http://192.168.2.225:8776/v1/$(tenant_id)s
> <http://192.168.2.225:8776/v1/$%28tenant_id%29s> |
> | d313aa76bf854dde94f33a49a9f0c8ac | RegionOne |          
> http://192.168.2.225:9292/v2          |          
> http://192.168.2.225:9292/v2          |      
>  http://192.168.2.225:9292/v2        |
> | e5aa4ecf3cbe4dd5aba9b204c74fee6a | RegionOne |        
>  http://192.168.2.225:5000/v2.0         |        
>  http://192.168.2.225:5000/v2.0         |    
>  http://192.168.2.225:35357/v2.0       |
> | fba6f790e3b444c890d114f13cd32b37 | RegionOne |    
> http://192.168.2.225:8773/services/Cloud    |    
> http://192.168.2.225:8773/services/Cloud    |
>  http://192.168.2.225:8773/services/Admin  |
> +----------------------------------+-----------+-------------------------------------------------+-------------------------------------------------+--------------------------------------------+
>
> *root@controller:~# ip link show*
> 1: lo: <LOOPBACK,UP,LOWER_UP> mtu 16436 qdisc noqueue state UNKNOWN 
>     link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00
> 2: eth0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc mq state UP
> qlen 1000
>     link/ether d4:ae:52:bb:aa:20 brd ff:ff:ff:ff:ff:ff
> 3: eth1: <BROADCAST,MULTICAST> mtu 1500 qdisc noop state DOWN qlen 1000
>     link/ether d4:ae:52:bb:aa:21 brd ff:ff:ff:ff:ff:ff
> 4: eth0.2@eth0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc
> noqueue state UP 
>     link/ether d4:ae:52:bb:aa:20 brd ff:ff:ff:ff:ff:ff
> 5: br-int: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc noqueue
> state UNKNOWN 
>     link/ether ba:7a:e9:dc:2b:41 brd ff:ff:ff:ff:ff:ff
> 7: br-ex: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc noqueue
> state UNKNOWN 
>     link/ether 9a:41:c8:8a:9e:49 brd ff:ff:ff:ff:ff:ff
> 8: tapf71b5b86-5c: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc
> noqueue state UNKNOWN 
>     link/ether 2a:44:a3:d1:7d:f3 brd ff:ff:ff:ff:ff:ff
> 9: qr-4d088f3a-78: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc
> noqueue state UNKNOWN 
>     link/ether ca:5b:8d:4d:6d:fb brd ff:ff:ff:ff:ff:ff
> 10: qg-6f8374cb-cb: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc
> noqueue state UNKNOWN 
>     link/ether 0e:7f:dd:3a:80:bc brd ff:ff:ff:ff:ff:ff
> 27: br-tun: <BROADCAST,MULTICAST> mtu 1500 qdisc noop state DOWN 
>     link/ether 8a:cf:ec:7c:15:40 brd ff:ff:ff:ff:ff:ff
>
> *root@controller:~# ip route show*
> default via 192.168.2.253 dev eth0.2 
> default via 192.168.2.253 dev eth0.2  metric 100 
> 10.5.5.0/24 <http://10.5.5.0/24> dev tapf71b5b86-5c  proto kernel
>  scope link  src 10.5.5.2 
> 10.5.5.0/24 <http://10.5.5.0/24> dev qr-4d088f3a-78  proto kernel
>  scope link  src 10.5.5.1 
> 10.10.10.0/24 <http://10.10.10.0/24> dev eth0  proto kernel  scope
> link  src 10.10.10.1 
> 192.168.2.0/24 <http://192.168.2.0/24> dev eth0.2  proto kernel  scope
> link  src 192.168.2.225 
> 192.168.2.0/24 <http://192.168.2.0/24> dev qg-6f8374cb-cb  proto
> kernel  scope link  src 192.168.2.151 
> 192.168.2.0/24 <http://192.168.2.0/24> dev br-ex  proto kernel  scope
> link  src 192.168.2.225
>
> *$ ip netns show (Did not return anything)*
>
> *root@controller:~# ovs-vsctl show*
> a566afae-d7a8-42a9-aefe-8b0f2f7054a3
>     Bridge br-tun
>         Port "gre-4"
>             Interface "gre-4"
>                 type: gre
>                 options: {in_key=flow, out_key=flow,
> remote_ip="10.10.10.12"}
>         Port "gre-3"
>             Interface "gre-3"
>                 type: gre
>                 options: {in_key=flow, out_key=flow,
> remote_ip="127.0.0.1"}
>         Port patch-int
>             Interface patch-int
>                 type: patch
>                 options: {peer=patch-tun}
>         Port br-tun
>             Interface br-tun
>                 type: internal
>         Port "gre-1"
>             Interface "gre-1"
>                 type: gre
>                 options: {in_key=flow, out_key=flow, remote_ip="10.0.0.3"}
>     Bridge br-ex
>         Port br-ex
>             Interface br-ex
>                 type: internal
>         Port "qg-6f8374cb-cb"
>             Interface "qg-6f8374cb-cb"
>                 type: internal
>         Port "br0"
>             Interface "br0"
>     Bridge br-int
>         Port br-int
>             Interface br-int
>                 type: internal
>         Port "tapf71b5b86-5c"
>             tag: 1
>             Interface "tapf71b5b86-5c"
>                 type: internal
>         Port patch-tun
>             Interface patch-tun
>                 type: patch
>                 options: {peer=patch-int}
>         Port "qr-4d088f3a-78"
>             tag: 1
>             Interface "qr-4d088f3a-78"
>                 type: internal
>     ovs_version: "1.4.0+build0"
>
>
> *Compute node:*
>
> *root@cronus:~# ip link show*
> 1: lo: <LOOPBACK,UP,LOWER_UP> mtu 16436 qdisc noqueue state UNKNOWN 
>     link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00
> 2: eth0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc mq state UP
> qlen 1000
>     link/ether d4:ae:52:bb:a1:9d brd ff:ff:ff:ff:ff:ff
> 3: eth1: <BROADCAST,MULTICAST> mtu 1500 qdisc noop state DOWN qlen 1000
>     link/ether d4:ae:52:bb:a1:9e brd ff:ff:ff:ff:ff:ff
> 4: eth0.2@eth0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc
> noqueue state UP 
>     link/ether d4:ae:52:bb:a1:9d brd ff:ff:ff:ff:ff:ff
> 5: br-int: <BROADCAST,MULTICAST> mtu 1500 qdisc noop state DOWN 
>     link/ether ae:9b:43:09:af:40 brd ff:ff:ff:ff:ff:ff
> 9: qbr256f5ed2-43: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc
> noqueue state UP 
>     link/ether c6:c0:df:64:c6:99 brd ff:ff:ff:ff:ff:ff
> 10: qvo256f5ed2-43: <BROADCAST,MULTICAST,PROMISC,UP,LOWER_UP> mtu 1500
> qdisc pfifo_fast state UP qlen 1000
>     link/ether 76:25:8b:fd:90:3b brd ff:ff:ff:ff:ff:ff
> 11: qvb256f5ed2-43: <BROADCAST,MULTICAST,PROMISC,UP,LOWER_UP> mtu 1500
> qdisc pfifo_fast master qbr256f5ed2-43 state UP qlen 1000
>     link/ether c6:c0:df:64:c6:99 brd ff:ff:ff:ff:ff:ff
> 13: br-tun: <BROADCAST,MULTICAST> mtu 1500 qdisc noop state DOWN 
>     link/ether be:8c:30:78:35:48 brd ff:ff:ff:ff:ff:ff
> 15: vnet0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc pfifo_fast
> master qbr256f5ed2-43 state UNKNOWN qlen 500
>     link/ether fe:16:3e:57:ec:ff brd ff:ff:ff:ff:ff:ff
>
> *root@cronus:~# ip route show*
> default via 192.168.2.253 dev eth0.2  metric 100 
> 10.10.10.0/24 <http://10.10.10.0/24> dev eth0  proto kernel  scope
> link  src 10.10.10.12 
> 192.168.2.0/24 <http://192.168.2.0/24> dev eth0.2  proto kernel  scope
> link  src 192.168.2.234 
>
> *root@cronus:~# ovs-vsctl show*
> d85bc334-6d64-4a13-b851-d56b18ff1549
>     Bridge br-int
>         Port "qvo0e743b01-89"
>             tag: 4095
>             Interface "qvo0e743b01-89"
>         Port "qvo256f5ed2-43"
>             tag: 1
>             Interface "qvo256f5ed2-43"
>         Port patch-tun
>             Interface patch-tun
>                 type: patch
>                 options: {peer=patch-int}
>         Port br-int
>             Interface br-int
>                 type: internal
>         Port "qvoee3d4131-2a"
>             tag: 4095
>             Interface "qvoee3d4131-2a"
>         Port "qvocbc816bd-3d"
>             tag: 4095
>             Interface "qvocbc816bd-3d"
>     Bridge br-tun
>         Port br-tun
>             Interface br-tun
>                 type: internal
>         Port "gre-2"
>             Interface "gre-2"
>                 type: gre
>                 options: {in_key=flow, out_key=flow,
> remote_ip="10.10.10.1"}
>         Port "gre-1"
>             Interface "gre-1"
>                 type: gre
>                 options: {in_key=flow, out_key=flow, remote_ip="10.0.0.3"}
>         Port patch-int
>             Interface patch-int
>                 type: patch
>                 options: {peer=patch-tun}
>         Port "gre-3"
>             Interface "gre-3"
>                 type: gre
>                 options: {in_key=flow, out_key=flow,
> remote_ip="127.0.0.1"}
>     ovs_version: "1.4.0+build0"
>
>
> Thanks I appreciate your help.
>
> On 16 February 2013 16:49, Jean-Baptiste RANSY
> <jean-baptiste.ransy@xxxxxxxxxx
> <mailto:jean-baptiste.ransy@xxxxxxxxxx>> wrote:
>
>     Please provide files listed bellow :
>
>     Controller Node :
>     /etc/nova/nova.conf
>     /etc/nova/api-paste.ini
>     /etc/quantum/l3_agent.ini
>     /etc/quantum/quantum.conf
>     /etc/quantum/dhcp_agent.ini
>     /etc/quantum/plugins/openvswitch/ovs_quantum_plugin.ini
>     /etc/quantum/api-paste.ini
>     /var/log/nova/*.log
>     /var/log/quantum/*.log
>
>     Compute Node :
>     /etc/nova/nova.conf
>     /etc/nova/nova-compute.conf
>     /etc/nova/api-paste.ini
>     /etc/quantum/quantum.conf
>     /etc/quantum/plugins/openvswitch/ovs_quantum_plugin.ini
>     /var/log/nova/*.log
>     /var/log/quantum/*.log
>
>     Plus, complete output of the following commands :
>
>     Controller Node :
>     $ keystone endpoint-list
>     $ ip link show
>     $ ip route show
>     $ ip netns show
>     $ ovs-vsctl show
>
>     Compute Node :
>     $ ip link show
>     $ ip route show
>     $ ovs-vsctl show
>
>     Regards,
>
>     Jean-Baptiste RANSY
>
>
>
>     On 02/16/2013 05:32 PM, Chathura M. Sarathchandra Magurawalage wrote:
>>     Hello Jean,
>>
>>     Thanks for your reply.
>>
>>     I followed the instructions
>>     in http://docs.openstack.org/folsom/basic-install/content/basic-install_network.html.
>>     And my Controller and the Network-node is installed in the same
>>     physical node.
>>
>>     I am using Folsom but without Network namespaces. 
>>
>>     But in the website you have provided it states that "If you run
>>     both L3 + DHCP services on the same node, you should enable
>>     namespaces to avoid conflicts with routes :"
>>
>>     But currently quantum-dhcp-agent and quantum-l3-agent are running
>>     in the same node? 
>>
>>     Additionally the control node serves as a DHCP server for the
>>     local network ( Don't know if that would make and difference)
>>
>>     Any idea what the problem could be?
>>
>>
>>     On 16 February 2013 16:21, Jean-Baptiste RANSY
>>     <jean-baptiste.ransy@xxxxxxxxxx
>>     <mailto:jean-baptiste.ransy@xxxxxxxxxx>> wrote:
>>
>>         Hello Chathura,
>>
>>         Are you using Folsom with Network Namespaces ?
>>
>>         If yes, have a look here :
>>         http://docs.openstack.org/folsom/openstack-network/admin/content/ch_limitations.html
>>
>>
>>         Regards,
>>
>>         Jean-Baptsite RANSY
>>
>>
>>
>>         On 02/16/2013 05:01 PM, Chathura M. Sarathchandra
>>         Magurawalage wrote:
>>>         Hello guys,
>>>
>>>         The problem still exists. Any ideas?
>>>
>>>         Thanks 
>>>
>>>         On 15 February 2013 14:37, Sylvain Bauza
>>>         <sylvain.bauza@xxxxxxxxxxxx
>>>         <mailto:sylvain.bauza@xxxxxxxxxxxx>> wrote:
>>>
>>>             Metadata API allows to fetch SSH credentials when
>>>             booting (pubkey I mean).
>>>             If a VM is unable to reach metadata service, then it
>>>             won't be able to get its public key, so you won't be
>>>             able to connect, unless you specifically go thru a
>>>             Password authentication (provided password auth is
>>>             enabled in /etc/ssh/sshd_config, which is not the case
>>>             with Ubuntu cloud archive).
>>>             There is also a side effect, the boot process is longer
>>>             as the instance is waiting for the curl timeout (60sec.)
>>>             to finish booting up.
>>>
>>>             Re: Quantum, the metadata API is actually DNAT'd from
>>>             Network node to the Nova-api node (here 172.16.0.1 as
>>>             internal management IP) :
>>>             Chain quantum-l3-agent-PREROUTING (1 references)
>>>
>>>             target     prot opt source               destination
>>>             DNAT       tcp  --  0.0.0.0/0 <http://0.0.0.0/0>        
>>>                169.254.169.254      tcp dpt:80 to:172.16.0.1:8775
>>>             <http://172.16.0.1:8775>
>>>
>>>
>>>             Anyway, the first step is to :
>>>             1. grab the console.log
>>>             2. access thru VNC to the desired instance
>>>
>>>             Troubleshooting will be easier once that done.
>>>
>>>             -Sylvain
>>>
>>>
>>>
>>>             Le 15/02/2013 14:24, Chathura M. Sarathchandra
>>>             Magurawalage a écrit :
>>>
>>>                 Hello Guys,
>>>
>>>                 Not sure if this is the right port but these are the
>>>                 results:
>>>
>>>                 *Compute node:*
>>>
>>>
>>>                 root@computenode:~# netstat -an | grep 8775
>>>                 tcp        0      0 0.0.0.0:8775
>>>                 <http://0.0.0.0:8775> <http://0.0.0.0:8775>
>>>                  0.0.0.0:*               LISTEN
>>>
>>>                 *Controller: *
>>>
>>>
>>>                 root@controller:~# netstat -an | grep 8775
>>>                 tcp        0      0 0.0.0.0:8775
>>>                 <http://0.0.0.0:8775> <http://0.0.0.0:8775>
>>>                  0.0.0.0:*               LISTEN
>>>
>>>                 *Additionally I cant curl 169.254.169.254 from the
>>>                 compute node. I am not sure if this is related to
>>>                 not being able to PING the VM.*
>>>
>>>
>>>                 curl -v http://169.254.169.254
>>>                 * About to connect() to 169.254.169.254 port 80 (#0)
>>>                 *   Trying 169.254.169.254...
>>>
>>>                 Thanks for your help
>>>
>>>
>>>                 -----------------------------------------------------------------------------------------------------------------------------
>>>                 Chathura Madhusanka Sarathchandra Magurawalage.
>>>                 1NW.2.1, Desk 2
>>>                 School of Computer Science and Electronic Engineering
>>>                 University Of Essex
>>>                 United Kingdom.
>>>
>>>                 Email: csarata@xxxxxxxxxxx
>>>                 <mailto:csarata@xxxxxxxxxxx>
>>>                 <mailto:csarata@xxxxxxxxxxx
>>>                 <mailto:csarata@xxxxxxxxxxx>>
>>>                           chathura.sarathchandra@xxxxxxxxx
>>>                 <mailto:chathura.sarathchandra@xxxxxxxxx>
>>>                 <mailto:77.chathura@xxxxxxxxx
>>>                 <mailto:77.chathura@xxxxxxxxx>>
>>>                 77.chathura@xxxxxxxxx <mailto:77.chathura@xxxxxxxxx>
>>>                 <mailto:77.chathura@xxxxxxxxx
>>>                 <mailto:77.chathura@xxxxxxxxx>>
>>>
>>>
>>>
>>>                 On 15 February 2013 11:03, Anil Vishnoi
>>>                 <vishnoianil@xxxxxxxxx
>>>                 <mailto:vishnoianil@xxxxxxxxx>
>>>                 <mailto:vishnoianil@xxxxxxxxx
>>>                 <mailto:vishnoianil@xxxxxxxxx>>> wrote:
>>>
>>>                     If you are using ubuntu cloud image then the
>>>                 only way to log-in is
>>>                     to do ssh with the public key. For that you have
>>>                 to create ssh key
>>>                     pair and download the ssh key. You can create
>>>                 this ssh pair using
>>>                     horizon/cli.
>>>
>>>
>>>                     On Fri, Feb 15, 2013 at 4:27 PM, Sylvain Bauza
>>>                     <sylvain.bauza@xxxxxxxxxxxx
>>>                 <mailto:sylvain.bauza@xxxxxxxxxxxx>
>>>                 <mailto:sylvain.bauza@xxxxxxxxxxxx
>>>                 <mailto:sylvain.bauza@xxxxxxxxxxxx>>>
>>>
>>>                     wrote:
>>>
>>>
>>>                         Le 15/02/2013 11:42, Chathura M.
>>>                 Sarathchandra Magurawalage a
>>>                         écrit :
>>>
>>>
>>>                             How can I log into the VM from VNC? What
>>>                 are the credentials?
>>>
>>>
>>>                         You have multiple ways to get VNC access.
>>>                 The easiest one is
>>>                         thru Horizon. Other can be looking at the
>>>                 KVM command-line for
>>>                         the desired instance (on the compute node)
>>>                 and check the vnc
>>>                         port in use (assuming KVM as hypervisor).
>>>                         This is basic knowledge of Nova.
>>>
>>>
>>>
>>>                             nova-api-metadata is running fine in the
>>>                 compute node.
>>>
>>>
>>>                         Make sure the metadata port is avaible
>>>                 thanks to telnet or
>>>                         netstat, nova-api can be running without
>>>                 listening on metadata
>>>                         port.
>>>
>>>
>>>
>>>
>>>                         _______________________________________________
>>>                         Mailing list:
>>>                 https://launchpad.net/~openstack
>>>                 <https://launchpad.net/%7Eopenstack>
>>>                         <https://launchpad.net/%7Eopenstack>
>>>                         Post to     : openstack@xxxxxxxxxxxxxxxxxxx
>>>                 <mailto:openstack@xxxxxxxxxxxxxxxxxxx>
>>>                         <mailto:openstack@xxxxxxxxxxxxxxxxxxx
>>>                 <mailto:openstack@xxxxxxxxxxxxxxxxxxx>>
>>>                         Unsubscribe :
>>>                 https://launchpad.net/~openstack
>>>                 <https://launchpad.net/%7Eopenstack>
>>>                         <https://launchpad.net/%7Eopenstack>
>>>
>>>                         More help   :
>>>                 https://help.launchpad.net/ListHelp
>>>
>>>
>>>
>>>
>>>                     --     Thanks & Regards
>>>                     --Anil Kumar Vishnoi
>>>
>>>
>>>
>>>
>>>
>>>
>>>         _______________________________________________ Mailing
>>>         list: https://launchpad.net/~openstack
>>>         <https://launchpad.net/%7Eopenstack>
>>>         Post to : openstack@xxxxxxxxxxxxxxxxxxx
>>>         <mailto:openstack@xxxxxxxxxxxxxxxxxxx> Unsubscribe :
>>>         https://launchpad.net/~openstack
>>>         <https://launchpad.net/%7Eopenstack> More help :
>>>         https://help.launchpad.net/ListHelp
>>
>>
>
>


Follow ups

References