← Back to team overview

openstack team mailing list archive

Re: Cant ping private or floating IP

 

Hello JB,

I changed the IP and restarted the quantum-l3-agent but still no luck :(

Thanks.

-----------------------------------------------------------------------------------------------------------------------------
Chathura Madhusanka Sarathchandra Magurawalage.
1NW.2.1, Desk 2
School of Computer Science and Electronic Engineering
University Of Essex
United Kingdom.

Email: csarata@xxxxxxxxxxx
          chathura.sarathchandra@xxxxxxxxx <77.chathura@xxxxxxxxx>
          77.chathura@xxxxxxxxx


On 17 February 2013 00:47, Jean-Baptiste RANSY <
jean-baptiste.ransy@xxxxxxxxxx> wrote:

>  and restart quantum-l3-agent :)
>
> JB
>
>
> On 02/17/2013 01:46 AM, Jean-Baptiste RANSY wrote:
>
> Found !
>
> On the controller node you must change the metadata_ip in
> /etc/quantum/l3_agent.ini
>
> This params is used to create the nat rule quantum-l3-agent-PREROUTING
>
> Just replace 127.0.0.1 by 192.168.2.225 and that should be ok.
>
> JB
>
>
> On 02/17/2013 01:04 AM, Jean-Baptiste RANSY wrote:
>
> Hi, Chathura
>
> The compute node log file /var/log/nova/nova-api.log is too light (maybe
> logrotate :p)
>
> Please, clear nova-api.log, restart nova-api service, start a new instance
> and wait cloud-init fail to retrieve metadata.
>
> Thx,
>
> JB
>
>
> On 02/16/2013 11:35 PM, Chathura M. Sarathchandra Magurawalage wrote:
>
>  Thanks for that.
> *
> *
> *root@controller:~# ip addr show*
> 1: lo: <LOOPBACK,UP,LOWER_UP> mtu 16436 qdisc noqueue state UNKNOWN
>     link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00
>     inet 127.0.0.1/8 scope host lo
>     inet6 ::1/128 scope host
>        valid_lft forever preferred_lft forever
> 2: eth0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc mq state UP qlen
> 1000
>     link/ether d4:ae:52:bb:aa:20 brd ff:ff:ff:ff:ff:ff
>     inet 10.10.10.1/24 brd 10.10.10.255 scope global eth0
>     inet6 fe80::d6ae:52ff:febb:aa20/64 scope link
>        valid_lft forever preferred_lft forever
> 3: eth1: <BROADCAST,MULTICAST> mtu 1500 qdisc noop state DOWN qlen 1000
>     link/ether d4:ae:52:bb:aa:21 brd ff:ff:ff:ff:ff:ff
> 4: eth0.2@eth0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc noqueue
> state UP
>     link/ether d4:ae:52:bb:aa:20 brd ff:ff:ff:ff:ff:ff
>     inet 192.168.2.225/24 brd 192.168.2.255 scope global eth0.2
>     inet6 fe80::d6ae:52ff:febb:aa20/64 scope link
>        valid_lft forever preferred_lft forever
> 5: br-int: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc noqueue state
> UNKNOWN
>     link/ether ba:7a:e9:dc:2b:41 brd ff:ff:ff:ff:ff:ff
>     inet6 fe80::b87a:e9ff:fedc:2b41/64 scope link
>        valid_lft forever preferred_lft forever
> 7: br-ex: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc noqueue state
> UNKNOWN
>      link/ether 9a:41:c8:8a:9e:49 brd ff:ff:ff:ff:ff:ff
>     inet 192.168.2.225/24 scope global br-ex
> 8: tapf71b5b86-5c: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc
> noqueue state UNKNOWN
>     link/ether 2a:44:a3:d1:7d:f3 brd ff:ff:ff:ff:ff:ff
>     inet 10.5.5.2/24 brd 10.5.5.255 scope global tapf71b5b86-5c
>     inet6 fe80::2844:a3ff:fed1:7df3/64 scope link
>        valid_lft forever preferred_lft forever
> 9: qr-4d088f3a-78: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc
> noqueue state UNKNOWN
>     link/ether ca:5b:8d:4d:6d:fb brd ff:ff:ff:ff:ff:ff
>     inet 10.5.5.1/24 brd 10.5.5.255 scope global qr-4d088f3a-78
>     inet6 fe80::c85b:8dff:fe4d:6dfb/64 scope link
>        valid_lft forever preferred_lft forever
> 10: qg-6f8374cb-cb: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc
> noqueue state UNKNOWN
>     link/ether 0e:7f:dd:3a:80:bc brd ff:ff:ff:ff:ff:ff
>     inet 192.168.2.151/24 brd 192.168.2.255 scope global qg-6f8374cb-cb
>     inet6 fe80::c7f:ddff:fe3a:80bc/64 scope link
>        valid_lft forever preferred_lft forever
> 27: br-tun: <BROADCAST,MULTICAST> mtu 1500 qdisc noop state DOWN
>     link/ether 8a:cf:ec:7c:15:40 brd ff:ff:ff:ff:ff:ff
>
>  *cat /proc/sys/net/ipv4/ip_forward*
> 1
>
>  *root@computenode:~# ip addr show*
> 1: lo: <LOOPBACK,UP,LOWER_UP> mtu 16436 qdisc noqueue state UNKNOWN
>     link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00
>     inet 127.0.0.1/8 scope host lo
>     inet6 ::1/128 scope host
>        valid_lft forever preferred_lft forever
> 2: eth0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc mq state UP qlen
> 1000
>     link/ether d4:ae:52:bb:a1:9d brd ff:ff:ff:ff:ff:ff
>     inet 10.10.10.12/24 brd 10.10.10.255 scope global eth0
>     inet6 fe80::d6ae:52ff:febb:a19d/64 scope link
>        valid_lft forever preferred_lft forever
> 3: eth1: <BROADCAST,MULTICAST> mtu 1500 qdisc noop state DOWN qlen 1000
>     link/ether d4:ae:52:bb:a1:9e brd ff:ff:ff:ff:ff:ff
> 4: eth0.2@eth0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc noqueue
> state UP
>     link/ether d4:ae:52:bb:a1:9d brd ff:ff:ff:ff:ff:ff
>     inet 192.168.2.234/24 brd 192.168.2.255 scope global eth0.2
>     inet6 fe80::d6ae:52ff:febb:a19d/64 scope link
>        valid_lft forever preferred_lft forever
> 5: br-int: <BROADCAST,MULTICAST> mtu 1500 qdisc noop state DOWN
>     link/ether ae:9b:43:09:af:40 brd ff:ff:ff:ff:ff:ff
> 9: qbr256f5ed2-43: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc
> noqueue state UP
>     link/ether c6:c0:df:64:c6:99 brd ff:ff:ff:ff:ff:ff
>     inet6 fe80::20e8:b9ff:fe6c:6f55/64 scope link
>        valid_lft forever preferred_lft forever
> 10: qvo256f5ed2-43: <BROADCAST,MULTICAST,PROMISC,UP,LOWER_UP> mtu 1500
> qdisc pfifo_fast state UP qlen 1000
>     link/ether 76:25:8b:fd:90:3b brd ff:ff:ff:ff:ff:ff
>     inet6 fe80::7425:8bff:fefd:903b/64 scope link
>        valid_lft forever preferred_lft forever
> 11: qvb256f5ed2-43: <BROADCAST,MULTICAST,PROMISC,UP,LOWER_UP> mtu 1500
> qdisc pfifo_fast master qbr256f5ed2-43 state UP qlen 1000
>     link/ether c6:c0:df:64:c6:99 brd ff:ff:ff:ff:ff:ff
>     inet6 fe80::c4c0:dfff:fe64:c699/64 scope link
>        valid_lft forever preferred_lft forever
> 13: br-tun: <BROADCAST,MULTICAST> mtu 1500 qdisc noop state DOWN
>     link/ether be:8c:30:78:35:48 brd ff:ff:ff:ff:ff:ff
> 15: vnet0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc pfifo_fast
> master qbr256f5ed2-43 state UNKNOWN qlen 500
>     link/ether fe:16:3e:57:ec:ff brd ff:ff:ff:ff:ff:ff
>     inet6 fe80::fc16:3eff:fe57:ecff/64 scope link
>        valid_lft forever preferred_lft forever
>
>  btw cronus is my compute node and I have renamed it to computenode to
> understand it better.
>
> On 16 February 2013 22:11, Jean-Baptiste RANSY <
> jean-baptiste.ransy@xxxxxxxxxx> wrote:
>
>>  Download in progress
>>
>> Can you send me the output of those commands i forgot :
>>
>> Controller Node:
>> $ ip addr show
>> $ cat /proc/sys/net/ipv4/ip_forward
>>
>> Compute Node:
>> $ ip addr show
>>
>>
>> JB
>>
>>
>>
>> On 02/16/2013 10:45 PM, Chathura M. Sarathchandra Magurawalage wrote:
>>
>>  Thanks Ransy,
>>
>>  I have created a tar file with the configuration and log files in it.
>> Please download it using the following URL. I have pasted the output of the
>> commands below.
>>
>>  https://www.dropbox.com/s/qyfcsn50060y304/confilesnlogs.tar
>>
>>  *Controller node:*
>> *root@controller:~# keystone endpoint-list*
>> +----------------------------------+-----------+-------------------------------------------------+-------------------------------------------------+--------------------------------------------+Controller
>> node
>> |                id                |   region  |
>>  publicurl                    |                   internalurl
>>     |                  adminurl                  |
>>
>> +----------------------------------+-----------+-------------------------------------------------+-------------------------------------------------+--------------------------------------------+
>> | 2c9a1cb0fe8247d9b7716432cf459fe5 | RegionOne |
>> http://192.168.2.225:8774/v2/$(tenant_id)s   |
>> http://192.168.2.225:8774/v2/$(tenant_id)s   |
>> http://192.168.2.225:8774/v2/$(tenant_id)s |
>> | 2d306903ed3342a8aaaac7c5680c116f | RegionOne |
>> http://192.168.2.225:9696/           |
>> http://192.168.2.225:9696/           |         http://192.168.2.225:9696/        |
>> | 3848114f120f42bf819bc2443b28ac9e | RegionOne |
>> http://192.168.2.225:8080/v1/AUTH_$(tenant_id)s |
>> http://192.168.2.225:8080/v1/AUTH_$(tenant_id)s |
>> http://192.168.2.225:8080/v1        |
>> | 4955173b8d9e4d33ae4a5b29dc12c74d | RegionOne |
>> http://192.168.2.225:8776/v1/$(tenant_id)s   |
>> http://192.168.2.225:8776/v1/$(tenant_id)s   |
>> http://192.168.2.225:8776/v1/$(tenant_id)s |
>> | d313aa76bf854dde94f33a49a9f0c8ac | RegionOne |
>> http://192.168.2.225:9292/v2          |
>> http://192.168.2.225:9292/v2          |
>> http://192.168.2.225:9292/v2        |
>> | e5aa4ecf3cbe4dd5aba9b204c74fee6a | RegionOne |
>> http://192.168.2.225:5000/v2.0         |
>> http://192.168.2.225:5000/v2.0         |
>> http://192.168.2.225:35357/v2.0       |
>> | fba6f790e3b444c890d114f13cd32b37 | RegionOne |
>> http://192.168.2.225:8773/services/Cloud    |
>> http://192.168.2.225:8773/services/Cloud    |
>> http://192.168.2.225:8773/services/Admin  |
>>
>> +----------------------------------+-----------+-------------------------------------------------+-------------------------------------------------+--------------------------------------------+
>>
>>  *root@controller:~# ip link show*
>> 1: lo: <LOOPBACK,UP,LOWER_UP> mtu 16436 qdisc noqueue state UNKNOWN
>>     link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00
>> 2: eth0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc mq state UP
>> qlen 1000
>>     link/ether d4:ae:52:bb:aa:20 brd ff:ff:ff:ff:ff:ff
>> 3: eth1: <BROADCAST,MULTICAST> mtu 1500 qdisc noop state DOWN qlen 1000
>>     link/ether d4:ae:52:bb:aa:21 brd ff:ff:ff:ff:ff:ff
>>  4: eth0.2@eth0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc
>> noqueue state UP
>>     link/ether d4:ae:52:bb:aa:20 brd ff:ff:ff:ff:ff:ff
>> 5: br-int: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc noqueue state
>> UNKNOWN
>>     link/ether ba:7a:e9:dc:2b:41 brd ff:ff:ff:ff:ff:ff
>> 7: br-ex: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc noqueue state
>> UNKNOWN
>>     link/ether 9a:41:c8:8a:9e:49 brd ff:ff:ff:ff:ff:ff
>> 8: tapf71b5b86-5c: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc
>> noqueue state UNKNOWN
>>     link/ether 2a:44:a3:d1:7d:f3 brd ff:ff:ff:ff:ff:ff
>> 9: qr-4d088f3a-78: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc
>> noqueue state UNKNOWN
>>     link/ether ca:5b:8d:4d:6d:fb brd ff:ff:ff:ff:ff:ff
>> 10: qg-6f8374cb-cb: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc
>> noqueue state UNKNOWN
>>     link/ether 0e:7f:dd:3a:80:bc brd ff:ff:ff:ff:ff:ff
>> 27: br-tun: <BROADCAST,MULTICAST> mtu 1500 qdisc noop state DOWN
>>     link/ether 8a:cf:ec:7c:15:40 brd ff:ff:ff:ff:ff:ff
>>
>>  *root@controller:~# ip route show*
>> default via 192.168.2.253 dev eth0.2
>> default via 192.168.2.253 dev eth0.2  metric 100
>> 10.5.5.0/24 dev tapf71b5b86-5c  proto kernel  scope link  src 10.5.5.2
>> 10.5.5.0/24 dev qr-4d088f3a-78  proto kernel  scope link  src 10.5.5.1
>> 10.10.10.0/24 dev eth0  proto kernel  scope link  src 10.10.10.1
>> 192.168.2.0/24 dev eth0.2  proto kernel  scope link  src 192.168.2.225
>> 192.168.2.0/24 dev qg-6f8374cb-cb  proto kernel  scope link  src
>> 192.168.2.151
>> 192.168.2.0/24 dev br-ex  proto kernel  scope link  src 192.168.2.225
>>
>> *$ ip netns show (Did not return anything)*
>>
>>  *root@controller:~# ovs-vsctl show*
>> a566afae-d7a8-42a9-aefe-8b0f2f7054a3
>>     Bridge br-tun
>>         Port "gre-4"
>>             Interface "gre-4"
>>                 type: gre
>>                 options: {in_key=flow, out_key=flow,
>> remote_ip="10.10.10.12"}
>>         Port "gre-3"
>>             Interface "gre-3"
>>                  type: gre
>>                 options: {in_key=flow, out_key=flow,
>> remote_ip="127.0.0.1"}
>>         Port patch-int
>>             Interface patch-int
>>                 type: patch
>>                 options: {peer=patch-tun}
>>         Port br-tun
>>              Interface br-tun
>>                 type: internal
>>         Port "gre-1"
>>             Interface "gre-1"
>>                 type: gre
>>                  options: {in_key=flow, out_key=flow,
>> remote_ip="10.0.0.3"}
>>     Bridge br-ex
>>          Port br-ex
>>             Interface br-ex
>>                 type: internal
>>         Port "qg-6f8374cb-cb"
>>             Interface "qg-6f8374cb-cb"
>>                 type: internal
>>         Port "br0"
>>              Interface "br0"
>>     Bridge br-int
>>         Port br-int
>>             Interface br-int
>>                 type: internal
>>          Port "tapf71b5b86-5c"
>>             tag: 1
>>             Interface "tapf71b5b86-5c"
>>                 type: internal
>>         Port patch-tun
>>             Interface patch-tun
>>                 type: patch
>>                 options: {peer=patch-int}
>>          Port "qr-4d088f3a-78"
>>             tag: 1
>>             Interface "qr-4d088f3a-78"
>>                 type: internal
>>     ovs_version: "1.4.0+build0"
>>
>>
>>  *Compute node:*
>>
>>  *root@cronus:~# ip link show*
>> 1: lo: <LOOPBACK,UP,LOWER_UP> mtu 16436 qdisc noqueue state UNKNOWN
>>      link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00
>> 2: eth0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc mq state UP
>> qlen 1000
>>     link/ether d4:ae:52:bb:a1:9d brd ff:ff:ff:ff:ff:ff
>> 3: eth1: <BROADCAST,MULTICAST> mtu 1500 qdisc noop state DOWN qlen 1000
>>     link/ether d4:ae:52:bb:a1:9e brd ff:ff:ff:ff:ff:ff
>> 4: eth0.2@eth0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc noqueue
>> state UP
>>     link/ether d4:ae:52:bb:a1:9d brd ff:ff:ff:ff:ff:ff
>> 5: br-int: <BROADCAST,MULTICAST> mtu 1500 qdisc noop state DOWN
>>     link/ether ae:9b:43:09:af:40 brd ff:ff:ff:ff:ff:ff
>> 9: qbr256f5ed2-43: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc
>> noqueue state UP
>>     link/ether c6:c0:df:64:c6:99 brd ff:ff:ff:ff:ff:ff
>> 10: qvo256f5ed2-43: <BROADCAST,MULTICAST,PROMISC,UP,LOWER_UP> mtu 1500
>> qdisc pfifo_fast state UP qlen 1000
>>     link/ether 76:25:8b:fd:90:3b brd ff:ff:ff:ff:ff:ff
>> 11: qvb256f5ed2-43: <BROADCAST,MULTICAST,PROMISC,UP,LOWER_UP> mtu 1500
>> qdisc pfifo_fast master qbr256f5ed2-43 state UP qlen 1000
>>     link/ether c6:c0:df:64:c6:99 brd ff:ff:ff:ff:ff:ff
>> 13: br-tun: <BROADCAST,MULTICAST> mtu 1500 qdisc noop state DOWN
>>     link/ether be:8c:30:78:35:48 brd ff:ff:ff:ff:ff:ff
>> 15: vnet0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc pfifo_fast
>> master qbr256f5ed2-43 state UNKNOWN qlen 500
>>     link/ether fe:16:3e:57:ec:ff brd ff:ff:ff:ff:ff:ff
>>
>>   *root@cronus:~# ip route show*
>> default via 192.168.2.253 dev eth0.2  metric 100
>> 10.10.10.0/24 dev eth0  proto kernel  scope link  src 10.10.10.12
>> 192.168.2.0/24 dev eth0.2  proto kernel  scope link  src 192.168.2.234
>>
>>   *root@cronus:~# ovs-vsctl show*
>> d85bc334-6d64-4a13-b851-d56b18ff1549
>>     Bridge br-int
>>         Port "qvo0e743b01-89"
>>             tag: 4095
>>             Interface "qvo0e743b01-89"
>>         Port "qvo256f5ed2-43"
>>             tag: 1
>>             Interface "qvo256f5ed2-43"
>>         Port patch-tun
>>             Interface patch-tun
>>                 type: patch
>>                 options: {peer=patch-int}
>>         Port br-int
>>             Interface br-int
>>                 type: internal
>>         Port "qvoee3d4131-2a"
>>             tag: 4095
>>             Interface "qvoee3d4131-2a"
>>         Port "qvocbc816bd-3d"
>>             tag: 4095
>>             Interface "qvocbc816bd-3d"
>>     Bridge br-tun
>>         Port br-tun
>>             Interface br-tun
>>                 type: internal
>>         Port "gre-2"
>>             Interface "gre-2"
>>                 type: gre
>>                 options: {in_key=flow, out_key=flow,
>> remote_ip="10.10.10.1"}
>>         Port "gre-1"
>>             Interface "gre-1"
>>                 type: gre
>>                 options: {in_key=flow, out_key=flow, remote_ip="10.0.0.3"}
>>         Port patch-int
>>             Interface patch-int
>>                 type: patch
>>                 options: {peer=patch-tun}
>>         Port "gre-3"
>>             Interface "gre-3"
>>                 type: gre
>>                 options: {in_key=flow, out_key=flow,
>> remote_ip="127.0.0.1"}
>>     ovs_version: "1.4.0+build0"
>>
>>
>>  Thanks I appreciate your help.
>>
>>   On 16 February 2013 16:49, Jean-Baptiste RANSY <
>> jean-baptiste.ransy@xxxxxxxxxx> wrote:
>>
>>>  Please provide files listed bellow :
>>>
>>> Controller Node :
>>> /etc/nova/nova.conf
>>> /etc/nova/api-paste.ini
>>> /etc/quantum/l3_agent.ini
>>> /etc/quantum/quantum.conf
>>> /etc/quantum/dhcp_agent.ini
>>> /etc/quantum/plugins/openvswitch/ovs_quantum_plugin.ini
>>> /etc/quantum/api-paste.ini
>>> /var/log/nova/*.log
>>> /var/log/quantum/*.log
>>>
>>> Compute Node :
>>> /etc/nova/nova.conf
>>> /etc/nova/nova-compute.conf
>>> /etc/nova/api-paste.ini
>>> /etc/quantum/quantum.conf
>>> /etc/quantum/plugins/openvswitch/ovs_quantum_plugin.ini
>>> /var/log/nova/*.log
>>> /var/log/quantum/*.log
>>>
>>> Plus, complete output of the following commands :
>>>
>>> Controller Node :
>>> $ keystone endpoint-list
>>> $ ip link show
>>> $ ip route show
>>> $ ip netns show
>>> $ ovs-vsctl show
>>>
>>> Compute Node :
>>> $ ip link show
>>> $ ip route show
>>> $ ovs-vsctl show
>>>
>>> Regards,
>>>
>>> Jean-Baptiste RANSY
>>>
>>>
>>>
>>> On 02/16/2013 05:32 PM, Chathura M. Sarathchandra Magurawalage wrote:
>>>
>>> Hello Jean,
>>>
>>>  Thanks for your reply.
>>>
>>>  I followed the instructions in
>>> http://docs.openstack.org/folsom/basic-install/content/basic-install_network.html.
>>> And my Controller and the Network-node is installed in the same physical
>>> node.
>>>
>>>  I am using Folsom but without Network namespaces.
>>>
>>>  But in the website you have provided it states that "If you run both
>>> L3 + DHCP services on the same node, you should enable namespaces to avoid
>>> conflicts with routes :"
>>>
>>>  But currently quantum-dhcp-agent and quantum-l3-agent are running in
>>> the same node?
>>>
>>>  Additionally the control node serves as a DHCP server for the local
>>> network ( Don't know if that would make and difference)
>>>
>>>  Any idea what the problem could be?
>>>
>>>
>>> On 16 February 2013 16:21, Jean-Baptiste RANSY <
>>> jean-baptiste.ransy@xxxxxxxxxx> wrote:
>>>
>>>>  Hello Chathura,
>>>>
>>>> Are you using Folsom with Network Namespaces ?
>>>>
>>>> If yes, have a look here :
>>>> http://docs.openstack.org/folsom/openstack-network/admin/content/ch_limitations.html
>>>>
>>>>
>>>> Regards,
>>>>
>>>> Jean-Baptsite RANSY
>>>>
>>>>
>>>>
>>>> On 02/16/2013 05:01 PM, Chathura M. Sarathchandra Magurawalage wrote:
>>>>
>>>>  Hello guys,
>>>>
>>>>  The problem still exists. Any ideas?
>>>>
>>>>  Thanks
>>>>
>>>>   On 15 February 2013 14:37, Sylvain Bauza <sylvain.bauza@xxxxxxxxxxxx>wrote:
>>>>
>>>>> Metadata API allows to fetch SSH credentials when booting (pubkey I
>>>>> mean).
>>>>> If a VM is unable to reach metadata service, then it won't be able to
>>>>> get its public key, so you won't be able to connect, unless you
>>>>> specifically go thru a Password authentication (provided password auth is
>>>>> enabled in /etc/ssh/sshd_config, which is not the case with Ubuntu cloud
>>>>> archive).
>>>>> There is also a side effect, the boot process is longer as the
>>>>> instance is waiting for the curl timeout (60sec.) to finish booting up.
>>>>>
>>>>> Re: Quantum, the metadata API is actually DNAT'd from Network node to
>>>>> the Nova-api node (here 172.16.0.1 as internal management IP) :
>>>>> Chain quantum-l3-agent-PREROUTING (1 references)
>>>>>
>>>>> target     prot opt source               destination
>>>>>  DNAT       tcp  --  0.0.0.0/0            169.254.169.254      tcp
>>>>> dpt:80 to:172.16.0.1:8775
>>>>>
>>>>>
>>>>> Anyway, the first step is to :
>>>>> 1. grab the console.log
>>>>> 2. access thru VNC to the desired instance
>>>>>
>>>>> Troubleshooting will be easier once that done.
>>>>>
>>>>> -Sylvain
>>>>>
>>>>>
>>>>>
>>>>> Le 15/02/2013 14:24, Chathura M. Sarathchandra Magurawalage a écrit :
>>>>>
>>>>>>  Hello Guys,
>>>>>>
>>>>>> Not sure if this is the right port but these are the results:
>>>>>>
>>>>>>  *Compute node:*
>>>>>>
>>>>>>
>>>>>> root@computenode:~# netstat -an | grep 8775
>>>>>>  tcp        0      0 0.0.0.0:8775 <http://0.0.0.0:8775>  0.0.0.0:*
>>>>>>             LISTEN
>>>>>>
>>>>>> *Controller: *
>>>>>>
>>>>>>
>>>>>> root@controller:~# netstat -an | grep 8775
>>>>>>  tcp        0      0 0.0.0.0:8775 <http://0.0.0.0:8775>  0.0.0.0:*
>>>>>>             LISTEN
>>>>>>
>>>>>> *Additionally I cant curl 169.254.169.254 from the compute node. I am
>>>>>> not sure if this is related to not being able to PING the VM.*
>>>>>>
>>>>>>
>>>>>> curl -v http://169.254.169.254
>>>>>> * About to connect() to 169.254.169.254 port 80 (#0)
>>>>>> *   Trying 169.254.169.254...
>>>>>>
>>>>>> Thanks for your help
>>>>>>
>>>>>>
>>>>>>
>>>>>> -----------------------------------------------------------------------------------------------------------------------------
>>>>>> Chathura Madhusanka Sarathchandra Magurawalage.
>>>>>> 1NW.2.1, Desk 2
>>>>>> School of Computer Science and Electronic Engineering
>>>>>> University Of Essex
>>>>>> United Kingdom.
>>>>>>
>>>>>>  Email: csarata@xxxxxxxxxxx <mailto:csarata@xxxxxxxxxxx>
>>>>>>           chathura.sarathchandra@xxxxxxxxx <mailto:
>>>>>> 77.chathura@xxxxxxxxx>
>>>>>> 77.chathura@xxxxxxxxx <mailto:77.chathura@xxxxxxxxx>
>>>>>>
>>>>>>
>>>>>>
>>>>>> On 15 February 2013 11:03, Anil Vishnoi <vishnoianil@xxxxxxxxx<mailto:
>>>>>> vishnoianil@xxxxxxxxx>> wrote:
>>>>>>
>>>>>>     If you are using ubuntu cloud image then the only way to log-in is
>>>>>>     to do ssh with the public key. For that you have to create ssh key
>>>>>>     pair and download the ssh key. You can create this ssh pair using
>>>>>>     horizon/cli.
>>>>>>
>>>>>>
>>>>>>     On Fri, Feb 15, 2013 at 4:27 PM, Sylvain Bauza
>>>>>>      <sylvain.bauza@xxxxxxxxxxxx <mailto:sylvain.bauza@xxxxxxxxxxxx>>
>>>>>>
>>>>>>
>>>>>>     wrote:
>>>>>>
>>>>>>
>>>>>>         Le 15/02/2013 11:42, Chathura M. Sarathchandra Magurawalage a
>>>>>>         écrit :
>>>>>>
>>>>>>
>>>>>>             How can I log into the VM from VNC? What are the
>>>>>> credentials?
>>>>>>
>>>>>>
>>>>>>         You have multiple ways to get VNC access. The easiest one is
>>>>>>         thru Horizon. Other can be looking at the KVM command-line for
>>>>>>         the desired instance (on the compute node) and check the vnc
>>>>>>         port in use (assuming KVM as hypervisor).
>>>>>>         This is basic knowledge of Nova.
>>>>>>
>>>>>>
>>>>>>
>>>>>>             nova-api-metadata is running fine in the compute node.
>>>>>>
>>>>>>
>>>>>>         Make sure the metadata port is avaible thanks to telnet or
>>>>>>         netstat, nova-api can be running without listening on metadata
>>>>>>         port.
>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>>         _______________________________________________
>>>>>>         Mailing list: https://launchpad.net/~openstack
>>>>>>          <https://launchpad.net/%7Eopenstack>
>>>>>>         Post to     : openstack@xxxxxxxxxxxxxxxxxxx
>>>>>>         <mailto:openstack@xxxxxxxxxxxxxxxxxxx>
>>>>>>         Unsubscribe : https://launchpad.net/~openstack
>>>>>>         <https://launchpad.net/%7Eopenstack>
>>>>>>
>>>>>>         More help   : https://help.launchpad.net/ListHelp
>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>>     --     Thanks & Regards
>>>>>>     --Anil Kumar Vishnoi
>>>>>>
>>>>>>
>>>>>>
>>>>>
>>>>
>>>>
>>>>  _______________________________________________
>>>> Mailing list: https://launchpad.net/~openstack
>>>>
>>>> Post to     : openstack@xxxxxxxxxxxxxxxxxxx
>>>> Unsubscribe : https://launchpad.net/~openstack
>>>> More help   : https://help.launchpad.net/ListHelp
>>>>
>>>>
>>>>
>>>
>>>
>>
>>
>
>
>
>

Follow ups

References