yahoo-eng-team team mailing list archive
-
yahoo-eng-team team
-
Mailing list archive
-
Message #93045
[Bug 2042598] [NEW] neutron_server container suspended in health:starting state
Public bug reported:
I installed OpenStack (zed) on a Raspberry Pi cluster with kolla-ansible
(version tagged for zed), all containers are healthy except the
neutron_server which is suspended in 'health: starting' state. Network
related part of OpenStack does not work. Some other commands commands
work as expected (e.g., can create an image which is reported by
openstack image list as 'active').
There are four Raspberry Pi 4B in the cluster (2 x 4GB RAM and 2 x 8GB RAM). They run Debian 11 (bullseaye) and kolla-ansible has been used for the installation.
Notably, I'm using a specific configuration of networking on my Pis to mimic two network interfaces on each host as kolla-ansible expects. These are provided as interfaces of veth pairs (more details on that below, too).
Below, one can find:
1. configuration commands I used to configure my Pi hosts (this panel)
2. environment details related to the Pis (the one serving as controller in OpenStack) and kolla-ansible install information (this panel)
3. ml2_conf.ini and nova-compute.conf configuration used in kolla-ansible
4. kolla-ansible files: globals.yml (4.1) and inventory multinode (4.2)
- changed parts - this panel
- complete versions - attachments
5. HttpException: 503 message from running init-runonce (kolla-ansible test script for new installation) (this panel)
6. status of containers on the control node as reported by 'docker ps -a' (this panel)
7. output form docker neutron_server inspect command (attachment)
8. log form neutron_server container (attachment)
*************************************************************
1. Debian configuration on the Pis
*************************************************************
Selected details fo the configuration are given in the following.
Basically, most of them are needed to configure Pis' host networking
using netplan. Another one relates to qemu-kvm.
(Note: initial configs to enable ssh access should be done locally (keyboard, monitor) on each Pi, in particular:
PermitRootLogin yes
PasswordAuthentication yes
I skip the details of enabling ssh access, though. Below, I assume ssh acces as a regular (non-root) user.
)
=========== Preparation for host networking setup ===========
$ sudo apt-get remove unattended-upgrades -y
$ sudo apt-get update -y && sudo apt-get upgrade -y
- updating $PATH for a user
$ sudo tee -a ~/.bashrc << EOT
export PATH=$PATH:/usr/local/sbin:/usr/sbin:/sbin
EOT
$ source ~/.bashrc
- enable systemd-networkd and configure eth0 for ssh access (neede to use ssh; not neede if one does stuff locally, attaching keyboard and monitor to each Pi)
- enabling systemd-networkd
$ sudo mv /etc/network/interfaces /etc/network/interfaces.save
$ sudo mv /etc/network/interfaces.d /etc/network/interfaces.d.save
$ sudo systemctl enable systemd-networkd && sudo systemctl start systemd-networkd
$ sudo systemctl status systemd-networkd
- configure eth0 (in may case, I've configured static DHCP for each Pi on my DHCP server)
$ sudo tee /etc/systemd/network/20-wired.network << EOT
[Match]
Name=eth0
[Network]
DHCP=yes
EOT
- install netplan
$ sudo apt update && sudo apt -y install netplan.io
$ sudo reboot
- enable ip forwarding
$ sudo nano /etc/sysctl.conf
===> uncomment the line: net.ipv4.ip_forward=1
$sudo sysctl -p
========= Host networking setup ==========
- network setup on each Pi host - drawing:
192.168.1.xy/24 bez adresu IP
+---------+ +---------+
| veth0 | | veth1 | <==== network-interface and network-external-interface for kolla-ansible
+---------+ +---------+
| veth pairs |
+---------+ +---------+
| veth0br | | veth1br |
+---------+ +---------+
+-┴-----------------┴-+
| brmux |
+----------┬----------+
+---------+
| eth0 |
+---------+
========== commands for achieving the above network setup (on each Pi)
-------- first veth pair
$ sudo tee /etc/systemd/network/veth-openstack-net-itf-veth0.netdev << EOT
#network-interface in kolla-ansible globals.yml
[NetDev]
Name=veth0
Kind=veth
[Peer]
Name=veth0br
EOT
--------- second veth pair
$sudo tee /etc/systemd/network/veth-openstack-net-itf-veth1.netdev << EOT
#network-external-interface in kolla-ansible globals.yml
[NetDev]
Name=veth1
Kind=veth
[Peer]
Name=veth1br
EOT
--------- set the networking (wlan0 is not necessary, just for any case)
sudo tee /etc/netplan/50-cloud-init.yaml << EOT
network:
version: 2
renderer: networkd
#-----------------------------------------#
# enable wlan for any case #
#-----------------------------------------#
# Interfejs wlan0 dostanie IPaddr z Linksysa przez DHCP.
wifis:
wlan0:
access-points:
FreshTomato06:
password: klasterek
dhcp4: true
optional: true
#-----------------------------------------#
# Kolla-Ansible required #
#-----------------------------------------#
# Interfaces
ethernets:
eth0:
dhcp4: false
dhcp6: false
# kolla-ansible network-interface
veth0:
addresses:
- 192.168.1.xy/24 # set appropriate address
nameservers:
addresses:
- 192.168.1.1 # dns on the local/home network router
- 8.8.8.8
routes:
- to: 0.0.0.0/0
via: 192.168.1.1
veth0br:
dhcp4: false
dhcp6: false
# kolla-ansible network-external-interface
veth1:
dhcp4: false
dhcp6: false
veth1br:
dhcp4: false
dhcp6: false
# Bridge - logically, a switch on the side of data center provider
network
bridges:
brmux:
interfaces:
- eth0
- veth0br
- veth1br
EOT
$ sudo netplan generate
$ sudo netplan apply
(now ssh using the new IP address of veth0)
============ Remaining Pi configurations ============
$ sudo visudo ==> add NOPASSWD: ALL as shown below
# Allow members of group sudo to execute any command
%sudo ALL=(ALL:ALL) NOPASSWD: ALL
$ sudo apt-get update
$ sudo apt-get install -y qemu-kvm
$ sudo apt-get upgrade -y
- on notification:
A new version (/tmp/tmp.EofC83AObD) of configuration file /etc/ssh/sshd_config is available
I choose ==> keep the local version currently installed
$ sudo apt-get install sshpass
$ sudo apt-get install ufw
$ sudo apt-get install net-tools <=== not necessary, I install it on the Pis just for any case
$ sudo nano /etc/sysctl.conf
===> uncomment the line: net.ipv4.ip_forward=1
$ sudo sysctl -p
************ End of Debian configuration *********
**************************************************
2. Environment details
**************************************************
( when related to a Pi this is the one serving as the control node)
* OS (e.g. from /etc/os-release):
ubuntu@ost64:~$ cat /etc/os-release
PRETTY_NAME="Debian GNU/Linux 11 (bullseye)"
NAME="Debian GNU/Linux"
VERSION_ID="11"
VERSION="11 (bullseye)"
VERSION_CODENAME=bullseye
ID=debian
HOME_URL="https://www.debian.org/"
SUPPORT_URL="https://www.debian.org/support"
BUG_REPORT_URL="https://bugs.debian.org/"
* Kernel (e.g. `uname -a`):
ubuntu@ost64:~$ uname -a
Linux ost64 5.10.0-20-arm64 #1 SMP Debian 5.10.158-2 (2022-12-13) aarch64 GNU/Linux
* Docker version if applicable (e.g. `docker version`):
ubuntu@ost64:~$ sudo docker version
Client: Docker Engine - Community
Version: 24.0.7
API version: 1.43
Go version: go1.20.10
Git commit: afdd53b
Built: Thu Oct 26 09:08:29 2023
OS/Arch: linux/arm64
Context: default
Server: Docker Engine - Community
Engine:
Version: 24.0.7
API version: 1.43 (minimum version 1.12)
Go version: go1.20.10
Git commit: 311b9ff
Built: Thu Oct 26 09:08:29 2023
OS/Arch: linux/arm64
Experimental: false
containerd:
Version: 1.6.24
GitCommit: 61f9fd88f79f081d64d6fa3bb1a0dc71ec870523
runc:
Version: 1.1.9
GitCommit: v1.1.9-0-gccaecfc
docker-init:
Version: 0.19.0
GitCommit: de40ad0
* Kolla-Ansible version (e.g. `git head or tag or stable branch` or pip package version if using release):
- kolla-ansible installed according to: https://docs.openstack.org/kolla-ansible/zed/user/quickstart.html), in particular:
$ pip install -U 'ansible>=4,<6'
$ pip install git+https://opendev.org/openstack/kolla-ansible@stable/zed
- kolla-ansible was run using virtual environment, strictly following the instructions form quickstart.html
* Docker image Install type of Openstack containers (source/binary):
binary
* Docker image distribution: quay.io, zed-debian-bullseye-aarch64, e.g. (from Ansible log during installation)
changed: [ost64] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'quay.io/openstack.kolla/fluentd:zed-debian-bullseye-aarch64', 'environment>
* Are you using official images from Docker Hub or self built?
I assume they are from quay.io (see above).
************************************************
3. ml2_conf.ini and nova-compute.conf is configured as follows
(Note: vlans enabled just for any case)
************************************************
=========== ml2_conf.ini
$ sudo tee /etc/kolla/config/neutron/ml2_conf.ini << EOT
[ml2]
type_drivers = flat,vlan
tenant_network_types = vxlan
[ml2_type_vlan]
network_vlan_ranges = physnet1:100:200
[ml2_type_flat]
flat_networks = physnet1
EOT
=========== nova-compute.conf
$ sudo tee /etc/kolla/config/nova/nova-compute.conf << EOT
[DEFAULT]
resume_guests_state_on_host_boot = true
*************************************************
4. globals.yml and inventory multinode files
*************************************************
4.1 globals.yml - basic settings
================================
Notes:
- remaining setting are default
- complete globals.yml - in attachment
------------
kolla_base_distro: "debian"
#openstack_release: "zed" <==== kolla-ansible has been generated for zed
openstack_tag_suffix: "-aarch64"
kolla_internal_vip_address: "192.168.1.60" <==== comment: ping 192.168.1.60 works well
network_interface: "veth0"
neutron_external_interface: "veth1"
enable_neutron_provider_networks: "yes"
#nova_compute_virt_type: "kvm" <===== comment: default (kvm) is left
4.2 inventory multinode - basic settings
========================================
Notes:
- remaining setting are default
- complete multinode - in attachment
------------
# These initial groups are the only groups required to be modified. The
# additional groups are for more control of the environment.
[control]
# These hostname must be resolvable from your deployment host
#control01
#control02
#control03
ost64 ansible_user=ubuntu ansible_password=ubuntu ansible_become=true
# The above can also be specified as follows:
#control[01:03] ansible_user=kolla
# The network nodes are where your l3-agent and loadbalancers will run
# This can be the same as a host in the control group
[network]
#network01
#network02
ost64
[compute]
#compute01
ost[61:63] ansible_user=ubuntu ansible_password=ubuntu ansible_become=true
ost64
[monitoring]
ost64
# When compute nodes and control nodes use different interfaces,
# you need to comment out "api_interface" and other interfaces from the globals.yml
# and specify like below:
#compute01 neutron_external_interface=eth0 api_interface=em1 tunnel_interface=em1
[storage]
#storage01
#ost61
[deployment]
localhost ansible_connection=local
=================================================
5. HttpException - message from init-runonce on the kolla-nsible node
(Note: cirros image has been stored successfully in glance and is reported as 'active' by openstack image list.)
=================================================
...
Configuring neutron.
HttpException: 503: Server Error for url: http://192.168.1.60:9696/v2.0/routers, No server is available to handle this request.: 503 Service Unavailable
=================================================================================
6. Status of containers reported by sudo docker ps -a on the control node
=================================================================================
ubuntu@ost64:~$ sudo docker ps -a
CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES
4f6598955a8d quay.io/openstack.kolla/horizon:zed-debian-bullseye-aarch64 "dumb-init --single-…" 4 hours ago Up 4 hours (healthy) horizon
77802d295317 quay.io/openstack.kolla/heat-engine:zed-debian-bullseye-aarch64 "dumb-init --single-…" 4 hours ago Up 4 hours (healthy) heat_engine
23dcd7083007 quay.io/openstack.kolla/heat-api-cfn:zed-debian-bullseye-aarch64 "dumb-init --single-…" 4 hours ago Up 4 hours (healthy) heat_api_cfn
3c7d7fb7dd85 quay.io/openstack.kolla/heat-api:zed-debian-bullseye-aarch64 "dumb-init --single-…" 4 hours ago Up 4 hours (healthy) heat_api
86420274224a quay.io/openstack.kolla/neutron-metadata-agent:zed-debian-bullseye-aarch64 "dumb-init --single-…" 4 hours ago Up 4 hours (healthy) neutron_metadata_agent
98a0f60fe048 quay.io/openstack.kolla/neutron-l3-agent:zed-debian-bullseye-aarch64 "dumb-init --single-…" 4 hours ago Up 4 hours (healthy) neutron_l3_agent
359fce2450aa quay.io/openstack.kolla/neutron-dhcp-agent:zed-debian-bullseye-aarch64 "dumb-init --single-…" 4 hours ago Up 4 hours (healthy) neutron_dhcp_agent
4c5e18147b87 quay.io/openstack.kolla/neutron-openvswitch-agent:zed-debian-bullseye-aarch64 "dumb-init --single-…" 4 hours ago Up 4 hours (healthy) neutron_openvswitch_agent
b5218aa4fb46 quay.io/openstack.kolla/neutron-server:zed-debian-bullseye-aarch64 "dumb-init --single-…" 4 hours ago Up 8 seconds (health: starting) neutron_server
aed0c1ec47a4 quay.io/openstack.kolla/openvswitch-vswitchd:zed-debian-bullseye-aarch64 "dumb-init --single-…" 4 hours ago Up 4 hours (healthy) openvswitch_vswitchd
463c6e360b8d quay.io/openstack.kolla/openvswitch-db-server:zed-debian-bullseye-aarch64 "dumb-init --single-…" 4 hours ago Up 4 hours (healthy) openvswitch_db
a53f653df1d1 quay.io/openstack.kolla/nova-compute:zed-debian-bullseye-aarch64 "dumb-init --single-…" 4 hours ago Up 4 hours (healthy) nova_compute
ce794c8d037c quay.io/openstack.kolla/nova-libvirt:zed-debian-bullseye-aarch64 "dumb-init --single-…" 4 hours ago Up 4 hours (healthy) nova_libvirt
16285faf8013 quay.io/openstack.kolla/nova-ssh:zed-debian-bullseye-aarch64 "dumb-init --single-…" 4 hours ago Up 4 hours (healthy) nova_ssh
c1e51c6bff9d quay.io/openstack.kolla/nova-novncproxy:zed-debian-bullseye-aarch64 "dumb-init --single-…" 4 hours ago Up 4 hours (healthy) nova_novncproxy
ca27ffc8d401 quay.io/openstack.kolla/nova-conductor:zed-debian-bullseye-aarch64 "dumb-init --single-…" 4 hours ago Up 4 hours (healthy) nova_conductor
2ca934ee1aa6 quay.io/openstack.kolla/nova-api:zed-debian-bullseye-aarch64 "dumb-init --single-…" 4 hours ago Up 4 hours (healthy) nova_api
f168dc3cc91f quay.io/openstack.kolla/nova-scheduler:zed-debian-bullseye-aarch64 "dumb-init --single-…" 4 hours ago Up 4 hours (healthy) nova_scheduler
690628e6cdb9 quay.io/openstack.kolla/placement-api:zed-debian-bullseye-aarch64 "dumb-init --single-…" 4 hours ago Up 4 hours (healthy) placement_api
b2676ebcc73a quay.io/openstack.kolla/glance-api:zed-debian-bullseye-aarch64 "dumb-init --single-…" 4 hours ago Up 4 hours (healthy) glance_api
2d5e935b8d4c quay.io/openstack.kolla/keystone:zed-debian-bullseye-aarch64 "dumb-init --single-…" 4 hours ago Up 4 hours (healthy) keystone
7a76f857056e quay.io/openstack.kolla/keystone-fernet:zed-debian-bullseye-aarch64 "dumb-init --single-…" 4 hours ago Up 4 hours (healthy) keystone_fernet
75c27f2df24f quay.io/openstack.kolla/keystone-ssh:zed-debian-bullseye-aarch64 "dumb-init --single-…" 4 hours ago Up 4 hours (healthy) keystone_ssh
ecbe00610982 quay.io/openstack.kolla/rabbitmq:zed-debian-bullseye-aarch64 "dumb-init --single-…" 4 hours ago Up 4 hours (healthy) rabbitmq
bfe9f55e263a quay.io/openstack.kolla/memcached:zed-debian-bullseye-aarch64 "dumb-init --single-…" 4 hours ago Up 4 hours (healthy) memcached
86a6a9a81e8d quay.io/openstack.kolla/mariadb-clustercheck:zed-debian-bullseye-aarch64 "dumb-init --single-…" 4 hours ago Up 4 hours mariadb_clustercheck
b377bd08e204 quay.io/openstack.kolla/mariadb-server:zed-debian-bullseye-aarch64 "dumb-init -- kolla_…" 4 hours ago Up 4 hours (healthy) mariadb
13c6eef85c13 quay.io/openstack.kolla/keepalived:zed-debian-bullseye-aarch64 "dumb-init --single-…" 4 hours ago Up 4 hours keepalived
5f0939c125a8 quay.io/openstack.kolla/haproxy:zed-debian-bullseye-aarch64 "dumb-init --single-…" 4 hours ago Up 4 hours (healthy) haproxy
8539f6ebdc7a quay.io/openstack.kolla/cron:zed-debian-bullseye-aarch64 "dumb-init --single-…" 4 hours ago Up 4 hours cron
1fbe92f1c594 quay.io/openstack.kolla/kolla-toolbox:zed-debian-bullseye-aarch64 "dumb-init --single-…" 4 hours ago Up 4 hours kolla_toolbox
444b3c2dd8b7 quay.io/openstack.kolla/fluentd:zed-debian-bullseye-aarch64 "dumb-init --single-…" 4 hours ago Up 4 hours fluentd
========================================================
7. Output form docker neutron_server inspect
========================================================
see attachment
=================================================================================
8. Log form neutron_server container (opening part of the log and ... the last part of the log)
=================================================================================
see attachment
############# end of problem description ###############
** Affects: neutron
Importance: Undecided
Status: New
--
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to neutron.
https://bugs.launchpad.net/bugs/2042598
Title:
neutron_server container suspended in health:starting state
Status in neutron:
New
Bug description:
I installed OpenStack (zed) on a Raspberry Pi cluster with kolla-
ansible (version tagged for zed), all containers are healthy except
the neutron_server which is suspended in 'health: starting' state.
Network related part of OpenStack does not work. Some other commands
commands work as expected (e.g., can create an image which is reported
by openstack image list as 'active').
There are four Raspberry Pi 4B in the cluster (2 x 4GB RAM and 2 x 8GB RAM). They run Debian 11 (bullseaye) and kolla-ansible has been used for the installation.
Notably, I'm using a specific configuration of networking on my Pis to mimic two network interfaces on each host as kolla-ansible expects. These are provided as interfaces of veth pairs (more details on that below, too).
Below, one can find:
1. configuration commands I used to configure my Pi hosts (this panel)
2. environment details related to the Pis (the one serving as controller in OpenStack) and kolla-ansible install information (this panel)
3. ml2_conf.ini and nova-compute.conf configuration used in kolla-ansible
4. kolla-ansible files: globals.yml (4.1) and inventory multinode (4.2)
- changed parts - this panel
- complete versions - attachments
5. HttpException: 503 message from running init-runonce (kolla-ansible test script for new installation) (this panel)
6. status of containers on the control node as reported by 'docker ps -a' (this panel)
7. output form docker neutron_server inspect command (attachment)
8. log form neutron_server container (attachment)
*************************************************************
1. Debian configuration on the Pis
*************************************************************
Selected details fo the configuration are given in the following.
Basically, most of them are needed to configure Pis' host networking
using netplan. Another one relates to qemu-kvm.
(Note: initial configs to enable ssh access should be done locally (keyboard, monitor) on each Pi, in particular:
PermitRootLogin yes
PasswordAuthentication yes
I skip the details of enabling ssh access, though. Below, I assume ssh acces as a regular (non-root) user.
)
=========== Preparation for host networking setup ===========
$ sudo apt-get remove unattended-upgrades -y
$ sudo apt-get update -y && sudo apt-get upgrade -y
- updating $PATH for a user
$ sudo tee -a ~/.bashrc << EOT
export PATH=$PATH:/usr/local/sbin:/usr/sbin:/sbin
EOT
$ source ~/.bashrc
- enable systemd-networkd and configure eth0 for ssh access (neede to use ssh; not neede if one does stuff locally, attaching keyboard and monitor to each Pi)
- enabling systemd-networkd
$ sudo mv /etc/network/interfaces /etc/network/interfaces.save
$ sudo mv /etc/network/interfaces.d /etc/network/interfaces.d.save
$ sudo systemctl enable systemd-networkd && sudo systemctl start systemd-networkd
$ sudo systemctl status systemd-networkd
- configure eth0 (in may case, I've configured static DHCP for each Pi on my DHCP server)
$ sudo tee /etc/systemd/network/20-wired.network << EOT
[Match]
Name=eth0
[Network]
DHCP=yes
EOT
- install netplan
$ sudo apt update && sudo apt -y install netplan.io
$ sudo reboot
- enable ip forwarding
$ sudo nano /etc/sysctl.conf
===> uncomment the line: net.ipv4.ip_forward=1
$sudo sysctl -p
========= Host networking setup ==========
- network setup on each Pi host - drawing:
192.168.1.xy/24 bez adresu IP
+---------+ +---------+
| veth0 | | veth1 | <==== network-interface and network-external-interface for kolla-ansible
+---------+ +---------+
| veth pairs |
+---------+ +---------+
| veth0br | | veth1br |
+---------+ +---------+
+-┴-----------------┴-+
| brmux |
+----------┬----------+
+---------+
| eth0 |
+---------+
========== commands for achieving the above network setup (on each Pi)
-------- first veth pair
$ sudo tee /etc/systemd/network/veth-openstack-net-itf-veth0.netdev << EOT
#network-interface in kolla-ansible globals.yml
[NetDev]
Name=veth0
Kind=veth
[Peer]
Name=veth0br
EOT
--------- second veth pair
$sudo tee /etc/systemd/network/veth-openstack-net-itf-veth1.netdev << EOT
#network-external-interface in kolla-ansible globals.yml
[NetDev]
Name=veth1
Kind=veth
[Peer]
Name=veth1br
EOT
--------- set the networking (wlan0 is not necessary, just for any
case)
sudo tee /etc/netplan/50-cloud-init.yaml << EOT
network:
version: 2
renderer: networkd
#-----------------------------------------#
# enable wlan for any case #
#-----------------------------------------#
# Interfejs wlan0 dostanie IPaddr z Linksysa przez DHCP.
wifis:
wlan0:
access-points:
FreshTomato06:
password: klasterek
dhcp4: true
optional: true
#-----------------------------------------#
# Kolla-Ansible required #
#-----------------------------------------#
# Interfaces
ethernets:
eth0:
dhcp4: false
dhcp6: false
# kolla-ansible network-interface
veth0:
addresses:
- 192.168.1.xy/24 # set appropriate address
nameservers:
addresses:
- 192.168.1.1 # dns on the local/home network router
- 8.8.8.8
routes:
- to: 0.0.0.0/0
via: 192.168.1.1
veth0br:
dhcp4: false
dhcp6: false
# kolla-ansible network-external-interface
veth1:
dhcp4: false
dhcp6: false
veth1br:
dhcp4: false
dhcp6: false
# Bridge - logically, a switch on the side of data center provider
network
bridges:
brmux:
interfaces:
- eth0
- veth0br
- veth1br
EOT
$ sudo netplan generate
$ sudo netplan apply
(now ssh using the new IP address of veth0)
============ Remaining Pi configurations ============
$ sudo visudo ==> add NOPASSWD: ALL as shown below
# Allow members of group sudo to execute any command
%sudo ALL=(ALL:ALL) NOPASSWD: ALL
$ sudo apt-get update
$ sudo apt-get install -y qemu-kvm
$ sudo apt-get upgrade -y
- on notification:
A new version (/tmp/tmp.EofC83AObD) of configuration file /etc/ssh/sshd_config is available
I choose ==> keep the local version currently installed
$ sudo apt-get install sshpass
$ sudo apt-get install ufw
$ sudo apt-get install net-tools <=== not necessary, I install it on the Pis just for any case
$ sudo nano /etc/sysctl.conf
===> uncomment the line: net.ipv4.ip_forward=1
$ sudo sysctl -p
************ End of Debian configuration *********
**************************************************
2. Environment details
**************************************************
( when related to a Pi this is the one serving as the control node)
* OS (e.g. from /etc/os-release):
ubuntu@ost64:~$ cat /etc/os-release
PRETTY_NAME="Debian GNU/Linux 11 (bullseye)"
NAME="Debian GNU/Linux"
VERSION_ID="11"
VERSION="11 (bullseye)"
VERSION_CODENAME=bullseye
ID=debian
HOME_URL="https://www.debian.org/"
SUPPORT_URL="https://www.debian.org/support"
BUG_REPORT_URL="https://bugs.debian.org/"
* Kernel (e.g. `uname -a`):
ubuntu@ost64:~$ uname -a
Linux ost64 5.10.0-20-arm64 #1 SMP Debian 5.10.158-2 (2022-12-13) aarch64 GNU/Linux
* Docker version if applicable (e.g. `docker version`):
ubuntu@ost64:~$ sudo docker version
Client: Docker Engine - Community
Version: 24.0.7
API version: 1.43
Go version: go1.20.10
Git commit: afdd53b
Built: Thu Oct 26 09:08:29 2023
OS/Arch: linux/arm64
Context: default
Server: Docker Engine - Community
Engine:
Version: 24.0.7
API version: 1.43 (minimum version 1.12)
Go version: go1.20.10
Git commit: 311b9ff
Built: Thu Oct 26 09:08:29 2023
OS/Arch: linux/arm64
Experimental: false
containerd:
Version: 1.6.24
GitCommit: 61f9fd88f79f081d64d6fa3bb1a0dc71ec870523
runc:
Version: 1.1.9
GitCommit: v1.1.9-0-gccaecfc
docker-init:
Version: 0.19.0
GitCommit: de40ad0
* Kolla-Ansible version (e.g. `git head or tag or stable branch` or pip package version if using release):
- kolla-ansible installed according to: https://docs.openstack.org/kolla-ansible/zed/user/quickstart.html), in particular:
$ pip install -U 'ansible>=4,<6'
$ pip install git+https://opendev.org/openstack/kolla-ansible@stable/zed
- kolla-ansible was run using virtual environment, strictly following the instructions form quickstart.html
* Docker image Install type of Openstack containers (source/binary):
binary
* Docker image distribution: quay.io, zed-debian-bullseye-aarch64, e.g. (from Ansible log during installation)
changed: [ost64] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'quay.io/openstack.kolla/fluentd:zed-debian-bullseye-aarch64', 'environment>
* Are you using official images from Docker Hub or self built?
I assume they are from quay.io (see above).
************************************************
3. ml2_conf.ini and nova-compute.conf is configured as follows
(Note: vlans enabled just for any case)
************************************************
=========== ml2_conf.ini
$ sudo tee /etc/kolla/config/neutron/ml2_conf.ini << EOT
[ml2]
type_drivers = flat,vlan
tenant_network_types = vxlan
[ml2_type_vlan]
network_vlan_ranges = physnet1:100:200
[ml2_type_flat]
flat_networks = physnet1
EOT
=========== nova-compute.conf
$ sudo tee /etc/kolla/config/nova/nova-compute.conf << EOT
[DEFAULT]
resume_guests_state_on_host_boot = true
*************************************************
4. globals.yml and inventory multinode files
*************************************************
4.1 globals.yml - basic settings
================================
Notes:
- remaining setting are default
- complete globals.yml - in attachment
------------
kolla_base_distro: "debian"
#openstack_release: "zed" <==== kolla-ansible has been generated for zed
openstack_tag_suffix: "-aarch64"
kolla_internal_vip_address: "192.168.1.60" <==== comment: ping 192.168.1.60 works well
network_interface: "veth0"
neutron_external_interface: "veth1"
enable_neutron_provider_networks: "yes"
#nova_compute_virt_type: "kvm" <===== comment: default (kvm) is left
4.2 inventory multinode - basic settings
========================================
Notes:
- remaining setting are default
- complete multinode - in attachment
------------
# These initial groups are the only groups required to be modified. The
# additional groups are for more control of the environment.
[control]
# These hostname must be resolvable from your deployment host
#control01
#control02
#control03
ost64 ansible_user=ubuntu ansible_password=ubuntu ansible_become=true
# The above can also be specified as follows:
#control[01:03] ansible_user=kolla
# The network nodes are where your l3-agent and loadbalancers will run
# This can be the same as a host in the control group
[network]
#network01
#network02
ost64
[compute]
#compute01
ost[61:63] ansible_user=ubuntu ansible_password=ubuntu ansible_become=true
ost64
[monitoring]
ost64
# When compute nodes and control nodes use different interfaces,
# you need to comment out "api_interface" and other interfaces from the globals.yml
# and specify like below:
#compute01 neutron_external_interface=eth0 api_interface=em1 tunnel_interface=em1
[storage]
#storage01
#ost61
[deployment]
localhost ansible_connection=local
=================================================
5. HttpException - message from init-runonce on the kolla-nsible node
(Note: cirros image has been stored successfully in glance and is reported as 'active' by openstack image list.)
=================================================
...
Configuring neutron.
HttpException: 503: Server Error for url: http://192.168.1.60:9696/v2.0/routers, No server is available to handle this request.: 503 Service Unavailable
=================================================================================
6. Status of containers reported by sudo docker ps -a on the control node
=================================================================================
ubuntu@ost64:~$ sudo docker ps -a
CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES
4f6598955a8d quay.io/openstack.kolla/horizon:zed-debian-bullseye-aarch64 "dumb-init --single-…" 4 hours ago Up 4 hours (healthy) horizon
77802d295317 quay.io/openstack.kolla/heat-engine:zed-debian-bullseye-aarch64 "dumb-init --single-…" 4 hours ago Up 4 hours (healthy) heat_engine
23dcd7083007 quay.io/openstack.kolla/heat-api-cfn:zed-debian-bullseye-aarch64 "dumb-init --single-…" 4 hours ago Up 4 hours (healthy) heat_api_cfn
3c7d7fb7dd85 quay.io/openstack.kolla/heat-api:zed-debian-bullseye-aarch64 "dumb-init --single-…" 4 hours ago Up 4 hours (healthy) heat_api
86420274224a quay.io/openstack.kolla/neutron-metadata-agent:zed-debian-bullseye-aarch64 "dumb-init --single-…" 4 hours ago Up 4 hours (healthy) neutron_metadata_agent
98a0f60fe048 quay.io/openstack.kolla/neutron-l3-agent:zed-debian-bullseye-aarch64 "dumb-init --single-…" 4 hours ago Up 4 hours (healthy) neutron_l3_agent
359fce2450aa quay.io/openstack.kolla/neutron-dhcp-agent:zed-debian-bullseye-aarch64 "dumb-init --single-…" 4 hours ago Up 4 hours (healthy) neutron_dhcp_agent
4c5e18147b87 quay.io/openstack.kolla/neutron-openvswitch-agent:zed-debian-bullseye-aarch64 "dumb-init --single-…" 4 hours ago Up 4 hours (healthy) neutron_openvswitch_agent
b5218aa4fb46 quay.io/openstack.kolla/neutron-server:zed-debian-bullseye-aarch64 "dumb-init --single-…" 4 hours ago Up 8 seconds (health: starting) neutron_server
aed0c1ec47a4 quay.io/openstack.kolla/openvswitch-vswitchd:zed-debian-bullseye-aarch64 "dumb-init --single-…" 4 hours ago Up 4 hours (healthy) openvswitch_vswitchd
463c6e360b8d quay.io/openstack.kolla/openvswitch-db-server:zed-debian-bullseye-aarch64 "dumb-init --single-…" 4 hours ago Up 4 hours (healthy) openvswitch_db
a53f653df1d1 quay.io/openstack.kolla/nova-compute:zed-debian-bullseye-aarch64 "dumb-init --single-…" 4 hours ago Up 4 hours (healthy) nova_compute
ce794c8d037c quay.io/openstack.kolla/nova-libvirt:zed-debian-bullseye-aarch64 "dumb-init --single-…" 4 hours ago Up 4 hours (healthy) nova_libvirt
16285faf8013 quay.io/openstack.kolla/nova-ssh:zed-debian-bullseye-aarch64 "dumb-init --single-…" 4 hours ago Up 4 hours (healthy) nova_ssh
c1e51c6bff9d quay.io/openstack.kolla/nova-novncproxy:zed-debian-bullseye-aarch64 "dumb-init --single-…" 4 hours ago Up 4 hours (healthy) nova_novncproxy
ca27ffc8d401 quay.io/openstack.kolla/nova-conductor:zed-debian-bullseye-aarch64 "dumb-init --single-…" 4 hours ago Up 4 hours (healthy) nova_conductor
2ca934ee1aa6 quay.io/openstack.kolla/nova-api:zed-debian-bullseye-aarch64 "dumb-init --single-…" 4 hours ago Up 4 hours (healthy) nova_api
f168dc3cc91f quay.io/openstack.kolla/nova-scheduler:zed-debian-bullseye-aarch64 "dumb-init --single-…" 4 hours ago Up 4 hours (healthy) nova_scheduler
690628e6cdb9 quay.io/openstack.kolla/placement-api:zed-debian-bullseye-aarch64 "dumb-init --single-…" 4 hours ago Up 4 hours (healthy) placement_api
b2676ebcc73a quay.io/openstack.kolla/glance-api:zed-debian-bullseye-aarch64 "dumb-init --single-…" 4 hours ago Up 4 hours (healthy) glance_api
2d5e935b8d4c quay.io/openstack.kolla/keystone:zed-debian-bullseye-aarch64 "dumb-init --single-…" 4 hours ago Up 4 hours (healthy) keystone
7a76f857056e quay.io/openstack.kolla/keystone-fernet:zed-debian-bullseye-aarch64 "dumb-init --single-…" 4 hours ago Up 4 hours (healthy) keystone_fernet
75c27f2df24f quay.io/openstack.kolla/keystone-ssh:zed-debian-bullseye-aarch64 "dumb-init --single-…" 4 hours ago Up 4 hours (healthy) keystone_ssh
ecbe00610982 quay.io/openstack.kolla/rabbitmq:zed-debian-bullseye-aarch64 "dumb-init --single-…" 4 hours ago Up 4 hours (healthy) rabbitmq
bfe9f55e263a quay.io/openstack.kolla/memcached:zed-debian-bullseye-aarch64 "dumb-init --single-…" 4 hours ago Up 4 hours (healthy) memcached
86a6a9a81e8d quay.io/openstack.kolla/mariadb-clustercheck:zed-debian-bullseye-aarch64 "dumb-init --single-…" 4 hours ago Up 4 hours mariadb_clustercheck
b377bd08e204 quay.io/openstack.kolla/mariadb-server:zed-debian-bullseye-aarch64 "dumb-init -- kolla_…" 4 hours ago Up 4 hours (healthy) mariadb
13c6eef85c13 quay.io/openstack.kolla/keepalived:zed-debian-bullseye-aarch64 "dumb-init --single-…" 4 hours ago Up 4 hours keepalived
5f0939c125a8 quay.io/openstack.kolla/haproxy:zed-debian-bullseye-aarch64 "dumb-init --single-…" 4 hours ago Up 4 hours (healthy) haproxy
8539f6ebdc7a quay.io/openstack.kolla/cron:zed-debian-bullseye-aarch64 "dumb-init --single-…" 4 hours ago Up 4 hours cron
1fbe92f1c594 quay.io/openstack.kolla/kolla-toolbox:zed-debian-bullseye-aarch64 "dumb-init --single-…" 4 hours ago Up 4 hours kolla_toolbox
444b3c2dd8b7 quay.io/openstack.kolla/fluentd:zed-debian-bullseye-aarch64 "dumb-init --single-…" 4 hours ago Up 4 hours fluentd
========================================================
7. Output form docker neutron_server inspect
========================================================
see attachment
=================================================================================
8. Log form neutron_server container (opening part of the log and ... the last part of the log)
=================================================================================
see attachment
############# end of problem description ###############
To manage notifications about this bug go to:
https://bugs.launchpad.net/neutron/+bug/2042598/+subscriptions
Follow ups