yahoo-eng-team team mailing list archive
-
yahoo-eng-team team
-
Mailing list archive
-
Message #95814
[Bug 2109676] [NEW] ovs_rcu(urcu11)|WARN|blocked xxxx ms waiting for handler1 to quiesce
Public bug reported:
Hello.
I am using OVN for my system. I have ovs_rcu(urcu11)|WARN|blocked xxxx
ms waiting for handler1 to quiesce when apply qos policy for networks.
How to reproduce this bug:
Create 2 instances: one uses a VLAN IP, and the other uses Geneve(with
floating IP), both same subnet
Run iperf -s on the first instance and run iperf -c <first instance IP>
-t 360 on the second to connect to it.
Create a network QoS policy:
openstack network qos policy create --project 5d895a6d62d241e495f9e29c3b04644b bw-limiter
openstack network qos rule create --type bandwidth-limit --max-kbps 3000 --max-burst-kbits 2400 --egress bw-limiter
openstack network qos rule create --type bandwidth-limit --max-kbps 3000 --max-burst-kbits 2400 --ingress bw-limiter
openstack network set --qos-policy bw-limiter mynetwork
Perform a hard reboot of the second instance, then check the logs on the
compute node hosting it:
tail -f /var/log/kolla/openvswitch/ovs-vswitchd.log
You will see logs like:
2025-04-30T07:53:42.726Z|00002|ovs_rcu(urcu11)|WARN|blocked 2000 ms waiting for handler1 to quiesce
2025-04-30T07:53:43.037Z|00132|ovs_rcu|WARN|blocked 2000 ms waiting for handler1 to quiesce
2025-04-30T07:53:44.726Z|00003|ovs_rcu(urcu11)|WARN|blocked 4000 ms waiting for handler1 to quiesce
...
Environment:
Ubuntu 24.04
OpenStack 2025.1
Deployment tool: Kolla-Ansible (2025 master branch)
** Affects: neutron
Importance: Undecided
Status: New
** Description changed:
Hello.
I am using OVN for my system. I have ovs_rcu(urcu11)|WARN|blocked xxxx
- ms waiting for handler1 to quiesce when apply qos policy for network
+ ms waiting for handler1 to quiesce when apply qos policy for networks.
How to reproduce this bug:
Create 2 instances: one uses a VLAN IP, and the other uses Geneve(with
floating IP), both same subnet
Run iperf -s on the first instance and run iperf -c <first instance IP>
-t 360 on the second to connect to it.
Create a network QoS policy:
openstack network qos policy create --project 5d895a6d62d241e495f9e29c3b04644b bw-limiter
openstack network qos rule create --type bandwidth-limit --max-kbps 3000 --max-burst-kbits 2400 --egress bw-limiter
openstack network qos rule create --type bandwidth-limit --max-kbps 3000 --max-burst-kbits 2400 --ingress bw-limiter
openstack network set --qos-policy bw-limiter mynetwork
Perform a hard reboot of the second instance, then check the logs on the
compute node hosting it:
tail -f /var/log/kolla/openvswitch/ovs-vswitchd.log
You will see logs like:
2025-04-30T07:53:42.726Z|00002|ovs_rcu(urcu11)|WARN|blocked 2000 ms waiting for handler1 to quiesce
2025-04-30T07:53:43.037Z|00132|ovs_rcu|WARN|blocked 2000 ms waiting for handler1 to quiesce
2025-04-30T07:53:44.726Z|00003|ovs_rcu(urcu11)|WARN|blocked 4000 ms waiting for handler1 to quiesce
...
Environment:
Ubuntu 24.04
OpenStack 2025.1
Deployment tool: Kolla-Ansible (2025 master branch)
--
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to neutron.
https://bugs.launchpad.net/bugs/2109676
Title:
ovs_rcu(urcu11)|WARN|blocked xxxx ms waiting for handler1 to quiesce
Status in neutron:
New
Bug description:
Hello.
I am using OVN for my system. I have ovs_rcu(urcu11)|WARN|blocked xxxx
ms waiting for handler1 to quiesce when apply qos policy for networks.
How to reproduce this bug:
Create 2 instances: one uses a VLAN IP, and the other uses Geneve(with
floating IP), both same subnet
Run iperf -s on the first instance and run iperf -c <first instance
IP> -t 360 on the second to connect to it.
Create a network QoS policy:
openstack network qos policy create --project 5d895a6d62d241e495f9e29c3b04644b bw-limiter
openstack network qos rule create --type bandwidth-limit --max-kbps 3000 --max-burst-kbits 2400 --egress bw-limiter
openstack network qos rule create --type bandwidth-limit --max-kbps 3000 --max-burst-kbits 2400 --ingress bw-limiter
openstack network set --qos-policy bw-limiter mynetwork
Perform a hard reboot of the second instance, then check the logs on
the compute node hosting it:
tail -f /var/log/kolla/openvswitch/ovs-vswitchd.log
You will see logs like:
2025-04-30T07:53:42.726Z|00002|ovs_rcu(urcu11)|WARN|blocked 2000 ms waiting for handler1 to quiesce
2025-04-30T07:53:43.037Z|00132|ovs_rcu|WARN|blocked 2000 ms waiting for handler1 to quiesce
2025-04-30T07:53:44.726Z|00003|ovs_rcu(urcu11)|WARN|blocked 4000 ms waiting for handler1 to quiesce
...
Environment:
Ubuntu 24.04
OpenStack 2025.1
Deployment tool: Kolla-Ansible (2025 master branch)
To manage notifications about this bug go to:
https://bugs.launchpad.net/neutron/+bug/2109676/+subscriptions