From 3800287f9ed97431aad5a75391cee9f86ecd76cb Mon Sep 17 00:00:00 2001 From: technowhizz <7688823+technowhizz@users.noreply.github.com> Date: Tue, 19 Nov 2024 14:51:15 +0000 Subject: [PATCH 1/2] Add warning for OVS clusters with rabbitmq upgrade --- doc/source/operations/upgrading-openstack.rst | 4 ++++ 1 file changed, 4 insertions(+) diff --git a/doc/source/operations/upgrading-openstack.rst b/doc/source/operations/upgrading-openstack.rst index 43c6efa6e..15e68d5c2 100644 --- a/doc/source/operations/upgrading-openstack.rst +++ b/doc/source/operations/upgrading-openstack.rst @@ -225,6 +225,10 @@ Some errors have been observed in testing when the upgrades are perfomed back-to-back. A 200s delay eliminates this issue. On particularly large or slow deployments, consider increasing this timeout. +Additionally errors have been observed at sites with OVS networking where after +the upgrade, tenant networking is broken and requires a reset of RabbitMQ. This +can be done by running the rabbitmq-reset playbook. + .. code-block:: bash kayobe overcloud service configuration generate --node-config-dir /tmp/ignore -kt none From 161cbe6fefc1a4c2b9ead33fe7087a01419ec4c6 Mon Sep 17 00:00:00 2001 From: technowhizz <7688823+technowhizz@users.noreply.github.com> Date: Tue, 19 Nov 2024 14:51:44 +0000 Subject: [PATCH 2/2] Fix typo in upgrade docs --- doc/source/operations/upgrading-openstack.rst | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/doc/source/operations/upgrading-openstack.rst b/doc/source/operations/upgrading-openstack.rst index 15e68d5c2..3821280a3 100644 --- a/doc/source/operations/upgrading-openstack.rst +++ b/doc/source/operations/upgrading-openstack.rst @@ -221,7 +221,7 @@ to 3.12, then to 3.13 on Antelope before the Caracal upgrade. This upgrade should not cause an API outage (though it should still be considered "at risk"). -Some errors have been observed in testing when the upgrades are perfomed +Some errors have been observed in testing when the upgrades are performed back-to-back. A 200s delay eliminates this issue. On particularly large or slow deployments, consider increasing this timeout.