Replies: 4 comments 4 replies
-
I should point out the upgrades were performed as a rolling upgrade. Node 1 was taken offline, packages upgraded, and added back to the cluster, followed by Node 2, etc. |
Beta Was this translation helpful? Give feedback.
-
I will convert this issue to a GitHub discussion. Currently GitHub will automatically close and lock the issue even though your question will be transferred and responded to elsewhere. This is to let you know that we do not intend to ignore this but this is how the current GitHub conversion mechanism makes it seem for the users :( |
Beta Was this translation helpful? Give feedback.
-
very likely means a classic queue leader process is not running. This has nothing to do with CLI tools.
means the node responded with an authentication code that usually means a mismatched cookie. Consider restarting nodes one by one and trying rebalancing again. Note that classic mirrored queues should be considered a deprecated feature that will be removed in a future version. |
Beta Was this translation helpful? Give feedback.
-
Hello,
Here’s the reference to converting this to a GitHub Discussion…
***@***.***
***@***.***<https://github.com/michaelklishin>
michaelklishin<https://github.com/michaelklishin>
on Mar 17, 2021<#2903 (comment)>
Maintainer
I will convert this issue to a GitHub discussion. Currently GitHub will automatically close and lock the issue<community/community#3197> even though your question will be transferred and responded to elsewhere. This is to let you know that we do not intend to ignore this but this is how the current GitHub conversion mechanism makes it seem for the users :(
.
Thanks for the info you provided. I will pass this on to the developers.
From: Michael Klishin ***@***.***>
Sent: Wednesday, May 3, 2023 1:41 PM
To: rabbitmq/rabbitmq-server ***@***.***>
Cc: Byron Carlson ***@***.***>; Comment ***@***.***>
Subject: Re: [rabbitmq/rabbitmq-server] rabbitmq-queues rebalance fails after an upgrade from 3.8.5 to 3.8.14 (Discussion #2903)
[External email -- Courriel externe]
I do not see any references to GitHub discussions in my reply from two years ago.
There were hardly any changes to classic mirrored queues since then and won't be in the future (classic non-mirrored queues have seen a non-trivial investment and the plan is to continue with their improvements).
My best guess is that this CM queue ended up without an electable leader<https://rabbitmq.com/ha.html#unsynchronised-mirrors>. Disabling mirroring may or may not help. Deletion and redeclaration of the queue should.
Quorum queues do not have this problem (but require a quorum of nodes to be online). They have been around for several years and there is a set of pretty well understood migration steps<https://blog.rabbitmq.com/posts/2023/02/quorum-queues-migration/> besides Blue Green deployment.
Classic mirrored queues will be completely removed in 4.0 which is expected early next year. This was announced over one year ago.
—
Reply to this email directly, view it on GitHub<#2903 (reply in thread)>, or unsubscribe<https://github.com/notifications/unsubscribe-auth/A7S3FLRMPMQXUPUGBXTNV33XEKRELANCNFSM4ZLYNRHQ>.
You are receiving this because you commented.Message ID: ***@***.******@***.***>>
Click here<https://www.mailcontrol.com/sr/UFCPharrJYzGX2PQPOmvUkHxlm0SZLPCtfsSz5svJHlr1wdUDsztnuDzLl8J6EdFfL5n6GRS7DmNI5tH4SJ_Mg==> to report this email as spam.
|
Beta Was this translation helpful? Give feedback.
Uh oh!
There was an error while loading. Please reload this page.
-
I upgraded a few RabbitMQ clusters from 3.8.5 to 3.8.14 using official deb packages today and noticed an issue with the
rabbitmq-queues
cli.From a user perspective it errors out with:
In the crash logs this is seen as:
This made me think it's a cookie issue, but I am able to run
rabbitmqctl
andrabbitmq-diagnostics
without issue.In addition, the info.log looks like the command connected in and started the process:
Erlang Version: 23.2.7
RabbitMQ 3.8.14
Ubuntu 16.04.6 LTS
Interestingly I was able to reproduce this on all nodes that upgraded from 3.8.5 to 3.8.14, but on nodes where the cluster was upgraded from 3.7.16, the problem does not exist.
Please let me know if there is any other info you'd like me to collect.
Beta Was this translation helpful? Give feedback.
All reactions