Replies: 3 comments 6 replies
-
Can you share a folder with such segments files? This would allow us to identify what's special about it. |
Beta Was this translation helpful? Give feedback.
-
If consumers on a quorum queue consume a message but do not acknowledge it, quorum queues won't be able to reclaim disk space from that moment on. In other words, a single stack acknowledgement can lead to a similar behavior. Restarting your consumers would be an easy way to rest this hypothesis. |
Beta Was this translation helpful? Give feedback.
-
Most likely something like a faulty consumer held a lock on a message and
eventually released the lock. You didn’t really provide any information for
us to use to even provide a guess with
…On Sun, 5 Nov 2023 at 09:48, 4danmi ***@***.***> wrote:
Hi,
After a few days, the folder was cleared without any explanation.
Do you have explanation for this behavior?
—
Reply to this email directly, view it on GitHub
<#9853 (reply in thread)>,
or unsubscribe
<https://github.com/notifications/unsubscribe-auth/AAJAHFAIVUT7FZJNZ3UIMBDYC5OIDAVCNFSM6AAAAAA62PKQLCVHI2DSMVQWIX3LMV43SRDJONRXK43TNFXW4Q3PNVWWK3TUHM3TINZYGE3DK>
.
You are receiving this because you commented.Message ID:
***@***.***
com>
|
Beta Was this translation helpful? Give feedback.
Uh oh!
There was an error while loading. Please reload this page.
-
Hi,
We had similar issue to #8860, though we upgraded a month ago from version 3.8.7 to version 3.12.4 (not in place, but in a completely different machines). After the upgrade it seemed that the issued had been resolved, but we now found that the segment files folder of one of the queues keeps growing uncontrollabley. This folder belongs (according to the config file inside it) to a quorum queue that is replicated on 3 nodes, and the folder keeps growing on all of its nodes.
As far as we can tell, it happens only in a single queue (out of hundreds), though this queue has much higher message rate compare to most other queues in the cluster, but other queues with slightly lower message rate don't seem to have this issue. The queue is almost always emtpy as the consumers keeps reading from it with no appearant problem. We also have the same queue with the same configuration and data, running as a backup in a different cluster that doesn't show the same behaviour.
We considered removing each one of the nodes from this specific queue, deleting the segements file folder and re-adding it back to the queue (using
rabbitmq-queues
), but we afraid that it might break the queue and we can't risk it at this time.All of our consumers (also for other queues) are using AMQP with Python's pika library. The server is running RHEL 8.4 with Erlang version 26.0.2.
The only solution we found for this problem is to reset one of the cluster nodes, and then join it back to the cluster, but during this process we risk the cluster availablity. We already performed this operation once, but the problem seems to occur again after a few hours.
We also considered upgrading to version 3.12.8, but non of the bug fixes in the changelog seems to address something that might resolve this problem.
What else can we do?
Thanks for your help.
Beta Was this translation helpful? Give feedback.
All reactions