Skip to content

Navigation Menu

Sign in
Appearance settings

Search code, repositories, users, issues, pull requests...

Provide feedback

We read every piece of feedback, and take your input very seriously.

Saved searches

Use saved searches to filter your results more quickly

Appearance settings
Discussion options

Getting execptions from a client trying to consume from a quorum queue, client error state

Exception (541) Reason: "INTERNAL_ERROR - timed out consuming from quorum queue 'xxx' in vhost '/': {'%2F_yyy'}

logs on one Rabbit node state

2025-10-03 22:01:37.917907+02:00 [info] <0.863.0> queue 'xxx' in vhost '/': term mismatch - follower had entry at 658 with term 81 but not with term 82. Asking leader {'%2F_yyy','rabbit@158.rabbitmq.service.consul'} to resend from 659
2025-10-03 22:02:07.919957+02:00 [info] <0.863.0> queue 'xxx' in vhost '/': term mismatch - follower had entry at 658 with term 81 but not with term 82. Asking leader {'%2F_yyy,'rabbit@0158.rabbitmq.service.consul'} to resend from 659
2025-10-03 22:02:11.782388+02:00 [error] <0.188296.0> Error on AMQP connection <0.188296.0> (10.20.128.0:60716 -> 10.20.222.109:5672, vhost: '/', user: 'USER', state: running), channel 5:
2025-10-03 22:02:11.782388+02:00 [error] <0.188296.0> operation basic.consume caused a connection exception internal_error: "timed out consuming from quorum queue 'xxx' in vhost '/': {'%2F_yyy',\n 'rabbit@158.rabbitmq.service.consul'}"
2025-10-03 22:02:11.806517+02:00 [info] <0.188296.0> closing AMQP connection (10.20.128.0:60716 -> 10.20.222.109:5672, vhost: '/', user: 'USER', duration: '2M, 15s')

This happens again and again, and seems like Rabbit cannot restore

Happens after netowork partition. Any idea if this is a potential bug or wrong use of quorum queus

Reproduction steps

Hard to say, will add steps when I have found a way to reproduce

Expected behavior

Should be able to survive network partition

Additional context

Running version 4.1.1

You must be logged in to vote

Replies: 1 comment · 1 reply

Comment options

@malmetom that's not enough information for us to conclude much.

All we can tell is that a quorum queue follower is behind the leader, so it asks the leader for the delta to reinstall the (follower's) Raft log. We don't know anything about what happens on the leader.

Without full logs from all cluster nodes or a reasonably reliable way to reproduce we can only guess as to what's going on, and we do not guess in this community. Guessing is a very expensive way of troubleshooting distributed infrastructure.

This could be a different manifestation of #13101, which is a known open issue for quorum queues. #14237 and #14241 are two competing solutions that we'll get to some time after 4.2.0 and the Ra 3.0 work concluding. Anyone is welcome to test #14241, which is now again ready for review (as of less than 24 hours ago).

You must be logged in to vote
1 reply
@malmetom
Comment options

Thanks for quick response, I fully understand, sorry for bad input to the issue. I will collect logs and attach if I can't figure out the issue on my end, probably the setup causing this issue

Just one question, this is the status of the quorum,
image
think it looks strange that Last applied is so low but thats probably due to all partitions going on, this is a test environment, but it shouldn't be an issue to recover from this state right?

I really appreciate the support given in this community

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
2 participants
Converted from issue

This discussion was converted from issue #14675 on October 03, 2025 20:56.

Morty Proxy This is a proxified and sanitized view of the page, visit original site.