-
DescriptionAfter 4-5 h of inactivity a tiny amount of published messages leads to a doubling the RAM usage. Reproduction
ExpectedThe cluster performs like 4 hours before using about the same amount of RAM (14 GB). ActualRight after the first messages are published each cluster node allocates another 15 GB RAM!! within 5 minutes hitting the high-watermark. Log (Info level)No log entries till the memory high watermark is hit. Rabbit MQ setupRabbit MQ-Version: 3.8.1 (Erlang 22.1.6) Reproduction setup
System load before errorCPU(user): < 10 % System load after "resuming"CPU(user): < 70 % |
Beta Was this translation helpful? Give feedback.
Replies: 5 comments 40 replies
-
Hello, and thanks for taking the time to report this. Since it seems as though you can easily reproduce this behavior, could you please do so using the latest version of Erlang (23.2.3) and RabbitMQ? (3.8.12-beta.1). It would also be extremely helpful if you could export your cluster definitions to a file and attach them so we can see exactly how everything is set up. Thanks! |
Beta Was this translation helpful? Give feedback.
-
You can ask the node what uses the memory. |
Beta Was this translation helpful? Give feedback.
-
We filled up the nodes again. Less but bigger messages (~40M) rabbitmq admin uimemory breakdown (node 1)
erlang grafana memory board (node 1) |
Beta Was this translation helpful? Give feedback.
-
We were able to reproduce the issue after 4-8h of inactivity followed by publishing messages. During this resume "event" the RabbitMQ grafana memory statistics are showing that every RabbitMQ node "tries" to allocate almost 70 GB at peak level. The VMs have 32 GB RAM.
As you can see there's not much going on till a smalll amount of messages are produced: Unfortunately we not only have a simulated testing issue with this. It already happened in production. Any help would be greatly appreciated. |
Beta Was this translation helpful? Give feedback.
-
This has been resolved part of VESC-1002 (private). The public resolution is here: erlang/otp#4571 (comment) If you agree @motmot80, please mark this comment as the answer 👍🏻 |
Beta Was this translation helpful? Give feedback.
This has been resolved part of VESC-1002 (private).
The public resolution is here: erlang/otp#4571 (comment)
If you agree @motmot80, please mark this comment as the answer 👍🏻