You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
We have Tempo Operator version 0.14.1-2 provided by Red Hat installed in our environment and we have created a TempoStack instance using below configuration.
With above configuration, it has deployed the TempoStack and other dependent component.
Here, i'm looking for 2 additional things as below:
Looking at above configuration, resources limits and quota is setup globally.
resources:
total:
limits:
cpu: '6'
memory: 15Gi
whenever we are trying to set resource quota for individually for each component e.g compactor i.e specs.template.compactor.resources.limits it is getting overwritten and due to which compactor pod is not getting enough compute resources and restarting with CrashloopBackoff.
We have noticed issues with the Grafana Tempo instance where some trace spans appeared to be missing (and were never linked to the parent trace).
Looking deeper, we noticed the following set of errors in the pods of the Tempo Stack instance:
Distributor pod:
level=error ts=2025-01-21T14:52:49.384256785Z caller=rate_limited_logger.go:27 msg="pusher failed to consume trace data" err="rpc error: code = FailedPrecondition desc = TRACE_TOO_LARGE: max size of trace (5000000) exceeded while adding 142414 bytes to trace 1b4722a6b98fd1da98f0370af4089927 for tenant single-tenant"
Ingestor pod:
level=warn ts=2025-01-21T20:28:55.532118428Z caller=server.go:1184 method=/tempopb.Pusher/PushBytesV2 duration=80.163µs msg=gRPC err="rpc error: code = FailedPrecondition desc = TRACE_TOO_LARGE: max size of trace (5000000) exceeded while adding 177635 bytes to trace c5ca65779266a32c8dd98a5cd54357f8 for tenant single-tenant"
Querier pod::
ResourceExhausted desc = grpc: received message after decompression larger than max (4993760 vs. 4194304)
From some research, it seems like we need to bump up the maximum trace size. By default, that is set to 5000000 (5 MiB).
As per here (https://grafana.com/docs/tempo/latest/configuration/#ingestion-limits), "overrides" can be used to increase this (there is caution against going to large, however). we have added parameter as below.
The issue with the querier pod seems to be due to there being a gRPC message size limit between TempoStack components. As suggested here (#1097), I think we need to change settings both in the tempo.yaml and tempo-query-frontend.yaml to increase these to at least the max_bytes_per_trace size.
we are trying to change the existing server stanza settings with below values however it is getting overwritten,
Can anyone help us here and suggest how to and where to change these settings.
Thanks in advance !!
The text was updated successfully, but these errors were encountered:
vsomwanshi
changed the title
unable to set grpc_server_max_recv_msg_size and grpc_server_max_send_msg_size in
unable to update grpc_server_max_recv_msg_size and grpc_server_max_send_msg_size in server stanza setting in TempoStack
Jan 24, 2025
vsomwanshi
changed the title
unable to update grpc_server_max_recv_msg_size and grpc_server_max_send_msg_size in server stanza setting in TempoStack
[Problem] Unable to update grpc_server_max_recv_msg_size and grpc_server_max_send_msg_size in server stanza setting in TempoStack
Jan 24, 2025
We have
Tempo Operator
version0.14.1-2 provided by Red Hat
installed in our environment and we have created aTempoStack
instance using below configuration.With above configuration, it has deployed the
TempoStack
and other dependent component.Here, i'm looking for 2 additional things as below:
whenever we are trying to set resource quota for individually for each component e.g
compactor
i.especs.template.compactor.resources.limits
it is getting overwritten and due to which compactor pod is not getting enough compute resources and restarting withCrashloopBackoff
.Looking deeper, we noticed the following set of errors in the pods of the Tempo Stack instance:
Distributor pod:
Ingestor pod:
Querier pod::
From some research, it seems like we need to bump up the maximum trace size. By default, that is set to 5000000 (5 MiB).
As per here (https://grafana.com/docs/tempo/latest/configuration/#ingestion-limits), "overrides" can be used to increase this (there is caution against going to large, however). we have added parameter as below.
The issue with the querier pod seems to be due to there being a gRPC message size limit between TempoStack components. As suggested here (#1097), I think we need to change settings both in the tempo.yaml and tempo-query-frontend.yaml to increase these to at least the max_bytes_per_trace size.
we are trying to change the existing server stanza settings with below values however it is getting overwritten,
Can anyone help us here and suggest how to and where to change these settings.
Thanks in advance !!
The text was updated successfully, but these errors were encountered: