-
Notifications
You must be signed in to change notification settings - Fork 486
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Metrics explosion since v0.35 for tracing collector component #5155
Comments
Could be related with #4764. Is there a way to disable this metrics in static config? |
Hi @ese 👋 Thank you for your report. In static mode we don't expect such a cardinality exposion, because the I just tried running the main branch of the Agent with a config like this: Agent config
Indeed, I do not see high cardinality metrics. My metrics look like this - with no
Would you mind telling us the precise version which you are using please? Also, it'd help if you could share a few example metrics. It may also help to try disabling some receivers by removing them from the config. I have only tested this with "otlp" - it is possible that other receivers don't honour the feature gate, but I feel like this is unlikely. |
I tested |
FYI I ran into this myself with an agent running |
I can still reproduce this with traces:
configs:
- name: default
receivers: # enable the receivers that you need
otlp:
protocols:
grpc:
http:
remote_write:
- endpoint: tempo-eu-west-0.grafana.net:443
basic_auth:
username: 274733
password_file: /var/lib/grafana-agent/rw.key and I am instrumenting a ruby/rails application with OTEL's SDK and sending the traces directly to the OTLP endpoint of the agent |
I just tested Agent v0.37.2 and v0.37.3, and while I do see the issue in v0.37.2, I do not see it in v0.37.3. This must be due to the OTel upgrade done in v0.37.3. On v0.37.2 I do see high cardinality labels such as
But on v0.37.3, those high cardinality labels are gone:
I'm not sure exactly what changed to fix this issue, but if anyone is affected I'd suggest upgrading to a version of the Agent which is v0.37.3 or above. |
I can confirm it is fixed in |
What's wrong?
Since v0.35 tracing component is generating a massive amount of metrics due to cardinality explosion. Before this version grafana-agent /metrics endpoint generated response of around 0.7Mb and after this release is generating over 250Mb responses. I tested upon the last release v0.36.1 getting the same results.
These are the metrics causing the issue (especially becouse the label
net_sock_peer_port
). they were not being generated before using the same config:traces_http_server_duration_bucket
traces_http_server_duration_sum
traces_http_server_duration_count
traces_http_server_response_content_length_total
Steps to reproduce
Upgrade agent to v0.35 or later and collect traces
System information
No response
Software version
v0.36.1
Configuration
The text was updated successfully, but these errors were encountered: