-
Notifications
You must be signed in to change notification settings - Fork 530
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
metrics-generator
idle and no traces are generated using Tempo 2.5.0
#3855
Comments
same here. but im using official grafana tempo-distributed helm chart
|
I'm not sure if this is the limitaion in distributed mode. When I tried it with Docker Compose, it worked because the metrics generator and ingester read and write data to the same local path. However, in a Kubernetes environment under distributed mode, all components run as separate containers with separate volumes. I'm not sure if the metrics generator and ingester share the same path. someone from Grafana team can enlighten us on this |
In our set-up,
But doesn't seem to be mounted into Another thing that we've spotted when running the tests was the following behaviour:
At some point during the tests (that in our case generate traffic across 30+ microservices) it basically stopped generating metrics. After I detected that in the morning next day, and restarted the
Nothing is being printed about even attempting to connect to Prometheus instance, which in our case, is completetly idle, not restarting, working as expected, always on the same URL. |
Hey all, can you review these steps: https://grafana.com/docs/tempo/latest/troubleshooting/metrics-generator/ This will help us identify where things are being dropped.
Generators and ingesters do not need shared access to block storage. They are fine to run independently. |
Hello @joe-elliott, thank you for reaching out. As a part of the troubleshooting, our Operations Team has reverted from Grafana Tempo I will follow the matching troubleshooting guide from here: https://grafana.com/docs/tempo/latest/troubleshooting/metrics-generator/ Testing of metrics from similar issue: #2514
Metrics like: Troubleshooting guide
It seems that It used to work when we had a reinstall of a Helm chart and after some time and tests, it stopped working, reporting I've attached the example body of the metrics-generator-example-metrics-endpoint-response-body.txt Prometheus reports it as scraping correctly: |
Also, we've restarted again the Grafana Tempo - clean deployed the Helm chart version What we can see now is that the
When we have a situation where
Where, for a working
On a idle
|
It stopped again:
As we used to have some metrics for a short period of time, I'd like to get the metrics from troubleshooting guide again: Troubleshooting guide
|
By checking out code locally,
The It has a chance to exit before
here:
Why it's not being called - I am not able to tell :( Any help needed here. Thank you. |
I would definitely not recommend upgrading from 2.1.x to 2.5.x. We aggressively rotate through block formats and it's quite possible blocks created by 2.1 are not supported by 2.5. also, if you're doing fresh installs. can you just install 2.5? But I'm not sure if that is causing your issue or not. You're saying that on a fresh install of 2.1.1 it worked for awhile and then quit? Those graphs suggest the distributors just stopped sending data to the generators. Can you check:
I believe it's not being called b/c if you don't send data to the generator it will never create an instance. it creates a per tenant instance on demand. the missing log line suggests it's just not receiving any data which is consistent with your other graphs. |
In my tempo distributor metrics I found this metric
metrics generator client is not configured it seems. |
@nc-gcz have you tried enabling metrics generator by setting values in global_overrides:
defaults:
metrics_generator:
processors:
- service-graphs
- span-metrics https://github.com/grafana/helm-charts/tree/main/charts/tempo-distributed#activate-metrics-generator |
I faced the same issue. |
Have the same issue with mostly fresh installation from tempo-distributed 2.6.0 - tempo was running for a day or two without generator. Here are relevant configmap parts:
Per-tenant override is empty. I've tried putting values from overrides into per-tenant config, but that changed nothing. Removed debug logs from comment for less visual spam. Edit: Reinstalled in fully fresh namespace and bucket and generator works there. |
I just ran into this issue after upgrading to 2.6.0 (but I believe it's been broken for me since 2.5.0), and this is how I resolved it:
|
I tried this |
turns out need to explicitly set this:
|
To enable metrics generator, the chart says to pass the names of the processors to overrides.metrics_generator_processors value like [service-graphs, span-metrics the following block must be added: global_overrides:
per_tenant_override_config: /runtime-config/overrides.yaml
overrides:
legacyOverrides:
metrics_generator_processors:
- service-graphs
- span-metrics this does not work (which is presented in the chart): overrides:
metrics_generator_processors:
- service-graphs
- span-metrics the following works in the single binary version: overrides:
defaults:
metrics_generator:
processors: [service-graphs, span-metrics, local-blocks] I opened this PR so see if I missed something or this is correct :) |
Also the configuration for single binary works here as well global_overrides:
per_tenant_override_config: /runtime-config/overrides.yaml
defaults:
metrics_generator:
processors: [service-graphs, span-metrics, local-blocks] |
In my case with tempo-distributed (helm chart 2.6.0), after 2 days debugging, I've found a solution. I put this block in values.yaml and metrics-generator started send metrics
|
Hi there! I noticed that we had a bunch of issues open for the overrides settings for the tempo-distributed Helm chart. We have two updates that may address the issue that you're having:
This doc may help if you're using the Helm chart. |
Never in a million years would I have thought about using |
Describe the bug
Good morning. We are running Grafana Tempo 2.5.0 using Bitnami Grafana Tempo distributed Helm chart in version: 3.6.3. The following components are deployed: Compactor, Distributed, Memcached, Metrics Generator, Querier, Query Frontend, Vulture and OpenTelemetry Collector in version: 0.83.0.
We have an issue after updating Grafana Tempo 2.1.1 to 2.5.0 with Metrics Generator. The issues with incompatible
vParquet
version has been mitigated according to documentation: https://grafana.com/docs/tempo/latest/setup/upgrade/#upgrade-to-tempo-25.The issue we are experiencing is that
metrics-generator
seems to be idle after start-up:This is the log from
metrics-generator
that has been started around 60 minutes before.The relevant configuration is as follows:
Yet, no service maps or span metrics are visible in Grafana. These metrics were correctly persisted in Prometheus via remote-write on Tempo 2.1.1 but broke after upgrading to 2.5.0. Also, internal metrics defined in the documentation page: https://grafana.com/docs/tempo/latest/troubleshooting/metrics-generator/ (first 3, 4) are reporting empty data in Prometheus - seems like they are not reported at all.
I think I've followed the documentation to enable the
processor
inmetrics_generator
yet, it's idling. Before, I was expecting something like:But I cannot see anything after joining cluster, nothing about connecting to Prometheus, nothing about not being able to connect to Prometheus remote-write, nothing.
Interestingly enough, running Grafana Tempo
./examples/docker-compose/distributed
I was able to see the metrics out of the box.To Reproduce:
Steps to reproduce the behavior:
Expected behavior:
Metrics Generator is not idle and metrics are visible in Prometheus metrics.
Environment:
v11.1.0 (5b85c4c2fc)
The text was updated successfully, but these errors were encountered: