Menu
Open source

Troubleshoot metrics-generator

If you are concerned with data quality issues in the metrics-generator, we’d first recommend:

  • Reviewing your telemetry pipeline to determine the number of dropped spans. We are only looking for major issues here.
  • Reviewing the service graph documentation to understand how they are built.

If everything seems ok from these two perspectives, consider the following topics to help resolve general issues with all metrics and span metrics specifically.

All metrics

Dropped spans in the distributor

The distributor has a queue of outgoing spans to the metrics-generators. If that queue is full then the distributor will drop spans before they reach the generator. Use the following metric to determine if that is happening:

sum(rate(tempo_distributor_queue_pushes_failures_total{}[1m]))

Failed pushes to the generator

For any number of reasons, the distributor can fail a push to the generators. Use the following metric to determine if that is happening:

sum(rate(tempo_distributor_metrics_generator_pushes_failures_total{}[1m]))

Discarded spans in the generator

Spans are rejected from being considered by the metrics-generator by a configurable slack time as well as due to user configurable filters. You can see the number of spans rejected by reason using this metric:

sum(rate(tempo_metrics_generator_spans_discarded_total{}[1m])) by (reason)

If a lot of spans are dropped in the metrics-generator due to your filters, you will need to adjust them. If spans are dropped due to the ingestion slack time, consider adjusting this setting:

metrics_generator:
  metrics_ingestion_time_range_slack: 30s

If spans are regularly exceeding this value you may want to consider reviewing your tracing pipeline to see if you have excessive buffering. Note that increasing this value allows the generator to consume more spans, but does reduce the accuracy of metrics because spans farther away from “now” are included.

Max active series

The generator protects itself and your remote-write target by having a maximum number of series the generator produces. Use the sum below to determine if series are being dropped due to this limit:

sum(rate(tempo_metrics_generator_registry_series_limited_total{}[1m]))

Use the following setting to update the limit:

yaml
overrides:
  defaults:
    metrics_generator:
      max_active_series: 0

Note that this value is per metrics generator. The actual max series remote written will be <# of metrics generators> * <metrics_generator.max_active_series>.

Remote write failures

For any number of reasons, the generator may fail a write to the remote write target. Use the following metrics to determine if that is happening:

sum(rate(prometheus_remote_storage_samples_failed_total{}[1m]))
sum(rate(prometheus_remote_storage_samples_dropped_total{}[1m]))
sum(rate(prometheus_remote_storage_exemplars_failed_total{}[1m]))
sum(rate(prometheus_remote_storage_exemplars_dropped_total{}[1m]))

Service graph metrics

Service graphs have additional configuration which can impact the quality of the output metrics.

Expired edges

The following metrics can be used to determine how many edges are failing to find a match.

Rate of edges that have expired without a match:

sum(rate(tempo_metrics_generator_processor_service_graphs_expired_edges{}[1m]))

Rate of all edges:

sum(rate(tempo_metrics_generator_processor_service_graphs_edges{}[1m]))

If you are seeing a large number of edges expire without a match, consider adjusting the wait setting. This controls how long the metrics generator waits to find a match before it gives up.

metrics_generator:
  processor:
    service_graphs:
      wait: 10s

Service graph max items

The service graph processor has a maximum number of edges it will track at once to limit the total amount of memory the processor uses. To determine if edges are being dropped due to this limit, check:

sum(rate(tempo_metrics_generator_processor_service_graphs_dropped_spans{}[1m]))

Use max_items to adjust the maximum amount of edges tracked:

metrics_generator:
  processor:
    service_graphs:
      max_items: 10000