Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

error [email protected]/accumulator.go:94 - failed to translate metric #26725

Closed
martinrw opened this issue Sep 18, 2023 · 3 comments
Closed
Labels
bug Something isn't working exporter/prometheus

Comments

@martinrw
Copy link

Component(s)

exporter/prometheus

What happened?

Description

We are seeing the error " failed to translate metric" in our logs for prometheus exporter for some of our metrics.
We are collecting metrics from the otel agent +libraries from a bunch of springboot, nodejs and python services, most of these are working perfectly fine but we see this error quite frequently in our logs.

I am unable to pinpoint these errors to a specific service or runtime or version of the agent etc even after enabling debug logs.

It is happening for a few different metrics such as:
http.server.duration
http.client.request.size
http.client.duration

Steps to Reproduce

Expected Result

Metrics are written to the /metrics endpoint to be scanned by prometheus

Actual Result

Error message in the logs, data point is dropped

Collector version

0.85

Environment information

No response

OpenTelemetry Collector configuration

config:
  exporters:
    prometheus:
      endpoint: "0.0.0.0:9464"
      resource_to_telemetry_conversion:
        enabled: true
      enable_open_metrics: true
      metric_expiration: 3m

  extensions:
    # The health_check extension is mandatory for this chart.
    # Without the health_check extension the collector will fail the readiness and liveliness probes.
    # The health_check extension can be modified, but should never be removed.
    health_check: {}
    zpages: {}
    pprof: {}
    memory_ballast:
      size_in_percentage: 30
  processors:
    memory_limiter:
      check_interval: 1s
      limit_percentage: 50
      spike_limit_percentage: 20
    batch:
      send_batch_size: 10000
      send_batch_max_size: 11000
      timeout: 2s
    #Delete unnecessary attributes from our metrics
    resource:
      attributes:
        - key: telemetry.sdk.name
          action: delete
        - key: telemetry.sdk.version
          action: delete
        - key: telemetry.sdk.language
          action: delete
        - key: telemetry.auto.version
          action: delete
        - key: container.id
          action: delete
        - key: process.command_args
          action: delete
        - key: process.command_line
          action: delete
        - key: process.command
          action: delete
        - key: process.executable.path
          action: delete
    transform:
      error_mode: ignore
      metric_statements:
        - context: metric
          statements:
            - set(description, "The duration of the inbound HTTP request") where name == "http.server.duration"
            - set(description, "The duration of the inbound HTTP request") where name == "http.client.duration"
            - set(description, "The current number of threads having NEW state") where name == "jvm.threads.states"
            - set(description, "The number of concurrent HTTP requests that are currently in-flight") where name == "http.server.active_requests" 
            - set(description, "") where name == "http.server.requests"
            - set(description, "") where name == "http.server.requests.max"
            - set(description, "Number of log events that were enabled by the effective log level") where name == "logback.events"
            - set(description, "") where name == "spring.data.repository.invocations.max"
            - set(description, "Duration of repository invocations") where name == "spring.data.repository.invocations"
            - set(description, "Time taken for the application to be ready to service requests") where name == "application.ready.time"
            - set(description, "Time taken (ms) to start the application") where name == "application.started.time"
            - set(description, "The size of HTTP request messages") where name == "http.client.request.size_bytes"
            - set(description, "The size of HTTP response messages") where name == "http.client.response.size_bytes"
    k8sattributes:
      extract:
        metadata:
          - k8s.namespace.name
          - k8s.deployment.name
          - k8s.statefulset.name
          - k8s.daemonset.name
          - k8s.cronjob.name
          - k8s.job.name
          - k8s.pod.name
          - k8s.namespace.name
          - k8s.node.name
  receivers:
    jaeger: null
    prometheus: null
    zipkin: null
    otlp:
      protocols:
        grpc:
          endpoint: 0.0.0.0:4317
        http:
          endpoint: 0.0.0.0:4318
  service:
    telemetry:
      #logs for the collector itself
      logs:
        level: info
    extensions:
      - health_check
      - memory_ballast
    pipelines:
      traces:
        exporters:
          - spanmetrics
        processors:
          - memory_limiter
          - batch
          - k8sattributes
        receivers:
          - otlp
      logs: null
      metrics:
        exporters:
          - prometheus
        processors:
          - memory_limiter
          - batch
          - resource
          - transform
          - k8sattributes
        receivers:
          - otlp
          - spanmetrics
  connectors:
    spanmetrics:
      namespace: span.metrics

Log output

error	[email protected]/accumulator.go:94	failed to translate metric	{"kind": "exporter", "data_type": "metrics", "name": "prometheus", "data_type": "\u0000", "metric_name": "http.client.request.size"}
github.com/open-telemetry/opentelemetry-collector-contrib/exporter/prometheusexporter.(*lastValueAccumulator).addMetric
	github.com/open-telemetry/opentelemetry-collector-contrib/exporter/[email protected]/accumulator.go:94
github.com/open-telemetry/opentelemetry-collector-contrib/exporter/prometheusexporter.(*lastValueAccumulator).Accumulate
	github.com/open-telemetry/opentelemetry-collector-contrib/exporter/[email protected]/accumulator.go:71
github.com/open-telemetry/opentelemetry-collector-contrib/exporter/prometheusexporter.(*collector).processMetrics
	github.com/open-telemetry/opentelemetry-collector-contrib/exporter/[email protected]/collector.go:92
github.com/open-telemetry/opentelemetry-collector-contrib/exporter/prometheusexporter.(*prometheusExporter).ConsumeMetrics
	github.com/open-telemetry/opentelemetry-collector-contrib/exporter/[email protected]/prometheus.go:85
go.opentelemetry.io/collector/exporter/exporterhelper.(*metricsRequest).Export
	go.opentelemetry.io/collector/[email protected]/exporterhelper/metrics.go:60
go.opentelemetry.io/collector/exporter/exporterhelper.(*timeoutSender).send
	go.opentelemetry.io/collector/[email protected]/exporterhelper/common.go:269
go.opentelemetry.io/collector/exporter/exporterhelper.(*baseRequestSender).send
	go.opentelemetry.io/collector/[email protected]/exporterhelper/common.go:54
go.opentelemetry.io/collector/exporter/exporterhelper.(*baseRequestSender).send
	go.opentelemetry.io/collector/[email protected]/exporterhelper/common.go:54
go.opentelemetry.io/collector/exporter/exporterhelper.(*metricsSenderWithObservability).send
	go.opentelemetry.io/collector/[email protected]/exporterhelper/metrics.go:179
go.opentelemetry.io/collector/exporter/exporterhelper.(*baseRequestSender).send
	go.opentelemetry.io/collector/[email protected]/exporterhelper/common.go:54
go.opentelemetry.io/collector/exporter/exporterhelper.(*baseExporter).send
	go.opentelemetry.io/collector/[email protected]/exporterhelper/common.go:216
go.opentelemetry.io/collector/exporter/exporterhelper.NewMetricsExporter.func1
	go.opentelemetry.io/collector/[email protected]/exporterhelper/metrics.go:100
go.opentelemetry.io/collector/consumer.ConsumeMetricsFunc.ConsumeMetrics
	go.opentelemetry.io/collector/[email protected]/metrics.go:25
github.com/open-telemetry/opentelemetry-collector-contrib/pkg/resourcetotelemetry.(*wrapperMetricsExporter).ConsumeMetrics
	github.com/open-telemetry/opentelemetry-collector-contrib/pkg/[email protected]/resource_to_telemetry.go:32
go.opentelemetry.io/collector/processor/processorhelper.NewMetricsProcessor.func1
	go.opentelemetry.io/collector/[email protected]/processorhelper/metrics.go:60
go.opentelemetry.io/collector/consumer.ConsumeMetricsFunc.ConsumeMetrics
	go.opentelemetry.io/collector/[email protected]/metrics.go:25
go.opentelemetry.io/collector/processor/processorhelper.NewMetricsProcessor.func1
	go.opentelemetry.io/collector/[email protected]/processorhelper/metrics.go:60
go.opentelemetry.io/collector/consumer.ConsumeMetricsFunc.ConsumeMetrics
	go.opentelemetry.io/collector/[email protected]/metrics.go:25
go.opentelemetry.io/collector/processor/processorhelper.NewMetricsProcessor.func1
	go.opentelemetry.io/collector/[email protected]/processorhelper/metrics.go:60
go.opentelemetry.io/collector/consumer.ConsumeMetricsFunc.ConsumeMetrics
	go.opentelemetry.io/collector/[email protected]/metrics.go:25
go.opentelemetry.io/collector/processor/batchprocessor.(*batchMetrics).export
	go.opentelemetry.io/collector/processor/[email protected]/batch_processor.go:442
go.opentelemetry.io/collector/processor/batchprocessor.(*shard).sendItems
	go.opentelemetry.io/collector/processor/[email protected]/batch_processor.go:256
go.opentelemetry.io/collector/processor/batchprocessor.(*shard).start
	go.opentelemetry.io/collector/processor/[email protected]/batch_processor.go:218

Additional context

No response

@martinrw martinrw added bug Something isn't working needs triage New item requiring triage labels Sep 18, 2023
@github-actions
Copy link
Contributor

Pinging code owners:

See Adding Labels via Comments if you do not have permissions to add labels yourself.

@vpmedia
Copy link

vpmedia commented Sep 23, 2023

I'm experiencing the same exception happening quite frequently.
Is "data_type": "\u0000" a valid data type? Just guessing..

@crobert-1
Copy link
Member

This is a duplicate of #13443, there's more discussion there.

@crobert-1 crobert-1 reopened this Sep 27, 2023
@crobert-1 crobert-1 closed this as not planned Won't fix, can't repro, duplicate, stale Sep 27, 2023
@crobert-1 crobert-1 removed the needs triage New item requiring triage label Oct 27, 2023
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
bug Something isn't working exporter/prometheus
Projects
None yet
Development

No branches or pull requests

3 participants