Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Issue is regarding the duplicate metrics and hostname discrepancies after setting up OpenTelemetry (Otel) along with Datadog agent #35448

Closed
zendesk-shweta opened this issue Sep 27, 2024 · 2 comments
Labels
bug Something isn't working needs triage New item requiring triage Stale

Comments

@zendesk-shweta
Copy link

Component(s)

No response

What happened?

Description

we have duplicated hosts being reported:

One Host i-xxxxxxxx which had for aliases ip-xxx-xx-xxx-xx.usw2.abc.com
One Host ip-xxx-xx-xxx-xx.usw2.abc.internal-podx which had no aliases

Steps to Reproduce

Expected Result

Actual Result

Collector version

VERSION=0.101.0

Environment information

Environment

OS: (e.g., "Ubuntu 20.04")
Compiler(if manually compiled): (e.g., "go 14.2")

OpenTelemetry Collector configuration

extensions:
  zpages:
    endpoint: :
  health_check:
    endpoint: 

receivers:
  datadog:
    endpoint: 
    read_timeout:
  otlp:
    protocols:
      grpc:
        endpoint: 
      http:
        endpoint: 
  prometheus:
   config:
     scrape_configs:
     - job_name: ' abc '
       scrape_interval: 60s
       static_configs:
       - targets: ['']
       metric_relabel_configs:
       - source_labels: [__name__]
         regex: '   '
         action: drop

processors:
  batch:
  memory_limiter:
    check_interval: 
    limit_percentage: 
    spike_limit_percentage: 

exporters:
  nop:
  logging:
    loglevel: info
  datadog:
    api:
      site: "datadoghq.com"
      key: ${env:DD_API_KEY}
    tls:
        insecure_skip_verify: true
    sending_queue:
      enabled: true
      queue_size: 20000
      num_consumers: 100
    timeout: 1s
    retry_on_failure:
      enabled: false
      initial_interval: 5s
      max_interval: 30s
      max_elapsed_time: 5m
    host_metadata:
      enabled: false
  otlphttp:
    endpoint: 
    tls:
      insecure: true

service:
  telemetry:
    metrics:
      level: basic
      address: ":8888"
  pipelines:
    traces:
      receivers: [otlp]
      processors: [memory_limiter, batch]
      exporters: [otlphttp]
    traces/datadog:
      receivers: [datadog]
      processors: [memory_limiter, batch]
      exporters: [otlphttp] # TBD: isolate it on aggr.
    metrics:
      receivers: [prometheus]
      processors: [memory_limiter, batch]
      exporters: [datadog] #TBD: make it nop for verification.

  extensions: [zpages, health_check]

Log output

No response

Additional context

No response

Copy link
Contributor

This issue has been inactive for 60 days. It will be closed in 60 days if there is no activity. To ping code owners by adding a component label, see Adding Labels via Comments, or if you are unsure of which component this issue relates to, please ping @open-telemetry/collector-contrib-triagers. If this issue is still relevant, please ping the code owners or leave a comment explaining why it is still relevant. Otherwise, please close it.

@atoulme
Copy link
Contributor

atoulme commented Dec 7, 2024

Please consider opening a support request with Datadog to receive help regarding your configuration. It is very difficult to diagnose the issue without more information.

@atoulme atoulme closed this as not planned Won't fix, can't repro, duplicate, stale Dec 7, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
bug Something isn't working needs triage New item requiring triage Stale
Projects
None yet
Development

No branches or pull requests

2 participants