-
Notifications
You must be signed in to change notification settings - Fork 1.5k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
scrape/scrape.go:1313 Scrape commit failed "error": "data refused due to high memory usage" #8217
Comments
I'm having the same problem with prometheus receiver performing scrapes since the container is started and have tried different collector versions and settings: resources:
requests:
cpu: 500m
memory: 4096Mi
limits:
memory: 4096Mi
processors:
memory_limiter:
check_interval: 1s
limit_mib: 4000
spike_limit_mib: 800
batch:
send_batch_size: 1000
timeout: 1s
send_batch_max_size: 1500
extensions:
memory_ballast:
size_mib: 2000 |
I have settings:
I should try memory_ballast as well ... |
Does anyone have normal scrape results from pods under load? |
Added definitions for batch
Let's see how it will be |
After a few days of work (5), history repeats itself. |
Hey @pilot513 , did you managed to fix this issue some how? Thanks |
In my case, I noticed that the number of metrics was constantly growing. I began to study this issue, and discovered that one application was generating a constant increase in unique metrics. It shouldn't be this way. I pointed this out to the developers, and they fixed it because their code for expose metrics was incorrect. As soon as I reinstalled the application, the problem went away. |
I'm seeing the same thing, memory usage keeps going up until the receiver starts falling. At that point I begin to see export failures and the export queue goes up as well. We're sending around 35K/s data points across 350 scrape targets.
|
Just tested on v0.97, same failure pattern. I did noticed this error message as well:
|
Met similar error on v0.97 |
I am facing similar issue.. Scrape continouly fails with the below error - github.com/prometheus/[email protected]/scrape/scrape.go:1306 This also leads high memory usage at otel-collector.. Do we have any work-around for this ? |
Same issue version 0.100. any work around? |
Describe the bug
ota collector can't scrape pod metrics
Steps to reproduce
Configure prometheus exporter with prometheus endpoint
What did you expect to see?
Scrape metrics from pods and receive it to 'prometheusremotewrite'
What did you see instead?
error scrape/scrape.go:1313 Scrape commit failed {"kind": "receiver", "name": "prometheus", "data_type": "metrics", "scrape_pool": "otel_kubernetes_podscraper", "target": "http://ip:port/metrics", "error": "data refused due to high memory usage"}
What version did you use?
Version: 0.82
What config did you use?
Config:
...
prometheus:
endpoint: 0.0.0.0:port
metric_expiration: 120m
resource_to_telemetry_conversion:
enabled: true
send_timestamps: true
prometheusremotewrite:
endpoint: http://hostname/prometheus/api/v1/write
extensions:
health_check: {}
memory_ballast: {}
processors:
batch: {}
memory_limiter:
check_interval: 3s
limit_mib: 6553
spike_limit_mib: 2048
...
Environment
k8s Pod (from helm chart) with Limit 8G 2cores
The text was updated successfully, but these errors were encountered: