Skip to content

Commit

Permalink
Apply suggestions from code review
Browse files Browse the repository at this point in the history
Co-authored-by: Nina Hingerl <[email protected]>
  • Loading branch information
a-thaler and NHingerl authored Sep 26, 2023
1 parent 4cad4cc commit 58ff617
Show file tree
Hide file tree
Showing 3 changed files with 6 additions and 6 deletions.
4 changes: 2 additions & 2 deletions docs/user/02-logs.md
Original file line number Diff line number Diff line change
Expand Up @@ -439,8 +439,8 @@ To avoid and detect these scenarios, you must monitor the instances by collectin
The relevant metrics are:
| Name | Threshold | Description |
|---|---|---|
| telemetry_fsbuffer_usage_bytes | (bytes/1000000000) * 100 > 90 | The metric indicates the current size of the persistent log buffer in bytes running on each instance. If the size reaches 1GB, logs will start getting dropped at that instance. At 90% buffer size an alert should get raised. |
| fluentbit_output_dropped_records_total| total[5m] > 0 | The metric indicates that the instance is actively dropping logs. That typically happens when a log message got rejected with a un-retryable status code like a 400. Any occurence of such drop should be alerted. |
| telemetry_fsbuffer_usage_bytes | (bytes/1000000000) * 100 > 90 | The metric indicates the current size (in bytes) of the persistent log buffer running on each instance. If the size reaches 1GB, logs are dropped at that instance. At 90% buffer size, an alert should be raised. |
| fluentbit_output_dropped_records_total| total[5m] > 0 | The metric indicates that the instance is actively dropping logs. That typically happens when a log message was rejected with a un-retryable status code like a 400. If logs are dropped, an alert should be raised. |


## Limitations
Expand Down
4 changes: 2 additions & 2 deletions docs/user/03-traces.md
Original file line number Diff line number Diff line change
Expand Up @@ -420,9 +420,9 @@ To avoid and detect these scenarios, you must monitor the instances by collectin
The relevant metrics are:
| Name | Threshold | Description |
|---|---|---|
| otelcol_exporter_enqueue_failed_spans | total[5m] > 0 | Indicates that new or retried items could not be added to the exporter buffer anymore as the buffer is exhausted. That usually happens when the configured backend cannot handle the load on time and is causing backpressure. |
| otelcol_exporter_enqueue_failed_spans | total[5m] > 0 | Indicates that new or retried items could not be added to the exporter buffer because the buffer is exhausted. Typically, that happens when the configured backend cannot handle the load on time and is causing back pressure. |
| otelcol_exporter_send_failed_spans | total[5m] > 0 | Indicates that items are refused in an non-retryable way like a 400 status |
| otelcol_processor_refused_spans | total[5m] > 0 | Indicates that items cannot be received anymore as a processor refuses them. That usually happens when memory of the collector is exhausted as too much data is arriving, then a throttling will start. |
| otelcol_processor_refused_spans | total[5m] > 0 | Indicates that items cannot be received anymore because a processor refuses them. Typically, that happens when memory of the collector is exhausted because too much data arrived and throttling started. |

## Limitations

Expand Down
4 changes: 2 additions & 2 deletions docs/user/04-metrics.md
Original file line number Diff line number Diff line change
Expand Up @@ -391,9 +391,9 @@ To avoid and detect these scenarios, you must monitor the instances by collectin
The relevant metrics are:
| Name | Threshold | Description |
|---|---|---|
| otelcol_exporter_enqueue_failed_metric_points | total[5m] > 0 | Indicates that new or retried items could not be added to the exporter buffer anymore as the buffer is exhausted. That usually happens when the configured backend cannot handle the load on time and is causing backpressure. |
| otelcol_exporter_enqueue_failed_metric_points | total[5m] > 0 | Indicates that new or retried items could not be added to the exporter buffer because the buffer is exhausted. Typically, that happens when the configured backend cannot handle the load on time and is causing back pressure. |
| otelcol_exporter_send_failed_metric_points | total[5m] > 0 | Indicates that items are refused in an non-retryable way like a 400 status |
| otelcol_processor_refused_metric_points | total[5m] > 0 | Indicates that items cannot be received anymore as a processor refuses them. That usually happens when memory of the collector is exhausted as too much data is arriving, then a throttling will start. |
| otelcol_processor_refused_metric_points | total[5m] > 0 | Indicates that items cannot be received because a processor refuses them. That usually happens when memory of the collector is exhausted because too much data arrived and throttling started.. |

## Limitations

Expand Down

0 comments on commit 58ff617

Please sign in to comment.