Internal telemetry of our otel-operator gateway-collector v9.3.1 emits a lot of errors reporting of sending errors to mOTLP endpoints.
{
"code.function.name": [
"go.opentelemetry.io/collector/exporter/exporterhelper/internal.NewQueueSender.func1"
],
"error": [
"not retryable error: Permanent error: rpc error: code = Internal desc = grpc: failed to read decompressed data: message after decompression larger than max (4194304), please check your batching config: https://ela.st/edot-batching-config"
],
"attributes.code.stacktrace": [
"go.opentelemetry.io/collector/exporter/exporterhelper/internal.NewQueueSender.func1\n\tgo.opentelemetry.io/collector/exporter/exporterhelper@v0.144.0/internal/queue_sender.go:50\ngo.opentelemetry.io/collector/exporter/exporterhelper/internal/queuebatch.(*partitionBatcher).flush.func1\n\tgo.opentelemetry.io/collector/exporter/exporterhelper@v0.144.0/internal/queuebatch/partition_batcher.go:246\ngo.opentelemetry.io/collector/exporter/exporterhelper/internal/queuebatch.(*workerPool).execute.func1\n\tgo.opentelemetry.io/collector/exporter/exporterhelper@v0.144.0/internal/queuebatch/partition_batcher.go:268"
]
}
Exporter configs are the ones of the elastic provided values.yaml file. Examples:
otlp/ingest_logs:
endpoint: ${env:ELASTIC_OTLP_ENDPOINT}
headers:
Authorization: ApiKey ${env:ELASTIC_API_KEY}
sending_queue:
enabled: true
sizer: bytes
queue_size: 50000000 # 50MB uncompressed
block_on_overflow: true
batch:
flush_timeout: 1s
min_size: 1_000_000 # 1MB uncompressed
max_size: 4_000_000 # 4MB uncompressed
timeout: 15s
otlp/ingest_metrics_traces:
endpoint: ${env:ELASTIC_OTLP_ENDPOINT}
headers:
Authorization: ApiKey ${env:ELASTIC_API_KEY}
sending_queue:
enabled: true
sizer: bytes
queue_size: 50000000 # 50MB uncompressed
block_on_overflow: true
batch:
flush_timeout: 1s
min_size: 1_000_000 # 1MB uncompressed
max_size: 4_000_000 # 4MB uncompressed
timeout: 15s
The queue settings are the same as recommended here.
It seems that either these settings or the settings on mOTEL are wrong and need to be fixed.
Internal telemetry of our otel-operator gateway-collector v9.3.1 emits a lot of errors reporting of sending errors to mOTLP endpoints.
{ "code.function.name": [ "go.opentelemetry.io/collector/exporter/exporterhelper/internal.NewQueueSender.func1" ], "error": [ "not retryable error: Permanent error: rpc error: code = Internal desc = grpc: failed to read decompressed data: message after decompression larger than max (4194304), please check your batching config: https://ela.st/edot-batching-config" ], "attributes.code.stacktrace": [ "go.opentelemetry.io/collector/exporter/exporterhelper/internal.NewQueueSender.func1\n\tgo.opentelemetry.io/collector/exporter/exporterhelper@v0.144.0/internal/queue_sender.go:50\ngo.opentelemetry.io/collector/exporter/exporterhelper/internal/queuebatch.(*partitionBatcher).flush.func1\n\tgo.opentelemetry.io/collector/exporter/exporterhelper@v0.144.0/internal/queuebatch/partition_batcher.go:246\ngo.opentelemetry.io/collector/exporter/exporterhelper/internal/queuebatch.(*workerPool).execute.func1\n\tgo.opentelemetry.io/collector/exporter/exporterhelper@v0.144.0/internal/queuebatch/partition_batcher.go:268" ] }Exporter configs are the ones of the elastic provided values.yaml file. Examples:
The queue settings are the same as recommended here.
It seems that either these settings or the settings on mOTEL are wrong and need to be fixed.