-
Notifications
You must be signed in to change notification settings - Fork 1.6k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
kafka source and aws s3 sink not work #18658
Comments
Hi @waney316 ! Can you clarify what you mean by "doesn't work". Are you seeing events consumed from Kafka (you could use It should be unrelated, but given you are using Kafka I do think adding a disk buffer to the Also, could you also add the debug logs here? |
tks @jszwedko ,It seems like you're encountering an issue where, under buffer.type = "disk", the buffer size remains unchanged. Additionally, Vector's top command shows that there are no incoming events from the Kafka source (source=input-kafka-backup), and the sink (Sinks=bak-to-cos) is not updating. I also noticed an error in Vector's logs: consumption error: PollExceeded (Local: Maximum application poll interval (max.poll.interval.ms) exceeded. What is the issue here? The following is a snippet of my configuration. [sources.input-kafka-backup]
type = "kafka"
bootstrap_servers = "${KAFKA}"
group_id = "xxxxxx"
topics = [ "^xxxxxx.*" ]
decoding.codec="json"
auto_offset_reset = "latest"
[sinks.bak-to-cos]
type = "aws_s3"
healthcheck.enabled = false
inputs = [ "input-kafka-backup" ]
auth.access_key_id = ""
auth.secret_access_key = ""
endpoint = "https://xxxxx.com"
key_prefix = "{{ topic }}/%Y/%m/%d/"
bucket = "xxxxxx"
region = "xxxxx"
compression = "gzip"
encoding.codec="json"
framing.method = "newline_delimited"
filename_time_format = "%Y%m%d%H%M%S%3f_archive"
batch.max_bytes = 209715200
batch.timeout_secs = 900
buffer.type = "disk"
buffer.max_size = 734003200
buffer.when_full = "block"
request.concurrency= "adaptive" |
Hi @waney316 , It seems like Vector is failing to fetch events from Kafka. I would check:
|
Closing due to lack of response. |
A note for the community
Problem
My vector version is 0.32.1. When I use kafka as the source and AWS S3 as the sink, it doesn't seem to work because the number of kafka consumer groups is 0. I suspect that this is due to the use of buffer.type=disk in the sink。
Configuration
No response
Version
0.32.1
Debug Output
No response
Example Data
No response
Additional Context
No response
References
No response
The text was updated successfully, but these errors were encountered: