Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Configure batch size on KafkaConnect BigQuery sink connector #396

Open
icyBlue27 opened this issue Feb 20, 2024 · 2 comments
Open

Configure batch size on KafkaConnect BigQuery sink connector #396

icyBlue27 opened this issue Feb 20, 2024 · 2 comments

Comments

@icyBlue27
Copy link

Using the KafkaConnect BigQuery sink connector, from time to time, we observe the following error since the streaming API has a limit on the size of the batch it can write to BigQuery com.google.cloud.bigquery.BigQueryException: Request size is too big: 12705398 limitation: 12582912. Would it be possible to have the batch size configurable on the sink connector to stay under certain limits?

Kafka offers different parameters to configure on the producer at the moment (https://kafka.apache.org/documentation/#producerconfigs).

@b-goyal
Copy link
Member

b-goyal commented Feb 21, 2024

@icyBlue27 , could you try setting consumer.override.max.poll.records to a lower number ? Default value for this config is 500.

@icyBlue27
Copy link
Author

thanks @b-goyal . If I understand correctly max.poll.records controls the number of records which are pulled. This is not exactly what we would like to have, which is a limit on the size in bytes of the batch. True that less records would probably be smaller in size, but this is not an optimal solution.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants