Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Big parameterised queries do not work with analytics #10

Open
MarieSaphira opened this issue Dec 7, 2020 · 2 comments
Open

Big parameterised queries do not work with analytics #10

MarieSaphira opened this issue Dec 7, 2020 · 2 comments

Comments

@MarieSaphira
Copy link

Describe the bug

When using the parameterised queries the message size is limited. So inserting

{"parameterTypes":["uuid","datetime","datetime","string","int","polygon"],
"query":"insert MeteologixWarnings { @id: ??id, time_start: ??mtime_start, time_end: ??mtime_end, warningType: ??mwarningType, severity: ??mseverity, area: ??marea } ",
"parameterNames":["id","mtime_start","mtime_end","mwarningType","mseverity","marea"],
"boundRows":[["5f2e3221-a0ae-4189-b06d-096b53520ec6","2020-12-19T00:00:00Z","2020-12-19T23:59:59Z","gust10m","0","POLYGON ((3.417 55.170308641975005, 3.51875 55.170308641975005, 3.51875 55.30487654321, 3.417 55.30487654321, 3.417 55.170308641975005))"],
...
]}

with more than ~3000 boundRows throws the following error in the API:

java.util.concurrent.ExecutionException: org.apache.kafka.common.errors.RecordTooLargeException: 
The message is 2584882 bytes when serialized which is larger than the maximum request size you have configured with the max.request.size configuration.

Workaround

Either split queries only using ~3000 boundRows or use api/noAnalytics/update endpoint.
Maybe the API also should have a look @Stratidakos

@MarieSaphira
Copy link
Author

This is still an issue, you find a big query here.

@zolotas4
Copy link
Contributor

Hi @MarieSaphira, there is no plan to have this issue fixed at the moment. It has to do with very large batch inserts which are done mostly to accommodate fast data ingestion. We believe that as soon as it supports some thousands of queries then it should be acceptable.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants