-
Notifications
You must be signed in to change notification settings - Fork 4.9k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
bulk_max_body_size support? - working around 413 Request Entity Too Large #3688
Comments
Note for those currently facing this issue. I am currently using the following. An AWS t2.small * 4 cluster.
|
We were also need a config like this due to AWS limitations (10MB). One suggestion was that
So this property is for max bytes of a bulk request before compression and not related to request body size. FYI, AWS ES doesn't support compression. Since our's is an Java application, if an error occurs continuously within a short period of time (worst case), we will reach the 10MB limit (on default Happy to help with the PR / pre dev discussions if this configuration can be included in the Roadmap. |
Please, I would really benefit from this. Typically messages are quite small (~5kb) but occassionally very large (best part of 1MB). We're using JSON mode and it's only really efficient with big batch sizes (>2000) most of the time. But then a few large messages screws everything up . I have to manually adjust down, then up again, on production. Also - |
This issue has been automatically marked as stale because it has not had recent activity. It will be closed if no further activity occurs. Thank you for your contributions. |
This issue doesn't have a |
(Follow up from : https://discuss.elastic.co/t/bulk-max-body-size-support/76611)
An alternative limit to bulk_max_size.
That functions based on the payload size instead.
This limitation is required due to managed Elasticsearch deployments (such as AWS)
having upload size limits of 10 MB for entry level.
See: http://docs.aws.amazon.com/elasticsearch-service/latest/developerguide/aes-limits.html
Because a single multiline message caps at 10MB by default, with 50 for batch processing.
The current "limit" is about 500MB with some overheads.
Currently when this happen a 413 error is perpetually repeated. Specifically the following.
As there is no way to increase the limit on AWS side, nor on the filebeat side,
other then to greatly decrease the max log size, and bulk_max_size.
This greatly limit the configuration options in such situations.
The text was updated successfully, but these errors were encountered: