-
Notifications
You must be signed in to change notification settings - Fork 25k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Stats about bulk sizes #47345
Comments
Pinging @elastic/es-core-features |
As a note here, it might make sense to track the average bulk sizes per shard as well as the average bulk request size, since a large bulk request may be chopped down into much smaller shard level bulk operation on an index with high numbers of shards. This makes more sense to me than just tracking at the shard level since most clients are not partitioning by shard already. |
* Add Bulk stats track the bulk sizes per shard and the time spent on the bulk shard request (elastic#50536)(elastic#47345)
This has been open for quite a while, and we haven't made much progress on this due to focus in other areas. For now I'm going to close this as something we aren't planning on implementing. We can re-open it later if needed. |
The size of bulk requests in bytes in one of the main factors of indexing performance. Yet we don't have any stats about it, and users don't always know what is the actual size of the bulk requests that get actually sent to Elasticsearch, which is something that is automated in many shippers. For instance Logstash allows to configure the size of batches as a number of documents, but the
pipeline.batch.delay
may cause incomplete batches to be sent, so it would be good to know what is the size of bulk requests that got sent to Elasticsearch in practice.The size of bulk requests per shard is probably the most relevant metric, but we could work with the per-index size as well if that proves easier to expose.
The text was updated successfully, but these errors were encountered: