You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
While using a cluster on 3.3.1 when there are millions of keys in etcd, we have seen very long durations (5-10 seconds) on the index compaction stat measured here. This coincided with increases in request latency and client side timeouts. We have further observed that almost all of the time spent in this function has been around the tree index compaction here. Looking into this function, we observed the the following comment. Since this function locks for the duration of compacting the B-Tree, this affects overall throughput performance. We wrote a benchmark in order to validate the run time in isolation that is provided below with our results. Unfortunately these results do not line up with the O(10ms) comment. We ran these benchmarks on a 16 logical CPU machine with Xeon E5-2670 v2 @ 2.50GHz. Is this an expected bottleneck?
While using a cluster on 3.3.1 when there are millions of keys in etcd, we have seen very long durations (5-10 seconds) on the index compaction stat measured here
the benchmark does not explain why the pause is at second level. maybe there is something else going on?
The benchmark shows that compaction scales worse than linearly. Adding a benchmark for 5000000 shows on this same machine a result in excess of 2 seconds. Extrapolating this to 10 million index entries it would seem is sufficient to explain that much pause.
While using a cluster on
3.3.1
when there are millions of keys in etcd, we have seen very long durations (5-10 seconds) on the index compaction stat measured here. This coincided with increases in request latency and client side timeouts. We have further observed that almost all of the time spent in this function has been around the tree index compaction here. Looking into this function, we observed the the following comment. Since this function locks for the duration of compacting the B-Tree, this affects overall throughput performance. We wrote a benchmark in order to validate the run time in isolation that is provided below with our results. Unfortunately these results do not line up with theO(10ms)
comment. We ran these benchmarks on a 16 logical CPU machine withXeon E5-2670 v2 @ 2.50GHz
. Is this an expected bottleneck?The text was updated successfully, but these errors were encountered: