-
Notifications
You must be signed in to change notification settings - Fork 24.9k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[ML] adding new KS test pipeline aggregation #73334
Conversation
Pinging @elastic/ml-core (Team:ML) |
@elasticmachine update branch |
…search into feature/ml-ks-test-agg
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I don't feel comfortable reviewing the math-heavy part in BucketCountKSTestAggregator.java
. In case you don't have another reviewer, it would be good to go through this code together so I can understand it in more detail.
experimental::[] | ||
|
||
A sibling pipeline aggregation which executes a two sample Kolmogorov–Smirnov test | ||
(referred to as a "K-S test" from now own) against a provided distribution and |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
(referred to as a "K-S test" from now own) against a provided distribution and | |
(referred to as a "K-S test" from now on) against a provided distribution and |
"range": { <2> | ||
"field": "latency", | ||
"ranges": [ | ||
{ "to": 0.0 }, |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Could 0.0
be replaced with 0
so that it is an integer, like numbers in other ranges?
<1> The term buckets containing a range aggregation and the bucket correlation aggregation. Both are utilized to calculate | ||
the correlation of the term values with the latency. | ||
<2> The range aggregation on the latency field. The ranges were created referencing the percentiles of the latency field. | ||
<3> The bucket count K-S test aggregation that determines if the count samples are from the same distribution as the uniform |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Could this sentence be simplified to:
The bucket count K-S test aggregation that determines if the count samples come from the uniform distribution
?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
technically, no. I will think of a better way of phrasing it.
The two-sampled K-S test (which we are doing here) is verifying that two samples come from the same distribution. In this particular case, we are testing if one sample (a uniform fraction sample) and another (these _counts) both come from the same distribution.
|
||
package org.elasticsearch.xpack.ml.aggs; | ||
|
||
public final class DoubleArray { |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Could you add a unit test for this class?
|
||
private DoubleArray() { } | ||
|
||
public static double[] cumulativeSum(double[] xs) { |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Could you add a method comment?
Especially a sentence that states that xs
is immutable.
sublistedPath, | ||
BucketHelpers.GapPolicy.INSERT_ZEROS | ||
); | ||
if (bucketValue != null && Double.isNaN(bucketValue) == false) { |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
What happens if this if
statement evaluates to false
and, in consequence, values.size()
will be less than the number of buckets?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
If I supplied a list of fractions
to the agg and that length of that list no longer matches the number of buckets how would that be resolved?
There should always be a bucket with a _count
value, if not is this exceptional and the code should throw?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
100%, I am throwing an execution exception here now. If we want to support something more nuanced later, we can change it.
throw invalidPathException(path); | ||
} | ||
|
||
private InvalidAggregationPathException invalidPathException(List<String> path) { |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
This method is only used once. Would it make sense to inline it?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
refactoring slightly as the inference agg result also uses this method (or one similar enough)
new double[] { 4, 6, 2, 3, 3, 2, 1, 1, 1, 1 } | ||
); | ||
|
||
final MlBucketsHelper.DoubleBucketValues UPPER_TAILED_VALUES = new MlBucketsHelper.DoubleBucketValues( |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
final MlBucketsHelper.DoubleBucketValues UPPER_TAILED_VALUES = new MlBucketsHelper.DoubleBucketValues( | |
private static final MlBucketsHelper.DoubleBucketValues UPPER_TAILED_VALUES = new MlBucketsHelper.DoubleBucketValues( |
new long[] { 10, 10, 10, 40, 40, 40, 40, 40, 40, 40 }, | ||
new double[] { 10, 10, 10, 40, 40, 40, 40, 40, 40, 40 } | ||
); | ||
final MlBucketsHelper.DoubleBucketValues UPPER_TAILED_VALUES_SPARSE = new MlBucketsHelper.DoubleBucketValues( |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
final MlBucketsHelper.DoubleBucketValues UPPER_TAILED_VALUES_SPARSE = new MlBucketsHelper.DoubleBucketValues( | |
private static final MlBucketsHelper.DoubleBucketValues UPPER_TAILED_VALUES_SPARSE = new MlBucketsHelper.DoubleBucketValues( |
allOf(hasKey(Alternative.GREATER.toString()), hasKey(Alternative.LESS.toString()), hasKey(Alternative.TWO_SIDED.toString())) | ||
); | ||
// Since these two distributions are the "same" (both uniform) | ||
// Assume that the p-value is greater than 0.9 |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
// Assume that the p-value is greater than 0.9 | |
// assume that the p-value is greater than 0.9 |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I'm not familiar with the KS test and haven't dived into the maths but everything I understood LGTM 👍
docs/reference/aggregations/pipeline/bucket-count-ks-test-aggregation.asciidoc
Show resolved
Hide resolved
|
||
<1> The term buckets containing a range aggregation and the bucket correlation aggregation. Both are utilized to calculate | ||
the correlation of the term values with the latency. | ||
<2> The range aggregation on the latency field. The ranges were created referencing the percentiles of the latency field. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
It would be nice to have a full example somewhere in docs of showing how the ranges were found
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Possibly, I think though that is a larger "how correlations + K-S tests work" type of docs.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Actually this isn't so bad as you explain that the latency ranges were calculated using a percentiles agg
sublistedPath, | ||
BucketHelpers.GapPolicy.INSERT_ZEROS | ||
); | ||
if (bucketValue != null && Double.isNaN(bucketValue) == false) { |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
If I supplied a list of fractions
to the agg and that length of that list no longer matches the number of buckets how would that be resolved?
There should always be a bucket with a _count
value, if not is this exceptional and the code should throw?
|
||
private InvalidAggregationPathException invalidPathException(List<String> path) { | ||
return new InvalidAggregationPathException( | ||
"unknown property " + path + " for " + InferencePipelineAggregationBuilder.NAME + " aggregation [" + getName() + "]" |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
"unknown property " + path + " for " + InferencePipelineAggregationBuilder.NAME + " aggregation [" + getName() + "]" | |
"unknown property " + path + " for " + BucketCountKSTestAggregationBuilder.NAME + " aggregation [" + getName() + "]" |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM
`sampling_method`:: | ||
(Optional, string) | ||
Indicates the sampling methodology when calculating the K-S test. Note, this is sampling | ||
of the returned values. This determines the cumulative distribution function (cdf) points |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
of the returned values. This determines the cumulative distribution function (cdf) points | |
of the returned values. This determines the cumulative distribution function (CDF) points |
Capitalised CDF
is used in the next sentence
|
||
<1> The term buckets containing a range aggregation and the bucket correlation aggregation. Both are utilized to calculate | ||
the correlation of the term values with the latency. | ||
<2> The range aggregation on the latency field. The ranges were created referencing the percentiles of the latency field. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Actually this isn't so bad as you explain that the latency ranges were calculated using a percentiles agg
if (alternative == null) { | ||
this.alternative = EnumSet.allOf(Alternative.class); | ||
} else { | ||
if (alternative.isEmpty()) { |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
The exact same check happens in validate()
but it throws a validation error rather than IllegalArgumentException
.
Should the gapPolicy
check below also be in validate()
(not the null part the bit about policy != INSERT_ZERORS). A quick survey of other agg builder classes does not show a strong convention either way.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I am gonna remove the check from validate()
. All the fields are final
and I already validate in the ctor.
@elasticmachine update branch |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I had some minor suggestions on documentation. Other than that, I think we need a different distribution for the sided test statistics. Let me get back to you with a suggestion for that.
A list of string values indicating which K-S test alternative to calculate. | ||
The valid values are: "greater", "less", "two_sided". This parameter is key for | ||
determining the K-S statistic used when calculating the K-S test. Default value is | ||
all possible alternative hypothesis. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
typo
all possible alternative hypothesis. | |
all possible alternative hypotheses. |
(referred to as a "K-S test" from now on) against a provided distribution and | ||
the distribution of documents counts in the configured sibling aggregation. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I think it is worth expanding this somewhat. For example, I think it is worth making clear that this is computing the K-S p-value between Y and E[Y | c] where c is the "event" selected by the sibling aggregation. The natural use case is when the sibling aggregation is a terms agg in which case it computes the K-S between Y and the restriction of Y to each term. This also means that one can't use any old values for fractions they have to match the actual proportion of the docs in each range bucket for Y.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I would propose something like the following:
A sibling pipeline aggregation which executes a two sample Kolmogorov–Smirnov test
(referred to as a "K-S test" from now on) against a provided distribution and the
distribution implied by the documents counts in the configured sibling aggregation.
Specifically, for some metric, assuming that the percentile intervals of the metric are
known beforehand or have been computed by an aggregation, then one would use range
aggregation for the sibling to compute the p-value of the distribution difference between
the metric and the restriction of that metric to a subset of the documents. A natural use
case is if the sibling aggregation range aggregation nested in a terms aggregation, in
which case one compares the overall distribution of metric to its restriction to each term.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
(I know you have some discussion of this later on, but I think this is crucial to understanding how this is meant to work for someone who knows what the K-S test actually does so worth including upfront.)
A list of doubles indicating the distribution of the samples with which to compare to the | ||
`buckets_path` results. The default is a uniform distribution of the same length as the | ||
`buckets_path` buckets. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
We have rather specific requirements on what fractions mean. To produce a meaningful result from this aggregation they should be related to some metric distribution which is then used to create the sibling aggregation. A natural choice is to use equal percentile range queries to construct the sibling aggregation in which case the default is correct. I think it is worth capturing something along these lines.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I'd propose something like:
A list of doubles indicating the distribution of the samples with which to compare to the
`buckets_path` results. In typical usage this is the overall proportion of documents in
each bucket, which is compared with the actual document proportions in each bucket
from the sibling aggregation counts. The default is to assume that overall documents
are uniformly distributed on these buckets, which they would be if one used equal
percentiles of a metric to define the bucket end points.
which was done utilizing the | ||
<<search-aggregations-metrics-percentile-aggregation,percentiles>> aggregation. | ||
|
||
This example is only using the 10s percentiles. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
This example is only using the 10s percentiles. | |
This example is only using the deciles of `latency`. |
"greater" : 4.58718174664754E-9, | ||
"two_sided" : 4.58718174664754E-9 |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I would expect greater
to be very nearly two_sided
/ 2 in this case.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Yeah, the two_sided
isn't accurate. I think Hodge's approximation only works in the sided case. We will probably have to do something else.
// And c(α)=√(−ln(α * 0.5) * 0.5) | ||
// Where Dₙ,ₘ is our statistic | ||
// Then solve for α | ||
// But I am not 100% sure which to choose in this case. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
This distribution is for the usual K-S test statistic not the sided versions. For small p-values the sided versions are very nearly half the two-sided one. I'll dig around for a similar asymptotic approximation for these alternatives.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Ah, yeah, hodges approximation is for the sided-statistic. So, we will have to have something else for two_sided.
The dictionary comes from Morfologik project. Morfologik uses data from | ||
Polish ispell/myspell dictionary hosted at http://www.sjp.pl/slownik/en/ and | ||
is licenced on the terms of (inter alia) LGPL and Creative Commons | ||
ShareAlike. The part-of-speech tags were added in Morfologik project and | ||
are not found in the data from sjp.pl. The tagset is similar to IPI PAN | ||
tagset. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I don't think this file contains the correct license for commons-math3. If I download https://mirrors.gethosted.online/apache/commons/math/binaries/commons-math3-3.6.1-bin.tar.gz then the license file in it is different to this one.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Ah, let me look into this.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM. Good stuff!
@elasticmachine update branch |
run elasticsearch-ci/part-2 |
@elasticmachine update branch |
This adds a new pipeline aggregation for calculating Kolmogorov–Smirnov test for a given sample and buckets path. For now, the buckets path resolution needs to be `_count`. But, this may be relaxed in the future. It accepts a parameter `fractions` that indicates the distribution of documents from some other pre-calculated sample. This particular version of the K-S test is Two-sample, meaning, it calculates if the `fractions` and the distribution of `_count` values in the buckets_path are taken from the same distribution. This in combination with the hypothesis alternatives (`less`, `greater`, `two_sided`) and sampling logic (`upper_tail`, `lower_tail`, `uniform`) allow for flexibility and usefulness when comparing two samples and determining the likelihood of them being from the same overall distribution. Usage: ``` POST correlate_latency/_search?size=0&filter_path=aggregations { "aggs": { "buckets": { "terms": { <1> "field": "version", "size": 2 }, "aggs": { "latency_ranges": { "range": { <2> "field": "latency", "ranges": [ { "to": 0.0 }, { "from": 0, "to": 105 }, { "from": 105, "to": 225 }, { "from": 225, "to": 445 }, { "from": 445, "to": 665 }, { "from": 665, "to": 885 }, { "from": 885, "to": 1115 }, { "from": 1115, "to": 1335 }, { "from": 1335, "to": 1555 }, { "from": 1555, "to": 1775 }, { "from": 1775 } ] } }, "ks_test": { <3> "bucket_count_ks_test": { "buckets_path": "latency_ranges>_count", "alternative": ["less", "greater", "two_sided"] } } } } } } ``` Co-authored-by: Elastic Machine <[email protected]>
This adds a new pipeline aggregation for calculating Kolmogorov–Smirnov test for a given sample and buckets path.
For now, the buckets path resolution needs to be
_count
. But, this may be relaxed in the future.It accepts a parameter
fractions
that indicates the distribution of documents from some other pre-calculated sample.This particular version of the K-S test is Two-sample, meaning, it calculates if the
fractions
and the distribution of_count
values in the buckets_path are taken from the same distribution.This in combination with the hypothesis alternatives (
less
,greater
,two_sided
) and sampling logic (upper_tail
,lower_tail
,uniform
) allow for flexibility and usefulness when comparing two samples and determining the likelihood of them being from the same overall distribution.Usage:
NOTE: one might notice that the
two_sided
alternative is always equal to eitherless
orgreater
. This is becausetwo_sided
is simply themax(less, greater)
.