-
Notifications
You must be signed in to change notification settings - Fork 915
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Fix floating point data generation in benchmarks #10372
Conversation
Codecov Report
@@ Coverage Diff @@
## branch-22.04 #10372 +/- ##
================================================
+ Coverage 10.42% 10.58% +0.15%
================================================
Files 119 125 +6
Lines 20603 21058 +455
================================================
+ Hits 2148 2228 +80
- Misses 18455 18830 +375
Continue to review full report at Codecov.
|
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
This looks good. Thank you for including the comment.
Tests appear to be failing with an unrelated error? #10150?
|
…bug-data_gen-limits
…nto bug-data_gen-limits
I think it's a CI weirdness caused by API changes in #10150. Merged latest, that usually helps in such cases. |
@gpucibot merge |
numeric_limits::lowest
andnumeric_limits::max
are used as bounds for numeric type generation. However, for normal generators, bounds are shifted to[0, upper_bound - lower_bound]
, and the random value is shifted back bylower_bound
.with
lowest
andmax
,upper_bound - lower_bound
is out of range for floats and generated values arenan
andinf
.This PR halves the ranges so that
upper_bound - lower_bound
is still within the type range.Expected to affect benchmarks that use floating point columns (e.g. Parquet reader benchmarks).