-
Notifications
You must be signed in to change notification settings - Fork 240
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[BUG] test_iceberg_parquet_read_round_trip FAILED "TypeError: object of type 'NoneType' has no len()" #6718
Comments
I can reproduce locally on 22.12 but wasn't able to on 22.10 |
gpu output seems to be missing a column entry, gpu has this as None cpu has it as: _c19=Row(child0=[110, -70, 109, -17, 97, 0, -66], child1=108, child2=-4.712884395247157e+25, child3=Decimal('3306845829.53')), on the first row of output |
cudf from 10/4 works so something must have changed there |
fails with spark-rapids-jni jar from 10/6 so likely something on 4th or 5th that went in |
Note this is happening when data type is: And it only happens when you select enough data to make the coalescing kick in. It also only happens with iceberg, reading the raw parquet files the coalescing reader works fine. With iceberg a ton of these columns come back with null instead of the actual values. |
I finally got a parquet file that would reproduce this and sent to cudf folks |
going to xfail the test temporarily |
Fixes NVIDIA/spark-rapids#6718 There was a bug introduced recently #11752 where an insufficient check for whether an input column contained repetition information could cause incorrect results for column hierarchies with structs at the root. Authors: - https://github.com/nvdbaranec Approvers: - Jim Brennan (https://github.com/jbrennan333) - Nghia Truong (https://github.com/ttnghia) - Mike Wilson (https://github.com/hyperbolic2346) URL: #11910
Describe the bug
iceberg_test.py::test_iceberg_parquet_read_round_trip[COALESCING-[Byte, Short, Integer, ...
TypeError: object of type 'NoneType' has no len()
The text was updated successfully, but these errors were encountered: