Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[BUG] YARN IT test test_optimized_hive_ctas_basic failures #7922

Closed
tgravescs opened this issue Mar 22, 2023 · 2 comments
Closed

[BUG] YARN IT test test_optimized_hive_ctas_basic failures #7922

tgravescs opened this issue Mar 22, 2023 · 2 comments
Labels
bug Something isn't working

Comments

@tgravescs
Copy link
Collaborator

Describe the bug
on the yanr cluster integration tests failed:

test_optimized_hive_ctas_basic[nativeorc-[Map(String(not_null),String), Map(Boolean(not_null),Boolean), Map(Byte(not_null),Byte), Map(Short(not_null),Short), Map(Integer(not_null),Integer), Map(Long(not_null),Long), Map(Float(not_null),Float), Map(Double(not_null),Double), Map(Timestamp(not_null),Timestamp), Map(Date(not_null),Date), Map(Decimal(not_null)(15,1),Decimal(15,1)), Map(Decimal(not_null)(36,5),Decimal(36,5))]]
 test_optimized_hive_ctas_basic[hiveorc-[Map(String(not_null),String), Map(Boolean(not_null),Boolean), Map(Byte(not_null),Byte), Map(Short(not_null),Short), Map(Integer(not_null),Integer), Map(Long(not_null),Long), Map(Float(not_null),Float), Map(Double(not_null),Double), Map(Timestamp(not_null),Timestamp), Map(Date(not_null),Date), Map(Decimal(not_null)(15,1),Decimal(15,1)), Map(Decimal(not_null)(36,5),Decimal(36,5))]] _�[0m

Exceptions:


[2023-03-22T16:52:34.020Z] �[1m�[31mE                   Caused by: java.lang.NullPointerException�[0m
[2023-03-22T16:52:34.020Z] �[1m�[31mE                   	at org.apache.hadoop.hive.ql.io.orc.RecordReaderImpl.close(RecordReaderImpl.java:151)�[0m
[2023-03-22T16:52:34.020Z] �[1m�[31mE                   	at org.apache.orc.impl.RecordReaderImpl.<init>(RecordReaderImpl.java:289)�[0m
[2023-03-22T16:52:34.020Z] �[1m�[31mE                   	at org.apache.hadoop.hive.ql.io.orc.RecordReaderImpl.<init>(RecordReaderImpl.java:63)�[0m
[2023-03-22T16:52:34.020Z] �[1m�[31mE                   	at org.apache.hadoop.hive.ql.io.orc.ReaderImpl.rowsOptions(ReaderImpl.java:80)�[0m
[2023-03-22T16:52:34.020Z] �[1m�[31mE                   	at org.apache.hadoop.hive.ql.io.orc.OrcInputFormat.createReaderFromFile(OrcInputFormat.java:324)�[0m
[2023-03-22T16:52:34.020Z] �[1m�[31mE                   	at org.apache.hadoop.hive.ql.io.orc.SparkOrcNewRecordReader.<init>(SparkOrcNewRecordReader.java:50)�[0m
[2023-03-22T16:52:34.020Z] �[1m�[31mE                   	at org.apache.****.sql.hive.orc.OrcFileFormat.$anonfun$buildReader$2(OrcFileFormat.scala:176)�[0m
[2023-03-22T16:52:34.020Z] �[1m�[31mE                   	at org.apache.****.sql.execution.datasources.FileFormat$$anon$1.apply(FileFormat.scala:147)�[0m
[2023-03-22T16:52:34.020Z] �[1m�[31mE                   	at org.apache.****.sql.execution.datasources.FileFormat$$anon$1.apply(FileFormat.scala:132)�[0m
[2023-03-22T16:52:34.020Z] �[1m�[31mE                   	at org.apache.****.sql.execution.datasources.FileScanRDD$$anon$1.org$apache$****$sql$execution$datasources$FileScanRDD$$anon$$readCurrentFile(FileScanRDD.scala:116)�[0m
[2023-03-22T16:52:34.020Z] �[1m�[31mE                   	at org.apache.****.sql.execution.datasources.FileScanRDD$$anon$1.nextIterator(FileScanRDD.scala:169)�[0m
[2023-03-22T16:52:34.020Z] �[1m�[31mE                   	at org.apache.****.sql.execution.datasources.FileScanRDD$$anon$1.hasNext(FileScanRDD.scala:93)�[0m
[2023-03-22T16:52:34.020Z] �[1m�[31mE                   	at scala.collection.Iterator$$anon$10.hasNext(Iterator.scala:458)�[0m
[2023-03-22T16:52:34.020Z] �[1m�[31mE                   	at org.apache.****.sql.execution.SparkPlan.$anonfun$getByteArrayRdd$1(SparkPlan.scala:345)�[0m
[2023-03-22T16:52:34.021Z] �[1m�[31mE                   	at org.apache.****.rdd.RDD.$anonfun$mapPartitionsInternal$2(RDD.scala:898)�[0m
[2023-03-22T16:52:34.021Z] �[1m�[31mE                   	at org.apache.****.rdd.RDD.$anonfun$mapPartitionsInternal$2$adapted(RDD.scala:898)�[0m
[2023-03-22T16:52:34.021Z] �[1m�[31mE                   	at org.apache.****.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:52)�[0m
[2023-03-22T16:52:34.021Z] �[1m�[31mE                   	at org.apache.****.rdd.RDD.computeOrReadCheckpoint(RDD.scala:373)�[0m
[2023-03-22T16:52:34.021Z] �[1m�[31mE                   	at org.apache.****.rdd.RDD.iterator(RDD.scala:337)�[0m
[2023-03-22T16:52:34.021Z] �[1m�[31mE                   	at org.apache.****.scheduler.ResultTask.runTask(ResultTask.scala:90)�[0m
[2023-03-22T16:52:34.021Z] �[1m�[31mE                   	at org.apache.****.scheduler.Task.run(Task.scala:131)�[0m
[2023-03-22T16:52:34.021Z] �[1m�[31mE                   	at org.apache.****.executor.Executor$TaskRunner.$anonfun$run$3(Executor.scala:497)�[0m
[2023-03-22T16:52:34.021Z] �[1m�[31mE                   	at org.apache.****.util.Utils$.tryWithSafeFinally(Utils.scala:1439)�[0m
[2023-03-22T16:52:34.021Z] �[1m�[31mE                   	at org.apache.****.executor.Executor$TaskRunner.run(Executor.scala:500)�[0m
[2023-03-22T16:52:34.021Z] �[1m�[31mE                   	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)�[0m
[2023-03-22T16:52:34.017Z] �[1m�[31mE                   Caused by: java.lang.NegativeArraySizeException�[0m
[2023-03-22T16:52:34.017Z] �[1m�[31mE                   	at org.apache.orc.impl.TreeReaderFactory$StringDictionaryTreeReader.readDictionaryLengthStream(TreeReaderFactory.java:1774)�[0m
[2023-03-22T16:52:34.017Z] �[1m�[31mE                   	at org.apache.orc.impl.TreeReaderFactory$StringDictionaryTreeReader.startStripe(TreeReaderFactory.java:1758)�[0m
[2023-03-22T16:52:34.017Z] �[1m�[31mE                   	at org.apache.orc.impl.TreeReaderFactory$StringTreeReader.startStripe(TreeReaderFactory.java:1500)�[0m
[2023-03-22T16:52:34.017Z] �[1m�[31mE                   	at org.apache.orc.impl.TreeReaderFactory$MapTreeReader.startStripe(TreeReaderFactory.java:2364)�[0m
[2023-03-22T16:52:34.017Z] �[1m�[31mE                   	at org.apache.orc.impl.TreeReaderFactory$StructTreeReader.startStripe(TreeReaderFactory.java:2090)�[0m
[2023-03-22T16:52:34.018Z] �[1m�[31mE                   	at org.apache.orc.impl.RecordReaderImpl.readStripe(RecordReaderImpl.java:1107)�[0m
[2023-03-22T16:52:34.018Z] �[1m�[31mE                   	at org.apache.orc.impl.RecordReaderImpl.advanceStripe(RecordReaderImpl.java:1256)�[0m
[2023-03-22T16:52:34.018Z] �[1m�[31mE                   	at org.apache.orc.impl.RecordReaderImpl.advanceToNextRow(RecordReaderImpl.java:1291)�[0m
[2023-03-22T16:52:34.018Z] �[1m�[31mE                   	at org.apache.orc.impl.RecordReaderImpl.<init>(RecordReaderImpl.java:286)�[0m
[2023-03-22T16:52:34.018Z] �[1m�[31mE                   	at org.apache.orc.impl.ReaderImpl.rows(ReaderImpl.java:669)�[0m
[2023-03-22T16:52:34.018Z] �[1m�[31mE                   	at org.apache.orc.mapreduce.OrcMapreduceRecordReader.<init>(OrcMapreduceRecordReader.java:59)�[0m
[2023-03-22T16:52:34.018Z] �[1m�[31mE                   	at org.apache.orc.mapreduce.OrcInputFormat.createRecordReader(OrcInputFormat.java:72)�[0m
[2023-03-22T16:52:34.018Z] �[1m�[31mE                   	at org.apache.****.sql.execution.datasources.orc.OrcFileFormat.$anonfun$buildReaderWithPartitionValues$1(OrcFileFormat.scala:227)�[0m
[2023-03-22T16:52:34.018Z] �[1m�[31mE                   	at org.apache.****.sql.execution.datasources.FileScanRDD$$anon$1.org$apache$****$sql$execution$datasources$FileScanRDD$$anon$$readCurrentFile(FileScanRDD.scala:116)�[0m
[2023-03-22T16:52:34.018Z] �[1m�[31mE                   	at org.apache.****.sql.execution.datasources.FileScanRDD$$anon$1.nextIterator(FileScanRDD.scala:169)�[0m
[2023-03-22T16:52:34.018Z] �[1m�[31mE                   	at org.apache.****.sql.execution.datasources.FileScanRDD$$anon$1.hasNext(FileScanRDD.scala:93)�[0m
[2023-03-22T16:52:34.018Z] �[1m�[31mE                   	at scala.collection.Iterator$$anon$10.hasNext(Iterator.scala:458)�[0m
[2023-03-22T16:52:34.018Z] �[1m�[31mE                   	at org.apache.****.sql.execution.SparkPlan.$anonfun$getByteArrayRdd$1(SparkPlan.scala:345)�[0m
[2023-03-22T16:52:34.018Z] �[1m�[31mE                   	at org.apache.****.rdd.RDD.$anonfun$mapPartitionsInternal$2(RDD.scala:898)�[0m
[2023-03-22T16:52:34.018Z] �[1m�[31mE                   	at org.apache.****.rdd.RDD.$anonfun$mapPartitionsInternal$2$adapted(RDD.scala:898)�[0m
[2023-03-22T16:52:34.018Z] �[1m�[31mE                   	at org.apache.****.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:52)�[0m
[2023-03-22T16:52:34.018Z] �[1m�[31mE                   	at org.apache.****.rdd.RDD.computeOrReadCheckpoint(RDD.scala:373)�[0m
[2023-03-22T16:52:34.018Z] �[1m�[31mE                   	at org.apache.****.rdd.RDD.iterator(RDD.scala:337)�[0m
[2023-03-22T16:52:34.018Z] �[1m�[31mE                   	at org.apache.****.scheduler.ResultTask.runTask(ResultTask.scala:90)�[0m
[2023-03-22T16:52:34.018Z] �[1m�[31mE                   	at org.apache.****.scheduler.Task.run(Task.scala:131)�[0m
[2023-03-22T16:52:34.018Z] �[1m�[31mE                   	at org.apache.****.executor.Executor$TaskRunner.$anonfun$run$3(Executor.scala:497)�[0m
[2023-03-22T16:52:34.018Z] �[1m�[31mE                   	at org.apache.****.util.Utils$.tryWithSafeFinally(Utils.scala:1439)�[0m
[2023-03-22T16:52:34.018Z] �[1m�[31mE                   	at org.apache.****.executor.Executor$TaskRunner.run(Executor.scala:500)�[0m
[2023-03-22T16:52:34.018Z] �[1m�[31mE                   	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)�[0m
[2023-03-22T16:52:34.018Z] �[1m�[31mE                   	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)�[0m
[2023-03-22T16:52:34.018Z] �[1m�[31mE                   	... 1 more�[0m
@tgravescs tgravescs added bug Something isn't working ? - Needs Triage Need team to review and classify labels Mar 22, 2023
@jlowe
Copy link
Member

jlowe commented Mar 23, 2023

This may be caused by the use-after-free introduced in rapidsai/cudf#12949 and later fixed by rapidsai/cudf#12984. We have yet to pickup the latter in a nightly build.

@mattahrens mattahrens removed the ? - Needs Triage Need team to review and classify label Mar 28, 2023
@mattahrens
Copy link
Collaborator

Builds are succeeding now, likely fixed by rapidsai/cudf#12984

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
bug Something isn't working
Projects
None yet
Development

No branches or pull requests

3 participants