From 1acc8e0744ff411ebb7ed63e993e2a278f018236 Mon Sep 17 00:00:00 2001 From: Jason Lowe Date: Tue, 23 Jan 2024 13:10:34 -0600 Subject: [PATCH] Mark hash reduction decimal overflow test as a permanent seed override (#10178) Signed-off-by: Jason Lowe --- integration_tests/src/main/python/hash_aggregate_test.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/integration_tests/src/main/python/hash_aggregate_test.py b/integration_tests/src/main/python/hash_aggregate_test.py index e5835a998b8..5d2bb7d658d 100644 --- a/integration_tests/src/main/python/hash_aggregate_test.py +++ b/integration_tests/src/main/python/hash_aggregate_test.py @@ -396,7 +396,7 @@ def test_hash_reduction_sum(data_gen, conf): @pytest.mark.parametrize('data_gen', numeric_gens + decimal_gens + [ DecimalGen(precision=38, scale=0), DecimalGen(precision=38, scale=-10)], ids=idfn) @pytest.mark.parametrize('conf', get_params(_confs, params_markers_for_confs), ids=idfn) -@datagen_overrides(seed=0, reason='https://github.com/NVIDIA/spark-rapids/issues/9779') +@datagen_overrides(seed=0, permanent=True, reason='https://github.com/NVIDIA/spark-rapids/issues/9779') def test_hash_reduction_sum_full_decimal(data_gen, conf): assert_gpu_and_cpu_are_equal_collect( lambda spark: unary_op_df(spark, data_gen, length=100).selectExpr("SUM(a)"),