diff --git a/cpp/src/io/parquet/decode_fixed.cu b/cpp/src/io/parquet/decode_fixed.cu
index 1ca2c404071..8d18c4c3e20 100644
--- a/cpp/src/io/parquet/decode_fixed.cu
+++ b/cpp/src/io/parquet/decode_fixed.cu
@@ -870,7 +870,7 @@ inline __device__ void bool_plain_decode(page_state_s* s, state_buf* sb, int t,
   if (t == 0) { s->dict_pos = pos; }
 }
 
-template <int decode_block_size_t, typename stream_type>
+template <int rolling_buf_size, typename stream_type>
 __device__ int skip_decode(stream_type& parquet_stream, int num_to_skip, int t)
 {
   // it could be that (e.g.) we skip 5000 but starting at row 4000 we have a run of length 2000:
@@ -879,7 +879,7 @@ __device__ int skip_decode(stream_type& parquet_stream, int num_to_skip, int t)
   int num_skipped = parquet_stream.skip_decode(t, num_to_skip);
   while (num_skipped < num_to_skip) {
     // TODO: Instead of decoding, skip within the run to the appropriate location
-    auto const to_decode = min(2 * decode_block_size_t, num_to_skip - num_skipped);
+    auto const to_decode = min(rolling_buf_size, num_to_skip - num_skipped);
     num_skipped += parquet_stream.decode_next(t, to_decode);
     __syncthreads();
   }
@@ -1036,11 +1036,11 @@ CUDF_KERNEL void __launch_bounds__(decode_block_size_t, 8)
     auto const skipped_leaf_values = s->page.skipped_leaf_values;
     if (skipped_leaf_values > 0) {
       if (should_process_nulls) {
-        skip_decode<decode_block_size_t>(def_decoder, skipped_leaf_values, t);
+        skip_decode<rolling_buf_size>(def_decoder, skipped_leaf_values, t);
       }
-      processed_count = skip_decode<decode_block_size_t>(rep_decoder, skipped_leaf_values, t);
+      processed_count = skip_decode<rolling_buf_size>(rep_decoder, skipped_leaf_values, t);
       if constexpr (has_dict_t) {
-        skip_decode<decode_block_size_t>(dict_stream, skipped_leaf_values, t);
+        skip_decode<rolling_buf_size>(dict_stream, skipped_leaf_values, t);
       }
     }
   }
diff --git a/cpp/src/io/parquet/parquet_gpu.hpp b/cpp/src/io/parquet/parquet_gpu.hpp
index 1c1185d8afd..8124be52552 100644
--- a/cpp/src/io/parquet/parquet_gpu.hpp
+++ b/cpp/src/io/parquet/parquet_gpu.hpp
@@ -994,7 +994,6 @@ void DecodePageDataFixedDict(cudf::detail::hostdevice_span<PageInfo> pages,
  * @param[in] level_type_size Size in bytes of the type for level decoding
  * @param[in] has_nesting Whether or not the data contains nested (but not list) data.
  * @param[in] is_list Whether or not the data contains list data.
- * @param[in] is_list Whether or not the data contains list data.
  * @param[out] error_code Error code for kernel failures
  * @param[in] stream CUDA stream to use
  */
@@ -1005,7 +1004,6 @@ void DecodeSplitPageFixedWidthData(cudf::detail::hostdevice_span<PageInfo> pages
                                    int level_type_size,
                                    bool has_nesting,
                                    bool is_list,
-                                   bool is_list,
                                    kernel_error::pointer error_code,
                                    rmm::cuda_stream_view stream);
 
diff --git a/cpp/src/io/parquet/rle_stream.cuh b/cpp/src/io/parquet/rle_stream.cuh
index 76a2b050b9e..3c49de0c997 100644
--- a/cpp/src/io/parquet/rle_stream.cuh
+++ b/cpp/src/io/parquet/rle_stream.cuh
@@ -216,7 +216,7 @@ struct rle_stream {
     decode_index = -1;  // signals the first iteration. Nothing to decode.
   }
 
-  __device__ inline int get_rle_run_info(rle_run<level_t>& run)
+  __device__ inline int get_rle_run_info(rle_run& run)
   {
     run.start     = cur;
     run.level_run = get_vlq32(run.start, end);
@@ -383,7 +383,7 @@ struct rle_stream {
     // started basically we're setting up the rle_stream vars necessary to start fill_run_batch for
     // the first time
     while (cur < end) {
-      rle_run<level_t> run;
+      rle_run run;
       int run_bytes = get_rle_run_info(run);
 
       if ((output_pos + run.size) > target_count) {