From 11964aee01d9e43aeddad585440bb8a79611e45e Mon Sep 17 00:00:00 2001 From: Alessandro Bellina Date: Mon, 14 Oct 2024 13:14:26 -0500 Subject: [PATCH] Remove an unused config shuffle.spillThreads (#11595) * Remove an unused config shuffle.spillThreads Signed-off-by: Alessandro Bellina * update configs.md --------- Signed-off-by: Alessandro Bellina --- docs/configs.md | 1 - .../src/main/scala/com/nvidia/spark/rapids/RapidsConf.scala | 6 ------ 2 files changed, 7 deletions(-) diff --git a/docs/configs.md b/docs/configs.md index 52338a3a1c3..2d757a40779 100644 --- a/docs/configs.md +++ b/docs/configs.md @@ -45,7 +45,6 @@ Name | Description | Default Value | Applicable at spark.rapids.sql.multiThreadedRead.numThreads|The maximum number of threads on each executor to use for reading small files in parallel. This can not be changed at runtime after the executor has started. Used with COALESCING and MULTITHREADED readers, see spark.rapids.sql.format.parquet.reader.type, spark.rapids.sql.format.orc.reader.type, or spark.rapids.sql.format.avro.reader.type for a discussion of reader types. If it is not set explicitly and spark.executor.cores is set, it will be tried to assign value of `max(MULTITHREAD_READ_NUM_THREADS_DEFAULT, spark.executor.cores)`, where MULTITHREAD_READ_NUM_THREADS_DEFAULT = 20.|20|Startup spark.rapids.sql.reader.batchSizeBytes|Soft limit on the maximum number of bytes the reader reads per batch. The readers will read chunks of data until this limit is met or exceeded. Note that the reader may estimate the number of bytes that will be used on the GPU in some cases based on the schema and number of rows in each batch.|2147483647|Runtime spark.rapids.sql.reader.batchSizeRows|Soft limit on the maximum number of rows the reader will read per batch. The orc and parquet readers will read row groups until this limit is met or exceeded. The limit is respected by the csv reader.|2147483647|Runtime -spark.rapids.sql.shuffle.spillThreads|Number of threads used to spill shuffle data to disk in the background.|6|Runtime spark.rapids.sql.udfCompiler.enabled|When set to true, Scala UDFs will be considered for compilation as Catalyst expressions|false|Runtime For more advanced configs, please refer to the [RAPIDS Accelerator for Apache Spark Advanced Configuration](./additional-functionality/advanced_configs.md) page. diff --git a/sql-plugin/src/main/scala/com/nvidia/spark/rapids/RapidsConf.scala b/sql-plugin/src/main/scala/com/nvidia/spark/rapids/RapidsConf.scala index 50dc457268c..c2f867f8a80 100644 --- a/sql-plugin/src/main/scala/com/nvidia/spark/rapids/RapidsConf.scala +++ b/sql-plugin/src/main/scala/com/nvidia/spark/rapids/RapidsConf.scala @@ -551,12 +551,6 @@ val GPU_COREDUMP_PIPE_PATTERN = conf("spark.rapids.gpu.coreDump.pipePattern") .integerConf .createWithDefault(2) - val SHUFFLE_SPILL_THREADS = conf("spark.rapids.sql.shuffle.spillThreads") - .doc("Number of threads used to spill shuffle data to disk in the background.") - .commonlyUsed() - .integerConf - .createWithDefault(6) - val GPU_BATCH_SIZE_BYTES = conf("spark.rapids.sql.batchSizeBytes") .doc("Set the target number of bytes for a GPU batch. Splits sizes for input data " + "is covered by separate configs. The maximum setting is 2 GB to avoid exceeding the " +