Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Removed references to spark.shuffle.service.enabled = true #145

Merged
merged 1 commit into from
Oct 24, 2024
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
8 changes: 0 additions & 8 deletions ons-spark/spark-overview/example-spark-sessions.md
Original file line number Diff line number Diff line change
Expand Up @@ -112,7 +112,6 @@ spark = (
.config("spark.dynamicAllocation.enabled", "true")
.config("spark.dynamicAllocation.maxExecutors", 3)
.config("spark.sql.shuffle.partitions", 12)
.config("spark.shuffle.service.enabled", "true")
.config("spark.ui.showConsoleProgress", "false")
.enableHiveSupport()
.getOrCreate()
Expand All @@ -127,7 +126,6 @@ small_config$spark.executor.cores <- 1
small_config$spark.dynamicAllocation.enabled <- "true"
small_config$spark.dynamicAllocation.maxExecutors <- 3
small_config$spark.sql.shuffle.partitions <- 12
small_config$spark.shuffle.service.enabled <- "true"

sc <- sparklyr::spark_connect(
master = "yarn-client",
Expand Down Expand Up @@ -164,7 +162,6 @@ spark = (
.config("spark.dynamicAllocation.enabled", "true")
.config("spark.dynamicAllocation.maxExecutors", 3)
.config("spark.sql.shuffle.partitions", 18)
.config("spark.shuffle.service.enabled", "true")
.config("spark.ui.showConsoleProgress", "false")
.enableHiveSupport()
.getOrCreate()
Expand All @@ -179,7 +176,6 @@ medium_config$spark.executor.cores <- 3
medium_config$spark.dynamicAllocation.enabled <- "true"
medium_config$spark.dynamicAllocation.maxExecutors <- 3
medium_config$spark.sql.shuffle.partitions <- 18
medium_config$spark.shuffle.service.enabled <- "true"

sc <- sparklyr::spark_connect(
master = "yarn-client",
Expand Down Expand Up @@ -218,7 +214,6 @@ spark = (
.config("spark.dynamicAllocation.enabled", "true")
.config("spark.dynamicAllocation.maxExecutors", 5)
.config("spark.sql.shuffle.partitions", 200)
.config("spark.shuffle.service.enabled", "true")
.config("spark.ui.showConsoleProgress", "false")
.enableHiveSupport()
.getOrCreate()
Expand All @@ -234,7 +229,6 @@ large_config$spark.executor.cores <- 5
large_config$spark.dynamicAllocation.enabled <- "true"
large_config$spark.dynamicAllocation.maxExecutors <- 5
large_config$spark.sql.shuffle.partitions <- 200
large_config$spark.shuffle.service.enabled <- "true"

sc <- sparklyr::spark_connect(
master = "yarn-client",
Expand Down Expand Up @@ -275,7 +269,6 @@ spark = (
.config("spark.dynamicAllocation.enabled", "true")
.config("spark.dynamicAllocation.maxExecutors", 12)
.config("spark.sql.shuffle.partitions", 240)
.config("spark.shuffle.service.enabled", "true")
.config("spark.ui.showConsoleProgress", "false")
.enableHiveSupport()
.getOrCreate()
Expand All @@ -291,7 +284,6 @@ xl_config$spark.executor.cores <- 5
xl_config$spark.dynamicAllocation.enabled <- "true"
xl_config$spark.dynamicAllocation.maxExecutors <- 12
xl_config$spark.sql.shuffle.partitions <- 240
xl_config$spark.shuffle.service.enabled <- "true"

sc <- sparklyr::spark_connect(
master = "yarn-client",
Expand Down
1 change: 0 additions & 1 deletion ons-spark/spark-overview/spark-defaults.md
Original file line number Diff line number Diff line change
Expand Up @@ -48,7 +48,6 @@ spark.executor.cores 1
spark.dynamicAllocation.enabled true
spark.dynamicAllocation.maxExecutors 3
spark.sql.shuffle.partitions 12
spark.shuffle.service.enabled true
spark.ui.showConsoleProgress false
```

Expand Down
5 changes: 2 additions & 3 deletions ons-spark/spark-overview/spark-session-guidance.md
Original file line number Diff line number Diff line change
Expand Up @@ -137,8 +137,8 @@ Also note when specifying this config, Spark will reserve the maxExecutors for y
stop your session- even if you don't need them. See the [reserving](#reserving) section of the calculations below for more details.

**Shuffle Service**
`.config("spark.shuffle.service.enabled", "true")`
This property helps with processes like switching executors off safely using dynamic allocation by making use of shuffle files.
`.config("spark.shuffle.service.enabled", "false")`
This property helps with processes like switching executors off safely using dynamic allocation by making use of shuffle files. The default value of this property is false.

**Hive Support**
`.enableHiveSupport()`
Expand All @@ -164,7 +164,6 @@ spark = (
.config("spark.executor.cores", 5)
.config("spark.dynamicAllocation.enabled", "true")
.config("spark.dynamicAllocation.maxExecutors", 5)
.config("spark.shuffle.service.enabled", "true")
.config("spark.ui.showConsoleProgress", "false")
.enableHiveSupport()
.getOrCreate()
Expand Down