From 3d1d60b434edd68bbe25b3e51015927fe0635854 Mon Sep 17 00:00:00 2001 From: Yuming Wang Date: Wed, 7 Aug 2019 18:48:06 +0800 Subject: [PATCH] Make spark.sql.function.preferIntegralDivision internal --- .../src/main/scala/org/apache/spark/sql/internal/SQLConf.scala | 3 ++- 1 file changed, 2 insertions(+), 1 deletion(-) diff --git a/sql/catalyst/src/main/scala/org/apache/spark/sql/internal/SQLConf.scala b/sql/catalyst/src/main/scala/org/apache/spark/sql/internal/SQLConf.scala index ae65f29c3226..8c127cef349c 100644 --- a/sql/catalyst/src/main/scala/org/apache/spark/sql/internal/SQLConf.scala +++ b/sql/catalyst/src/main/scala/org/apache/spark/sql/internal/SQLConf.scala @@ -1532,8 +1532,9 @@ object SQLConf { .createWithDefault(false) val PREFER_INTEGRAL_DIVISION = buildConf("spark.sql.function.preferIntegralDivision") + .internal() .doc("When true, will perform integral division with the / operator " + - "if both sides are integral types.") + "if both sides are integral types. This is for PostgreSQL test cases only.") .booleanConf .createWithDefault(false)