diff --git a/common/utils/src/main/scala/org/apache/spark/ErrorClassesJSONReader.scala b/common/utils/src/main/scala/org/apache/spark/ErrorClassesJSONReader.scala index e2dd0da1aac8..85d460f618a7 100644 --- a/common/utils/src/main/scala/org/apache/spark/ErrorClassesJSONReader.scala +++ b/common/utils/src/main/scala/org/apache/spark/ErrorClassesJSONReader.scala @@ -62,7 +62,8 @@ class ErrorClassesJsonReader(jsonFileURLs: Seq[URL]) { } if (util.SparkEnvUtils.isTesting) { val placeHoldersNum = ErrorClassesJsonReader.TEMPLATE_REGEX.findAllIn(messageTemplate).length - if (placeHoldersNum < sanitizedParameters.size) { + if (placeHoldersNum < sanitizedParameters.size && + !ErrorClassesJsonReader.MORE_PARAMS_ALLOWLIST.contains(errorClass)) { throw SparkException.internalError( s"Found unused message parameters of the error class '$errorClass'. " + s"Its error message format has $placeHoldersNum placeholders, " + @@ -123,6 +124,8 @@ class ErrorClassesJsonReader(jsonFileURLs: Seq[URL]) { private object ErrorClassesJsonReader { private val TEMPLATE_REGEX = "<([a-zA-Z0-9_-]+)>".r + private val MORE_PARAMS_ALLOWLIST = Array("CAST_INVALID_INPUT", "CAST_OVERFLOW") + private val mapper: JsonMapper = JsonMapper.builder() .addModule(DefaultScalaModule) .build() diff --git a/sql/api/src/main/scala/org/apache/spark/sql/errors/DataTypeErrors.scala b/sql/api/src/main/scala/org/apache/spark/sql/errors/DataTypeErrors.scala index 664501d22cb3..1e2b2e691cd3 100644 --- a/sql/api/src/main/scala/org/apache/spark/sql/errors/DataTypeErrors.scala +++ b/sql/api/src/main/scala/org/apache/spark/sql/errors/DataTypeErrors.scala @@ -207,7 +207,8 @@ private[sql] object DataTypeErrors extends DataTypeErrorsBase { messageParameters = Map( "expression" -> convertedValueStr, "sourceType" -> toSQLType(StringType), - "targetType" -> toSQLType(to)), + "targetType" -> toSQLType(to), + "ansiConfig" -> toSQLConf("spark.sql.ansi.enabled")), context = getQueryContext(context), summary = getSummary(context)) } @@ -225,8 +226,11 @@ private[sql] object DataTypeErrors extends DataTypeErrorsBase { def castingCauseOverflowError(t: String, from: DataType, to: DataType): ArithmeticException = { new SparkArithmeticException( errorClass = "CAST_OVERFLOW", - messageParameters = - Map("value" -> t, "sourceType" -> toSQLType(from), "targetType" -> toSQLType(to)), + messageParameters = Map( + "value" -> t, + "sourceType" -> toSQLType(from), + "targetType" -> toSQLType(to), + "ansiConfig" -> toSQLConf("spark.sql.ansi.enabled")), context = Array.empty, summary = "") } diff --git a/sql/api/src/main/scala/org/apache/spark/sql/errors/ExecutionErrors.scala b/sql/api/src/main/scala/org/apache/spark/sql/errors/ExecutionErrors.scala index 443bd33ad60b..8124b1a4ab19 100644 --- a/sql/api/src/main/scala/org/apache/spark/sql/errors/ExecutionErrors.scala +++ b/sql/api/src/main/scala/org/apache/spark/sql/errors/ExecutionErrors.scala @@ -109,7 +109,8 @@ private[sql] trait ExecutionErrors extends DataTypeErrorsBase { messageParameters = Map( "expression" -> sqlValue, "sourceType" -> toSQLType(from), - "targetType" -> toSQLType(to)), + "targetType" -> toSQLType(to), + "ansiConfig" -> toSQLConf("spark.sql.ansi.enabled")), context = getQueryContext(context), summary = getSummary(context)) } diff --git a/sql/catalyst/src/main/scala/org/apache/spark/sql/errors/QueryExecutionErrors.scala b/sql/catalyst/src/main/scala/org/apache/spark/sql/errors/QueryExecutionErrors.scala index d0ece5baff43..5f07d7d73f1c 100644 --- a/sql/catalyst/src/main/scala/org/apache/spark/sql/errors/QueryExecutionErrors.scala +++ b/sql/catalyst/src/main/scala/org/apache/spark/sql/errors/QueryExecutionErrors.scala @@ -79,7 +79,8 @@ private[sql] object QueryExecutionErrors extends QueryErrorsBase with ExecutionE messageParameters = Map( "value" -> toSQLValue(t, from), "sourceType" -> toSQLType(from), - "targetType" -> toSQLType(to)), + "targetType" -> toSQLType(to), + "ansiConfig" -> toSQLConf("spark.sql.ansi.enabled")), context = Array.empty, summary = "") } @@ -123,7 +124,8 @@ private[sql] object QueryExecutionErrors extends QueryErrorsBase with ExecutionE messageParameters = Map( "expression" -> toSQLValue(s, StringType), "sourceType" -> toSQLType(StringType), - "targetType" -> toSQLType(BooleanType)), + "targetType" -> toSQLType(BooleanType), + "ansiConfig" -> toSQLConf("spark.sql.ansi.enabled")), context = getQueryContext(context), summary = getSummary(context)) } @@ -137,7 +139,8 @@ private[sql] object QueryExecutionErrors extends QueryErrorsBase with ExecutionE messageParameters = Map( "expression" -> toSQLValue(s, StringType), "sourceType" -> toSQLType(StringType), - "targetType" -> toSQLType(to)), + "targetType" -> toSQLType(to), + "ansiConfig" -> toSQLConf("spark.sql.ansi.enabled")), context = getQueryContext(context), summary = getSummary(context)) } diff --git a/sql/catalyst/src/test/scala/org/apache/spark/sql/catalyst/util/DateTimeUtilsSuite.scala b/sql/catalyst/src/test/scala/org/apache/spark/sql/catalyst/util/DateTimeUtilsSuite.scala index 81c57f1aca91..ab57e476b1f2 100644 --- a/sql/catalyst/src/test/scala/org/apache/spark/sql/catalyst/util/DateTimeUtilsSuite.scala +++ b/sql/catalyst/src/test/scala/org/apache/spark/sql/catalyst/util/DateTimeUtilsSuite.scala @@ -1163,7 +1163,8 @@ class DateTimeUtilsSuite extends SparkFunSuite with Matchers with SQLHelper { parameters = Map( "expression" -> s"'$invalidTime'", "sourceType" -> "\"STRING\"", - "targetType" -> "\"TIME(6)\"")) + "targetType" -> "\"TIME(6)\"", + "ansiConfig" -> "\"spark.sql.ansi.enabled\"")) } } } diff --git a/sql/catalyst/src/test/scala/org/apache/spark/sql/catalyst/util/TimeFormatterSuite.scala b/sql/catalyst/src/test/scala/org/apache/spark/sql/catalyst/util/TimeFormatterSuite.scala index d47f103554b1..d99ea2bd1042 100644 --- a/sql/catalyst/src/test/scala/org/apache/spark/sql/catalyst/util/TimeFormatterSuite.scala +++ b/sql/catalyst/src/test/scala/org/apache/spark/sql/catalyst/util/TimeFormatterSuite.scala @@ -171,6 +171,7 @@ class TimeFormatterSuite extends SparkFunSuite with SQLHelper { parameters = Map( "expression" -> "'x123'", "sourceType" -> "\"STRING\"", - "targetType" -> "\"TIME(6)\"")) + "targetType" -> "\"TIME(6)\"", + "ansiConfig" -> "\"spark.sql.ansi.enabled\"")) } } diff --git a/sql/catalyst/src/test/scala/org/apache/spark/sql/types/DecimalSuite.scala b/sql/catalyst/src/test/scala/org/apache/spark/sql/types/DecimalSuite.scala index 821d59796753..4af2c8367819 100644 --- a/sql/catalyst/src/test/scala/org/apache/spark/sql/types/DecimalSuite.scala +++ b/sql/catalyst/src/test/scala/org/apache/spark/sql/types/DecimalSuite.scala @@ -382,7 +382,8 @@ class DecimalSuite extends SparkFunSuite with PrivateMethodTester with SQLHelper parameters = Map( "expression" -> "'str'", "sourceType" -> "\"STRING\"", - "targetType" -> "\"DECIMAL(10,0)\"")) + "targetType" -> "\"DECIMAL(10,0)\"", + "ansiConfig" -> "\"spark.sql.ansi.enabled\"")) } test("SPARK-35841: Casting string to decimal type doesn't work " + diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/execute-immediate.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/execute-immediate.sql.out index ce510527c878..78bf1ccb1678 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/execute-immediate.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/execute-immediate.sql.out @@ -471,6 +471,7 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'invalid_cast_error_expected'", "sourceType" : "\"STRING\"", "targetType" : "\"INT\"" @@ -661,6 +662,7 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'name1'", "sourceType" : "\"STRING\"", "targetType" : "\"INT\"" diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/postgreSQL/window_part2.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/postgreSQL/window_part2.sql.out index 330e1c1cad7e..cdcd563de4f6 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/postgreSQL/window_part2.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/postgreSQL/window_part2.sql.out @@ -449,6 +449,7 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'NaN'", "sourceType" : "\"STRING\"", "targetType" : "\"INT\"" diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/sql-session-variables.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/sql-session-variables.sql.out index b83cc337ae3a..6f22f598498c 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/sql-session-variables.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/sql-session-variables.sql.out @@ -842,6 +842,7 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'hello'", "sourceType" : "\"STRING\"", "targetType" : "\"INT\"" @@ -884,6 +885,7 @@ org.apache.spark.SparkArithmeticException "errorClass" : "CAST_OVERFLOW", "sqlState" : "22003", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "sourceType" : "\"INT\"", "targetType" : "\"SMALLINT\"", "value" : "100000" @@ -1000,6 +1002,7 @@ org.apache.spark.SparkArithmeticException "errorClass" : "CAST_OVERFLOW", "sqlState" : "22003", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "sourceType" : "\"DOUBLE\"", "targetType" : "\"INT\"", "value" : "1.0E10D" @@ -1059,6 +1062,7 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'hello'", "sourceType" : "\"STRING\"", "targetType" : "\"INT\"" diff --git a/sql/core/src/test/resources/sql-tests/results/cast.sql.out b/sql/core/src/test/resources/sql-tests/results/cast.sql.out index 0dbdf1d9975c..7dd7180165f2 100644 --- a/sql/core/src/test/resources/sql-tests/results/cast.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/cast.sql.out @@ -9,6 +9,7 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'1.23'", "sourceType" : "\"STRING\"", "targetType" : "\"INT\"" @@ -33,6 +34,7 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'1.23'", "sourceType" : "\"STRING\"", "targetType" : "\"BIGINT\"" @@ -57,6 +59,7 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'-4.56'", "sourceType" : "\"STRING\"", "targetType" : "\"INT\"" @@ -81,6 +84,7 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'-4.56'", "sourceType" : "\"STRING\"", "targetType" : "\"BIGINT\"" @@ -105,6 +109,7 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'abc'", "sourceType" : "\"STRING\"", "targetType" : "\"INT\"" @@ -129,6 +134,7 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'abc'", "sourceType" : "\"STRING\"", "targetType" : "\"BIGINT\"" @@ -153,6 +159,7 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'abc'", "sourceType" : "\"STRING\"", "targetType" : "\"FLOAT\"" @@ -177,6 +184,7 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'abc'", "sourceType" : "\"STRING\"", "targetType" : "\"DOUBLE\"" @@ -201,6 +209,7 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'1234567890123'", "sourceType" : "\"STRING\"", "targetType" : "\"INT\"" @@ -225,6 +234,7 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'12345678901234567890123'", "sourceType" : "\"STRING\"", "targetType" : "\"BIGINT\"" @@ -249,6 +259,7 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "''", "sourceType" : "\"STRING\"", "targetType" : "\"INT\"" @@ -273,6 +284,7 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "''", "sourceType" : "\"STRING\"", "targetType" : "\"BIGINT\"" @@ -297,6 +309,7 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "''", "sourceType" : "\"STRING\"", "targetType" : "\"FLOAT\"" @@ -321,6 +334,7 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "''", "sourceType" : "\"STRING\"", "targetType" : "\"DOUBLE\"" @@ -361,6 +375,7 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'123.a'", "sourceType" : "\"STRING\"", "targetType" : "\"INT\"" @@ -385,6 +400,7 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'123.a'", "sourceType" : "\"STRING\"", "targetType" : "\"BIGINT\"" @@ -409,6 +425,7 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'123.a'", "sourceType" : "\"STRING\"", "targetType" : "\"FLOAT\"" @@ -433,6 +450,7 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'123.a'", "sourceType" : "\"STRING\"", "targetType" : "\"DOUBLE\"" @@ -465,6 +483,7 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'-2147483649'", "sourceType" : "\"STRING\"", "targetType" : "\"INT\"" @@ -497,6 +516,7 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'2147483648'", "sourceType" : "\"STRING\"", "targetType" : "\"INT\"" @@ -529,6 +549,7 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'-9223372036854775809'", "sourceType" : "\"STRING\"", "targetType" : "\"BIGINT\"" @@ -561,6 +582,7 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'9223372036854775808'", "sourceType" : "\"STRING\"", "targetType" : "\"BIGINT\"" @@ -951,6 +973,7 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'1中文'", "sourceType" : "\"STRING\"", "targetType" : "\"TINYINT\"" @@ -975,6 +998,7 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'1中文'", "sourceType" : "\"STRING\"", "targetType" : "\"SMALLINT\"" @@ -999,6 +1023,7 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'1中文'", "sourceType" : "\"STRING\"", "targetType" : "\"INT\"" @@ -1023,6 +1048,7 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'中文1'", "sourceType" : "\"STRING\"", "targetType" : "\"BIGINT\"" @@ -1047,6 +1073,7 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'1中文'", "sourceType" : "\"STRING\"", "targetType" : "\"BIGINT\"" @@ -1089,6 +1116,7 @@ org.apache.spark.SparkRuntimeException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'\t\n xyz \t\r'", "sourceType" : "\"STRING\"", "targetType" : "\"BOOLEAN\"" @@ -1146,6 +1174,7 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'xyz'", "sourceType" : "\"STRING\"", "targetType" : "\"DECIMAL(4,2)\"" @@ -1178,6 +1207,7 @@ org.apache.spark.SparkDateTimeException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'a'", "sourceType" : "\"STRING\"", "targetType" : "\"DATE\"" @@ -1210,6 +1240,7 @@ org.apache.spark.SparkDateTimeException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'a'", "sourceType" : "\"STRING\"", "targetType" : "\"TIMESTAMP\"" @@ -1242,6 +1273,7 @@ org.apache.spark.SparkDateTimeException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'a'", "sourceType" : "\"STRING\"", "targetType" : "\"TIMESTAMP_NTZ\"" @@ -1266,6 +1298,7 @@ org.apache.spark.SparkDateTimeException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "Infinity", "sourceType" : "\"DOUBLE\"", "targetType" : "\"TIMESTAMP\"" @@ -1290,6 +1323,7 @@ org.apache.spark.SparkDateTimeException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "Infinity", "sourceType" : "\"DOUBLE\"", "targetType" : "\"TIMESTAMP\"" @@ -1346,6 +1380,7 @@ org.apache.spark.SparkArithmeticException "errorClass" : "CAST_OVERFLOW", "sqlState" : "22003", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "sourceType" : "\"INTERVAL HOUR TO SECOND\"", "targetType" : "\"SMALLINT\"", "value" : "INTERVAL '23:59:59' HOUR TO SECOND" @@ -1379,6 +1414,7 @@ org.apache.spark.SparkArithmeticException "errorClass" : "CAST_OVERFLOW", "sqlState" : "22003", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "sourceType" : "\"INTERVAL MONTH\"", "targetType" : "\"TINYINT\"", "value" : "INTERVAL '-1000' MONTH" @@ -1396,6 +1432,7 @@ org.apache.spark.SparkArithmeticException "errorClass" : "CAST_OVERFLOW", "sqlState" : "22003", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "sourceType" : "\"INTERVAL SECOND\"", "targetType" : "\"SMALLINT\"", "value" : "INTERVAL '1000000' SECOND" @@ -1485,6 +1522,7 @@ org.apache.spark.SparkArithmeticException "errorClass" : "CAST_OVERFLOW", "sqlState" : "22003", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "sourceType" : "\"INT\"", "targetType" : "\"INTERVAL YEAR\"", "value" : "2147483647" @@ -1502,6 +1540,7 @@ org.apache.spark.SparkArithmeticException "errorClass" : "CAST_OVERFLOW", "sqlState" : "22003", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "sourceType" : "\"BIGINT\"", "targetType" : "\"INTERVAL DAY\"", "value" : "-9223372036854775808L" @@ -1632,6 +1671,7 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'1.23'", "sourceType" : "\"STRING\"", "targetType" : "\"INT\"" @@ -1656,6 +1696,7 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'abc'", "sourceType" : "\"STRING\"", "targetType" : "\"INT\"" @@ -1680,6 +1721,7 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'12345678901234567890123'", "sourceType" : "\"STRING\"", "targetType" : "\"BIGINT\"" @@ -1704,6 +1746,7 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "''", "sourceType" : "\"STRING\"", "targetType" : "\"INT\"" @@ -1736,6 +1779,7 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'123.a'", "sourceType" : "\"STRING\"", "targetType" : "\"INT\"" @@ -1883,6 +1927,7 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'1.23'", "sourceType" : "\"STRING\"", "targetType" : "\"INT\"" @@ -1907,6 +1952,7 @@ org.apache.spark.SparkArithmeticException "errorClass" : "CAST_OVERFLOW", "sqlState" : "22003", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "sourceType" : "\"BIGINT\"", "targetType" : "\"INT\"", "value" : "2147483648L" @@ -1924,6 +1970,7 @@ org.apache.spark.SparkArithmeticException "errorClass" : "CAST_OVERFLOW", "sqlState" : "22003", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "sourceType" : "\"BIGINT\"", "targetType" : "\"INT\"", "value" : "2147483648L" diff --git a/sql/core/src/test/resources/sql-tests/results/conditional-functions.sql.out b/sql/core/src/test/resources/sql-tests/results/conditional-functions.sql.out index aa8a600f8756..26293cad10ce 100644 --- a/sql/core/src/test/resources/sql-tests/results/conditional-functions.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/conditional-functions.sql.out @@ -145,6 +145,7 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'abc'", "sourceType" : "\"STRING\"", "targetType" : "\"BIGINT\"" @@ -180,6 +181,7 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'abc'", "sourceType" : "\"STRING\"", "targetType" : "\"BIGINT\"" diff --git a/sql/core/src/test/resources/sql-tests/results/date.sql.out b/sql/core/src/test/resources/sql-tests/results/date.sql.out index 37cf05c6386d..5c47f5add920 100644 --- a/sql/core/src/test/resources/sql-tests/results/date.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/date.sql.out @@ -312,6 +312,7 @@ org.apache.spark.SparkDateTimeException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'xx'", "sourceType" : "\"STRING\"", "targetType" : "\"DATE\"" @@ -470,6 +471,7 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'1.2'", "sourceType" : "\"STRING\"", "targetType" : "\"INT\"" @@ -644,6 +646,7 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'1.2'", "sourceType" : "\"STRING\"", "targetType" : "\"INT\"" diff --git a/sql/core/src/test/resources/sql-tests/results/datetime-legacy.sql.out b/sql/core/src/test/resources/sql-tests/results/datetime-legacy.sql.out index feae373b52ca..518f83fe0300 100644 --- a/sql/core/src/test/resources/sql-tests/results/datetime-legacy.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/datetime-legacy.sql.out @@ -312,6 +312,7 @@ org.apache.spark.SparkDateTimeException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'xx'", "sourceType" : "\"STRING\"", "targetType" : "\"DATE\"" @@ -470,6 +471,7 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'1.2'", "sourceType" : "\"STRING\"", "targetType" : "\"INT\"" @@ -644,6 +646,7 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'1.2'", "sourceType" : "\"STRING\"", "targetType" : "\"INT\"" diff --git a/sql/core/src/test/resources/sql-tests/results/datetime-parsing-invalid.sql.out b/sql/core/src/test/resources/sql-tests/results/datetime-parsing-invalid.sql.out index cdc027846c34..e03bd33b0715 100644 --- a/sql/core/src/test/resources/sql-tests/results/datetime-parsing-invalid.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/datetime-parsing-invalid.sql.out @@ -427,6 +427,7 @@ org.apache.spark.SparkDateTimeException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'Unparseable'", "sourceType" : "\"STRING\"", "targetType" : "\"TIMESTAMP\"" @@ -451,6 +452,7 @@ org.apache.spark.SparkDateTimeException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'Unparseable'", "sourceType" : "\"STRING\"", "targetType" : "\"DATE\"" diff --git a/sql/core/src/test/resources/sql-tests/results/execute-immediate.sql.out b/sql/core/src/test/resources/sql-tests/results/execute-immediate.sql.out index 21ea4436f4fa..9249d7eb3e51 100644 --- a/sql/core/src/test/resources/sql-tests/results/execute-immediate.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/execute-immediate.sql.out @@ -392,6 +392,7 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'invalid_cast_error_expected'", "sourceType" : "\"STRING\"", "targetType" : "\"INT\"" @@ -602,6 +603,7 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'name1'", "sourceType" : "\"STRING\"", "targetType" : "\"INT\"" diff --git a/sql/core/src/test/resources/sql-tests/results/interval.sql.out b/sql/core/src/test/resources/sql-tests/results/interval.sql.out index 4e012df792de..c32ecc19448a 100644 --- a/sql/core/src/test/resources/sql-tests/results/interval.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/interval.sql.out @@ -130,6 +130,7 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'a'", "sourceType" : "\"STRING\"", "targetType" : "\"DOUBLE\"" @@ -154,6 +155,7 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'a'", "sourceType" : "\"STRING\"", "targetType" : "\"DOUBLE\"" @@ -178,6 +180,7 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'a'", "sourceType" : "\"STRING\"", "targetType" : "\"DOUBLE\"" @@ -202,6 +205,7 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'a'", "sourceType" : "\"STRING\"", "targetType" : "\"DOUBLE\"" @@ -242,6 +246,7 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'a'", "sourceType" : "\"STRING\"", "targetType" : "\"DOUBLE\"" @@ -266,6 +271,7 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'a'", "sourceType" : "\"STRING\"", "targetType" : "\"DOUBLE\"" @@ -1997,6 +2003,7 @@ org.apache.spark.SparkDateTimeException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'4 11:11'", "sourceType" : "\"STRING\"", "targetType" : "\"TIMESTAMP\"" @@ -2021,6 +2028,7 @@ org.apache.spark.SparkDateTimeException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'4 12:12:12'", "sourceType" : "\"STRING\"", "targetType" : "\"TIMESTAMP\"" @@ -2101,6 +2109,7 @@ org.apache.spark.SparkDateTimeException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'1'", "sourceType" : "\"STRING\"", "targetType" : "\"TIMESTAMP\"" @@ -2125,6 +2134,7 @@ org.apache.spark.SparkDateTimeException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'1'", "sourceType" : "\"STRING\"", "targetType" : "\"TIMESTAMP\"" diff --git a/sql/core/src/test/resources/sql-tests/results/math.sql.out b/sql/core/src/test/resources/sql-tests/results/math.sql.out index fb60a920040e..e2abcb099130 100644 --- a/sql/core/src/test/resources/sql-tests/results/math.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/math.sql.out @@ -881,6 +881,7 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'invalid'", "sourceType" : "\"STRING\"", "targetType" : "\"DOUBLE\"" diff --git a/sql/core/src/test/resources/sql-tests/results/nonansi/cast.sql.out b/sql/core/src/test/resources/sql-tests/results/nonansi/cast.sql.out index 738697c63883..6f74c63da354 100644 --- a/sql/core/src/test/resources/sql-tests/results/nonansi/cast.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/nonansi/cast.sql.out @@ -642,6 +642,7 @@ org.apache.spark.SparkArithmeticException "errorClass" : "CAST_OVERFLOW", "sqlState" : "22003", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "sourceType" : "\"INTERVAL HOUR TO SECOND\"", "targetType" : "\"SMALLINT\"", "value" : "INTERVAL '23:59:59' HOUR TO SECOND" @@ -675,6 +676,7 @@ org.apache.spark.SparkArithmeticException "errorClass" : "CAST_OVERFLOW", "sqlState" : "22003", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "sourceType" : "\"INTERVAL MONTH\"", "targetType" : "\"TINYINT\"", "value" : "INTERVAL '-1000' MONTH" @@ -692,6 +694,7 @@ org.apache.spark.SparkArithmeticException "errorClass" : "CAST_OVERFLOW", "sqlState" : "22003", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "sourceType" : "\"INTERVAL SECOND\"", "targetType" : "\"SMALLINT\"", "value" : "INTERVAL '1000000' SECOND" @@ -781,6 +784,7 @@ org.apache.spark.SparkArithmeticException "errorClass" : "CAST_OVERFLOW", "sqlState" : "22003", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "sourceType" : "\"INT\"", "targetType" : "\"INTERVAL YEAR\"", "value" : "2147483647" @@ -798,6 +802,7 @@ org.apache.spark.SparkArithmeticException "errorClass" : "CAST_OVERFLOW", "sqlState" : "22003", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "sourceType" : "\"BIGINT\"", "targetType" : "\"INTERVAL DAY\"", "value" : "-9223372036854775808L" diff --git a/sql/core/src/test/resources/sql-tests/results/postgreSQL/boolean.sql.out b/sql/core/src/test/resources/sql-tests/results/postgreSQL/boolean.sql.out index 052e7b4f2522..12660768b95c 100644 --- a/sql/core/src/test/resources/sql-tests/results/postgreSQL/boolean.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/postgreSQL/boolean.sql.out @@ -57,6 +57,7 @@ org.apache.spark.SparkRuntimeException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'test'", "sourceType" : "\"STRING\"", "targetType" : "\"BOOLEAN\"" @@ -89,6 +90,7 @@ org.apache.spark.SparkRuntimeException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'foo'", "sourceType" : "\"STRING\"", "targetType" : "\"BOOLEAN\"" @@ -129,6 +131,7 @@ org.apache.spark.SparkRuntimeException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'yeah'", "sourceType" : "\"STRING\"", "targetType" : "\"BOOLEAN\"" @@ -169,6 +172,7 @@ org.apache.spark.SparkRuntimeException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'nay'", "sourceType" : "\"STRING\"", "targetType" : "\"BOOLEAN\"" @@ -193,6 +197,7 @@ org.apache.spark.SparkRuntimeException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'on'", "sourceType" : "\"STRING\"", "targetType" : "\"BOOLEAN\"" @@ -217,6 +222,7 @@ org.apache.spark.SparkRuntimeException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'off'", "sourceType" : "\"STRING\"", "targetType" : "\"BOOLEAN\"" @@ -241,6 +247,7 @@ org.apache.spark.SparkRuntimeException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'of'", "sourceType" : "\"STRING\"", "targetType" : "\"BOOLEAN\"" @@ -265,6 +272,7 @@ org.apache.spark.SparkRuntimeException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'o'", "sourceType" : "\"STRING\"", "targetType" : "\"BOOLEAN\"" @@ -289,6 +297,7 @@ org.apache.spark.SparkRuntimeException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'on_'", "sourceType" : "\"STRING\"", "targetType" : "\"BOOLEAN\"" @@ -313,6 +322,7 @@ org.apache.spark.SparkRuntimeException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'off_'", "sourceType" : "\"STRING\"", "targetType" : "\"BOOLEAN\"" @@ -345,6 +355,7 @@ org.apache.spark.SparkRuntimeException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'11'", "sourceType" : "\"STRING\"", "targetType" : "\"BOOLEAN\"" @@ -377,6 +388,7 @@ org.apache.spark.SparkRuntimeException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'000'", "sourceType" : "\"STRING\"", "targetType" : "\"BOOLEAN\"" @@ -401,6 +413,7 @@ org.apache.spark.SparkRuntimeException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "''", "sourceType" : "\"STRING\"", "targetType" : "\"BOOLEAN\"" @@ -522,6 +535,7 @@ org.apache.spark.SparkRuntimeException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "' tru e '", "sourceType" : "\"STRING\"", "targetType" : "\"BOOLEAN\"" @@ -546,6 +560,7 @@ org.apache.spark.SparkRuntimeException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "''", "sourceType" : "\"STRING\"", "targetType" : "\"BOOLEAN\"" diff --git a/sql/core/src/test/resources/sql-tests/results/postgreSQL/float4.sql.out b/sql/core/src/test/resources/sql-tests/results/postgreSQL/float4.sql.out index 1a15610b4ded..6b4b343d9cca 100644 --- a/sql/core/src/test/resources/sql-tests/results/postgreSQL/float4.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/postgreSQL/float4.sql.out @@ -97,6 +97,7 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'N A N'", "sourceType" : "\"STRING\"", "targetType" : "\"FLOAT\"" @@ -121,6 +122,7 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'NaN x'", "sourceType" : "\"STRING\"", "targetType" : "\"FLOAT\"" @@ -145,6 +147,7 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "' INFINITY x'", "sourceType" : "\"STRING\"", "targetType" : "\"FLOAT\"" @@ -193,6 +196,7 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'nan'", "sourceType" : "\"STRING\"", "targetType" : "\"DECIMAL(10,0)\"" @@ -389,6 +393,7 @@ org.apache.spark.SparkArithmeticException "errorClass" : "CAST_OVERFLOW", "sqlState" : "22003", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "sourceType" : "\"FLOAT\"", "targetType" : "\"INT\"", "value" : "2.14748365E9" @@ -414,6 +419,7 @@ org.apache.spark.SparkArithmeticException "errorClass" : "CAST_OVERFLOW", "sqlState" : "22003", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "sourceType" : "\"FLOAT\"", "targetType" : "\"INT\"", "value" : "-2.1474839E9" @@ -455,6 +461,7 @@ org.apache.spark.SparkArithmeticException "errorClass" : "CAST_OVERFLOW", "sqlState" : "22003", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "sourceType" : "\"FLOAT\"", "targetType" : "\"BIGINT\"", "value" : "-9.22338E18" diff --git a/sql/core/src/test/resources/sql-tests/results/postgreSQL/float8.sql.out b/sql/core/src/test/resources/sql-tests/results/postgreSQL/float8.sql.out index b1a114bea30e..e1b880f34370 100644 --- a/sql/core/src/test/resources/sql-tests/results/postgreSQL/float8.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/postgreSQL/float8.sql.out @@ -129,6 +129,7 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'N A N'", "sourceType" : "\"STRING\"", "targetType" : "\"DOUBLE\"" @@ -153,6 +154,7 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'NaN x'", "sourceType" : "\"STRING\"", "targetType" : "\"DOUBLE\"" @@ -177,6 +179,7 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "' INFINITY x'", "sourceType" : "\"STRING\"", "targetType" : "\"DOUBLE\"" @@ -225,6 +228,7 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'nan'", "sourceType" : "\"STRING\"", "targetType" : "\"DECIMAL(10,0)\"" @@ -894,6 +898,7 @@ org.apache.spark.SparkArithmeticException "errorClass" : "CAST_OVERFLOW", "sqlState" : "22003", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "sourceType" : "\"DOUBLE\"", "targetType" : "\"BIGINT\"", "value" : "-9.22337203685478E18D" diff --git a/sql/core/src/test/resources/sql-tests/results/postgreSQL/int8.sql.out b/sql/core/src/test/resources/sql-tests/results/postgreSQL/int8.sql.out index 5e8abc273b12..f6e4bd8bd7e0 100755 --- a/sql/core/src/test/resources/sql-tests/results/postgreSQL/int8.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/postgreSQL/int8.sql.out @@ -737,6 +737,7 @@ org.apache.spark.SparkArithmeticException "errorClass" : "CAST_OVERFLOW", "sqlState" : "22003", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "sourceType" : "\"BIGINT\"", "targetType" : "\"INT\"", "value" : "4567890123456789L" @@ -762,6 +763,7 @@ org.apache.spark.SparkArithmeticException "errorClass" : "CAST_OVERFLOW", "sqlState" : "22003", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "sourceType" : "\"BIGINT\"", "targetType" : "\"SMALLINT\"", "value" : "4567890123456789L" @@ -807,6 +809,7 @@ org.apache.spark.SparkArithmeticException "errorClass" : "CAST_OVERFLOW", "sqlState" : "22003", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "sourceType" : "\"DOUBLE\"", "targetType" : "\"BIGINT\"", "value" : "9.223372036854776E20D" @@ -895,6 +898,7 @@ org.apache.spark.SparkArithmeticException "errorClass" : "CAST_OVERFLOW", "sqlState" : "22003", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "sourceType" : "\"BIGINT\"", "targetType" : "\"INT\"", "value" : "-9223372036854775808L" diff --git a/sql/core/src/test/resources/sql-tests/results/postgreSQL/text.sql.out b/sql/core/src/test/resources/sql-tests/results/postgreSQL/text.sql.out index 0a940f5f3c74..37b8a3e8fd19 100755 --- a/sql/core/src/test/resources/sql-tests/results/postgreSQL/text.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/postgreSQL/text.sql.out @@ -66,6 +66,7 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'four: 2'", "sourceType" : "\"STRING\"", "targetType" : "\"BIGINT\"" @@ -90,6 +91,7 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'four: 2'", "sourceType" : "\"STRING\"", "targetType" : "\"BIGINT\"" diff --git a/sql/core/src/test/resources/sql-tests/results/postgreSQL/union.sql.out b/sql/core/src/test/resources/sql-tests/results/postgreSQL/union.sql.out index 94692a57300f..7c920bbd32b3 100644 --- a/sql/core/src/test/resources/sql-tests/results/postgreSQL/union.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/postgreSQL/union.sql.out @@ -700,6 +700,7 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'foo'", "sourceType" : "\"STRING\"", "targetType" : "\"DOUBLE\"" diff --git a/sql/core/src/test/resources/sql-tests/results/postgreSQL/window_part2.sql.out b/sql/core/src/test/resources/sql-tests/results/postgreSQL/window_part2.sql.out index 352c5f05cb06..6cf5e69758d2 100644 --- a/sql/core/src/test/resources/sql-tests/results/postgreSQL/window_part2.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/postgreSQL/window_part2.sql.out @@ -489,6 +489,7 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'NaN'", "sourceType" : "\"STRING\"", "targetType" : "\"INT\"" diff --git a/sql/core/src/test/resources/sql-tests/results/predicate-functions.sql.out b/sql/core/src/test/resources/sql-tests/results/predicate-functions.sql.out index e2d0563a0c45..10ff268a5606 100644 --- a/sql/core/src/test/resources/sql-tests/results/predicate-functions.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/predicate-functions.sql.out @@ -217,6 +217,7 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'1.0'", "sourceType" : "\"STRING\"", "targetType" : "\"BIGINT\"" @@ -241,6 +242,7 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'2.0'", "sourceType" : "\"STRING\"", "targetType" : "\"BIGINT\"" @@ -265,6 +267,7 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'2.2'", "sourceType" : "\"STRING\"", "targetType" : "\"BIGINT\"" @@ -321,6 +324,7 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'1.0'", "sourceType" : "\"STRING\"", "targetType" : "\"BIGINT\"" @@ -345,6 +349,7 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'2.0'", "sourceType" : "\"STRING\"", "targetType" : "\"BIGINT\"" @@ -409,6 +414,7 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'1.0'", "sourceType" : "\"STRING\"", "targetType" : "\"BIGINT\"" @@ -433,6 +439,7 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'2.0'", "sourceType" : "\"STRING\"", "targetType" : "\"BIGINT\"" @@ -497,6 +504,7 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'1.0'", "sourceType" : "\"STRING\"", "targetType" : "\"BIGINT\"" @@ -521,6 +529,7 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'2.0'", "sourceType" : "\"STRING\"", "targetType" : "\"BIGINT\"" diff --git a/sql/core/src/test/resources/sql-tests/results/sql-session-variables.sql.out b/sql/core/src/test/resources/sql-tests/results/sql-session-variables.sql.out index 249a03fdfbf8..73d3ec737085 100644 --- a/sql/core/src/test/resources/sql-tests/results/sql-session-variables.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/sql-session-variables.sql.out @@ -943,6 +943,7 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'hello'", "sourceType" : "\"STRING\"", "targetType" : "\"INT\"" @@ -989,6 +990,7 @@ org.apache.spark.SparkArithmeticException "errorClass" : "CAST_OVERFLOW", "sqlState" : "22003", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "sourceType" : "\"INT\"", "targetType" : "\"SMALLINT\"", "value" : "100000" @@ -1102,6 +1104,7 @@ org.apache.spark.SparkArithmeticException "errorClass" : "CAST_OVERFLOW", "sqlState" : "22003", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "sourceType" : "\"DOUBLE\"", "targetType" : "\"INT\"", "value" : "1.0E10D" @@ -1168,6 +1171,7 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'hello'", "sourceType" : "\"STRING\"", "targetType" : "\"INT\"" diff --git a/sql/core/src/test/resources/sql-tests/results/string-functions.sql.out b/sql/core/src/test/resources/sql-tests/results/string-functions.sql.out index 7bbef71854d2..34a560d771e8 100644 --- a/sql/core/src/test/resources/sql-tests/results/string-functions.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/string-functions.sql.out @@ -101,6 +101,7 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'a'", "sourceType" : "\"STRING\"", "targetType" : "\"INT\"" @@ -141,6 +142,7 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'a'", "sourceType" : "\"STRING\"", "targetType" : "\"INT\"" @@ -505,6 +507,7 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'invalid_length'", "sourceType" : "\"STRING\"", "targetType" : "\"INT\"" @@ -529,6 +532,7 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'invalid_length'", "sourceType" : "\"STRING\"", "targetType" : "\"INT\"" diff --git a/sql/core/src/test/resources/sql-tests/results/timestampNTZ/timestamp-ansi.sql.out b/sql/core/src/test/resources/sql-tests/results/timestampNTZ/timestamp-ansi.sql.out index 08f462bdd5e3..9e138a95fc37 100644 --- a/sql/core/src/test/resources/sql-tests/results/timestampNTZ/timestamp-ansi.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/timestampNTZ/timestamp-ansi.sql.out @@ -387,6 +387,7 @@ org.apache.spark.SparkDateTimeException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'1'", "sourceType" : "\"STRING\"", "targetType" : "\"TIMESTAMP_NTZ\"" diff --git a/sql/core/src/test/resources/sql-tests/results/typeCoercion/native/caseWhenCoercion.sql.out b/sql/core/src/test/resources/sql-tests/results/typeCoercion/native/caseWhenCoercion.sql.out index 35ff9e79d980..252d8c884ca7 100644 --- a/sql/core/src/test/resources/sql-tests/results/typeCoercion/native/caseWhenCoercion.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/typeCoercion/native/caseWhenCoercion.sql.out @@ -1217,6 +1217,7 @@ org.apache.spark.SparkDateTimeException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'1'", "sourceType" : "\"STRING\"", "targetType" : "\"TIMESTAMP\"" @@ -1241,6 +1242,7 @@ org.apache.spark.SparkDateTimeException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'1'", "sourceType" : "\"STRING\"", "targetType" : "\"DATE\"" diff --git a/sql/core/src/test/resources/sql-tests/results/typeCoercion/native/concat.sql.out b/sql/core/src/test/resources/sql-tests/results/typeCoercion/native/concat.sql.out index 0f42834d2824..3e28c696a259 100644 --- a/sql/core/src/test/resources/sql-tests/results/typeCoercion/native/concat.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/typeCoercion/native/concat.sql.out @@ -318,6 +318,7 @@ org.apache.spark.SparkDateTimeException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'a'", "sourceType" : "\"STRING\"", "targetType" : "\"DATE\"" diff --git a/sql/core/src/test/resources/sql-tests/results/typeCoercion/native/ifCoercion.sql.out b/sql/core/src/test/resources/sql-tests/results/typeCoercion/native/ifCoercion.sql.out index b2e0f50028a4..ffcfba2717d6 100644 --- a/sql/core/src/test/resources/sql-tests/results/typeCoercion/native/ifCoercion.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/typeCoercion/native/ifCoercion.sql.out @@ -1217,6 +1217,7 @@ org.apache.spark.SparkDateTimeException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'1'", "sourceType" : "\"STRING\"", "targetType" : "\"TIMESTAMP\"" @@ -1241,6 +1242,7 @@ org.apache.spark.SparkDateTimeException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'1'", "sourceType" : "\"STRING\"", "targetType" : "\"DATE\"" diff --git a/sql/core/src/test/resources/sql-tests/results/typeCoercion/native/inConversion.sql.out b/sql/core/src/test/resources/sql-tests/results/typeCoercion/native/inConversion.sql.out index 7c9152a66a9c..6005466b8e2f 100644 --- a/sql/core/src/test/resources/sql-tests/results/typeCoercion/native/inConversion.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/typeCoercion/native/inConversion.sql.out @@ -1217,6 +1217,7 @@ org.apache.spark.SparkDateTimeException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'1'", "sourceType" : "\"STRING\"", "targetType" : "\"TIMESTAMP\"" @@ -1241,6 +1242,7 @@ org.apache.spark.SparkDateTimeException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'1'", "sourceType" : "\"STRING\"", "targetType" : "\"DATE\"" @@ -1945,6 +1947,7 @@ org.apache.spark.SparkDateTimeException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'2'", "sourceType" : "\"STRING\"", "targetType" : "\"TIMESTAMP\"" @@ -2201,6 +2204,7 @@ org.apache.spark.SparkDateTimeException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'2'", "sourceType" : "\"STRING\"", "targetType" : "\"DATE\"" @@ -3489,6 +3493,7 @@ org.apache.spark.SparkDateTimeException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'1'", "sourceType" : "\"STRING\"", "targetType" : "\"TIMESTAMP\"" @@ -3513,6 +3518,7 @@ org.apache.spark.SparkDateTimeException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'1'", "sourceType" : "\"STRING\"", "targetType" : "\"DATE\"" diff --git a/sql/core/src/test/resources/sql-tests/results/typeCoercion/native/promoteStrings.sql.out b/sql/core/src/test/resources/sql-tests/results/typeCoercion/native/promoteStrings.sql.out index a97abd5dd181..6fb21f001447 100644 --- a/sql/core/src/test/resources/sql-tests/results/typeCoercion/native/promoteStrings.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/typeCoercion/native/promoteStrings.sql.out @@ -323,6 +323,7 @@ org.apache.spark.SparkDateTimeException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'1'", "sourceType" : "\"STRING\"", "targetType" : "\"TIMESTAMP\"" @@ -347,6 +348,7 @@ org.apache.spark.SparkDateTimeException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'1'", "sourceType" : "\"STRING\"", "targetType" : "\"DATE\"" @@ -1333,6 +1335,7 @@ org.apache.spark.SparkDateTimeException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'1'", "sourceType" : "\"STRING\"", "targetType" : "\"TIMESTAMP\"" @@ -1357,6 +1360,7 @@ org.apache.spark.SparkDateTimeException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'1'", "sourceType" : "\"STRING\"", "targetType" : "\"DATE\"" @@ -2069,6 +2073,7 @@ org.apache.spark.SparkDateTimeException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'1'", "sourceType" : "\"STRING\"", "targetType" : "\"TIMESTAMP\"" @@ -2093,6 +2098,7 @@ org.apache.spark.SparkDateTimeException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'1'", "sourceType" : "\"STRING\"", "targetType" : "\"DATE\"" @@ -2189,6 +2195,7 @@ org.apache.spark.SparkDateTimeException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'1'", "sourceType" : "\"STRING\"", "targetType" : "\"TIMESTAMP\"" @@ -2213,6 +2220,7 @@ org.apache.spark.SparkDateTimeException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'1'", "sourceType" : "\"STRING\"", "targetType" : "\"DATE\"" @@ -2317,6 +2325,7 @@ org.apache.spark.SparkDateTimeException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'1'", "sourceType" : "\"STRING\"", "targetType" : "\"TIMESTAMP\"" @@ -2341,6 +2350,7 @@ org.apache.spark.SparkDateTimeException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'1'", "sourceType" : "\"STRING\"", "targetType" : "\"DATE\"" @@ -2437,6 +2447,7 @@ org.apache.spark.SparkDateTimeException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'1'", "sourceType" : "\"STRING\"", "targetType" : "\"TIMESTAMP\"" @@ -2461,6 +2472,7 @@ org.apache.spark.SparkDateTimeException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'1'", "sourceType" : "\"STRING\"", "targetType" : "\"DATE\"" @@ -2565,6 +2577,7 @@ org.apache.spark.SparkDateTimeException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'1'", "sourceType" : "\"STRING\"", "targetType" : "\"TIMESTAMP\"" @@ -2589,6 +2602,7 @@ org.apache.spark.SparkDateTimeException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'1'", "sourceType" : "\"STRING\"", "targetType" : "\"DATE\"" @@ -2693,6 +2707,7 @@ org.apache.spark.SparkDateTimeException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'1'", "sourceType" : "\"STRING\"", "targetType" : "\"TIMESTAMP\"" @@ -2717,6 +2732,7 @@ org.apache.spark.SparkDateTimeException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'1'", "sourceType" : "\"STRING\"", "targetType" : "\"DATE\"" @@ -2821,6 +2837,7 @@ org.apache.spark.SparkDateTimeException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'1'", "sourceType" : "\"STRING\"", "targetType" : "\"TIMESTAMP\"" @@ -2845,6 +2862,7 @@ org.apache.spark.SparkDateTimeException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'1'", "sourceType" : "\"STRING\"", "targetType" : "\"DATE\"" @@ -2949,6 +2967,7 @@ org.apache.spark.SparkDateTimeException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'1'", "sourceType" : "\"STRING\"", "targetType" : "\"TIMESTAMP\"" @@ -2973,6 +2992,7 @@ org.apache.spark.SparkDateTimeException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'1'", "sourceType" : "\"STRING\"", "targetType" : "\"DATE\"" @@ -3077,6 +3097,7 @@ org.apache.spark.SparkDateTimeException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'1'", "sourceType" : "\"STRING\"", "targetType" : "\"TIMESTAMP\"" @@ -3101,6 +3122,7 @@ org.apache.spark.SparkDateTimeException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'1'", "sourceType" : "\"STRING\"", "targetType" : "\"DATE\"" @@ -3205,6 +3227,7 @@ org.apache.spark.SparkDateTimeException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'1'", "sourceType" : "\"STRING\"", "targetType" : "\"TIMESTAMP\"" @@ -3229,6 +3252,7 @@ org.apache.spark.SparkDateTimeException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'1'", "sourceType" : "\"STRING\"", "targetType" : "\"DATE\"" @@ -3333,6 +3357,7 @@ org.apache.spark.SparkDateTimeException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'1'", "sourceType" : "\"STRING\"", "targetType" : "\"TIMESTAMP\"" @@ -3357,6 +3382,7 @@ org.apache.spark.SparkDateTimeException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'1'", "sourceType" : "\"STRING\"", "targetType" : "\"DATE\"" @@ -3461,6 +3487,7 @@ org.apache.spark.SparkDateTimeException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'1'", "sourceType" : "\"STRING\"", "targetType" : "\"TIMESTAMP\"" @@ -3485,6 +3512,7 @@ org.apache.spark.SparkDateTimeException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'1'", "sourceType" : "\"STRING\"", "targetType" : "\"DATE\"" @@ -3589,6 +3617,7 @@ org.apache.spark.SparkDateTimeException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'1'", "sourceType" : "\"STRING\"", "targetType" : "\"TIMESTAMP\"" @@ -3613,6 +3642,7 @@ org.apache.spark.SparkDateTimeException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'1'", "sourceType" : "\"STRING\"", "targetType" : "\"DATE\"" @@ -3717,6 +3747,7 @@ org.apache.spark.SparkDateTimeException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'1'", "sourceType" : "\"STRING\"", "targetType" : "\"TIMESTAMP\"" @@ -3741,6 +3772,7 @@ org.apache.spark.SparkDateTimeException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'1'", "sourceType" : "\"STRING\"", "targetType" : "\"DATE\"" diff --git a/sql/core/src/test/resources/sql-tests/results/typeCoercion/native/stringCastAndExpressions.sql.out b/sql/core/src/test/resources/sql-tests/results/typeCoercion/native/stringCastAndExpressions.sql.out index 3d42e866ea17..98b468621503 100644 --- a/sql/core/src/test/resources/sql-tests/results/typeCoercion/native/stringCastAndExpressions.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/typeCoercion/native/stringCastAndExpressions.sql.out @@ -17,6 +17,7 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'aa'", "sourceType" : "\"STRING\"", "targetType" : "\"TINYINT\"" @@ -41,6 +42,7 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'aa'", "sourceType" : "\"STRING\"", "targetType" : "\"SMALLINT\"" @@ -65,6 +67,7 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'aa'", "sourceType" : "\"STRING\"", "targetType" : "\"INT\"" @@ -89,6 +92,7 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'aa'", "sourceType" : "\"STRING\"", "targetType" : "\"BIGINT\"" @@ -113,6 +117,7 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'aa'", "sourceType" : "\"STRING\"", "targetType" : "\"FLOAT\"" @@ -137,6 +142,7 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'aa'", "sourceType" : "\"STRING\"", "targetType" : "\"DOUBLE\"" @@ -161,6 +167,7 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'aa'", "sourceType" : "\"STRING\"", "targetType" : "\"DECIMAL(10,0)\"" @@ -185,6 +192,7 @@ org.apache.spark.SparkRuntimeException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'aa'", "sourceType" : "\"STRING\"", "targetType" : "\"BOOLEAN\"" @@ -209,6 +217,7 @@ org.apache.spark.SparkDateTimeException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'aa'", "sourceType" : "\"STRING\"", "targetType" : "\"TIMESTAMP\"" @@ -233,6 +242,7 @@ org.apache.spark.SparkDateTimeException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'aa'", "sourceType" : "\"STRING\"", "targetType" : "\"DATE\"" @@ -337,6 +347,7 @@ org.apache.spark.SparkDateTimeException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'aa'", "sourceType" : "\"STRING\"", "targetType" : "\"TIMESTAMP\"" @@ -444,6 +455,7 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'aa'", "sourceType" : "\"STRING\"", "targetType" : "\"BIGINT\"" @@ -468,6 +480,7 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'2018-01-01'", "sourceType" : "\"STRING\"", "targetType" : "\"BIGINT\"" @@ -492,6 +505,7 @@ org.apache.spark.SparkDateTimeException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'aa'", "sourceType" : "\"STRING\"", "targetType" : "\"DATE\"" @@ -531,6 +545,7 @@ org.apache.spark.SparkDateTimeException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'aa'", "sourceType" : "\"STRING\"", "targetType" : "\"DATE\"" @@ -571,6 +586,7 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'aa'", "sourceType" : "\"STRING\"", "targetType" : "\"INT\"" diff --git a/sql/core/src/test/resources/sql-tests/results/typeCoercion/native/widenSetOperationTypes.sql.out b/sql/core/src/test/resources/sql-tests/results/typeCoercion/native/widenSetOperationTypes.sql.out index 678d00a39b2d..4b5d8e71c1da 100644 --- a/sql/core/src/test/resources/sql-tests/results/typeCoercion/native/widenSetOperationTypes.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/typeCoercion/native/widenSetOperationTypes.sql.out @@ -1366,6 +1366,7 @@ org.apache.spark.SparkDateTimeException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'1'", "sourceType" : "\"STRING\"", "targetType" : "\"TIMESTAMP\"" @@ -1390,6 +1391,7 @@ org.apache.spark.SparkDateTimeException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'1'", "sourceType" : "\"STRING\"", "targetType" : "\"DATE\"" @@ -1891,6 +1893,7 @@ org.apache.spark.SparkRuntimeException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'2'", "sourceType" : "\"STRING\"", "targetType" : "\"BOOLEAN\"" @@ -2193,6 +2196,7 @@ org.apache.spark.SparkDateTimeException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'2'", "sourceType" : "\"STRING\"", "targetType" : "\"TIMESTAMP\"" @@ -2478,6 +2482,7 @@ org.apache.spark.SparkDateTimeException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'2'", "sourceType" : "\"STRING\"", "targetType" : "\"DATE\"" diff --git a/sql/core/src/test/resources/sql-tests/results/udf/udf-union.sql.out b/sql/core/src/test/resources/sql-tests/results/udf/udf-union.sql.out index c61a6f841c88..6aa2148fa1e6 100644 --- a/sql/core/src/test/resources/sql-tests/results/udf/udf-union.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/udf/udf-union.sql.out @@ -44,6 +44,7 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'a'", "sourceType" : "\"STRING\"", "targetType" : "\"BIGINT\"" @@ -124,6 +125,7 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'str'", "sourceType" : "\"STRING\"", "targetType" : "\"BIGINT\"" @@ -150,6 +152,7 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'str'", "sourceType" : "\"STRING\"", "targetType" : "\"BIGINT\"" diff --git a/sql/core/src/test/resources/sql-tests/results/union.sql.out b/sql/core/src/test/resources/sql-tests/results/union.sql.out index 84e21a535f35..ce4d3421ae83 100644 --- a/sql/core/src/test/resources/sql-tests/results/union.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/union.sql.out @@ -44,6 +44,7 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'a'", "sourceType" : "\"STRING\"", "targetType" : "\"BIGINT\"" @@ -124,6 +125,7 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'str'", "sourceType" : "\"STRING\"", "targetType" : "\"BIGINT\"" @@ -150,6 +152,7 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'str'", "sourceType" : "\"STRING\"", "targetType" : "\"BIGINT\"" diff --git a/sql/core/src/test/resources/sql-tests/results/view-schema-binding-config.sql.out b/sql/core/src/test/resources/sql-tests/results/view-schema-binding-config.sql.out index 80260f8fcaa7..270f15e0b493 100644 --- a/sql/core/src/test/resources/sql-tests/results/view-schema-binding-config.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/view-schema-binding-config.sql.out @@ -701,6 +701,7 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'a'", "sourceType" : "\"STRING\"", "targetType" : "\"INT\"" diff --git a/sql/core/src/test/resources/sql-tests/results/view-schema-compensation.sql.out b/sql/core/src/test/resources/sql-tests/results/view-schema-compensation.sql.out index 94948d1a9a24..442d1f28744d 100644 --- a/sql/core/src/test/resources/sql-tests/results/view-schema-compensation.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/view-schema-compensation.sql.out @@ -187,6 +187,7 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'a'", "sourceType" : "\"STRING\"", "targetType" : "\"INT\"" diff --git a/sql/core/src/test/scala/org/apache/spark/sql/SQLInsertTestSuite.scala b/sql/core/src/test/scala/org/apache/spark/sql/SQLInsertTestSuite.scala index 13eb2dba6d95..c9feedc9645d 100644 --- a/sql/core/src/test/scala/org/apache/spark/sql/SQLInsertTestSuite.scala +++ b/sql/core/src/test/scala/org/apache/spark/sql/SQLInsertTestSuite.scala @@ -446,7 +446,8 @@ trait SQLInsertTestSuite extends QueryTest with SQLTestUtils with AdaptiveSparkP parameters = Map( "expression" -> "'ansi'", "sourceType" -> "\"STRING\"", - "targetType" -> "\"INT\"" + "targetType" -> "\"INT\"", + "ansiConfig" -> "\"spark.sql.ansi.enabled\"" ), context = ExpectedContext("insert into t partition(a='ansi')", 0, 32) ) diff --git a/sql/core/src/test/scala/org/apache/spark/sql/connector/ProcedureSuite.scala b/sql/core/src/test/scala/org/apache/spark/sql/connector/ProcedureSuite.scala index ba63b560309e..43bec881073a 100644 --- a/sql/core/src/test/scala/org/apache/spark/sql/connector/ProcedureSuite.scala +++ b/sql/core/src/test/scala/org/apache/spark/sql/connector/ProcedureSuite.scala @@ -282,7 +282,8 @@ class ProcedureSuite extends QueryTest with SharedSparkSession with BeforeAndAft parameters = Map( "expression" -> toSQLValue("A"), "sourceType" -> toSQLType("STRING"), - "targetType" -> toSQLType("INT")), + "targetType" -> toSQLType("INT"), + "ansiConfig" -> f"\"${SQLConf.ANSI_ENABLED.key}\""), context = ExpectedContext(fragment = call, start = 0, stop = call.length - 1)) } } diff --git a/sql/core/src/test/scala/org/apache/spark/sql/errors/QueryExecutionAnsiErrorsSuite.scala b/sql/core/src/test/scala/org/apache/spark/sql/errors/QueryExecutionAnsiErrorsSuite.scala index 753f612238e9..267b633c7676 100644 --- a/sql/core/src/test/scala/org/apache/spark/sql/errors/QueryExecutionAnsiErrorsSuite.scala +++ b/sql/core/src/test/scala/org/apache/spark/sql/errors/QueryExecutionAnsiErrorsSuite.scala @@ -49,7 +49,8 @@ class QueryExecutionAnsiErrorsSuite extends QueryTest condition = "CAST_OVERFLOW", parameters = Map("value" -> "TIMESTAMP '9999-12-31 04:13:14.56789'", "sourceType" -> "\"TIMESTAMP\"", - "targetType" -> "\"INT\""), + "targetType" -> "\"INT\"", + "ansiConfig" -> ansiConf), sqlState = "22003") } @@ -214,7 +215,8 @@ class QueryExecutionAnsiErrorsSuite extends QueryTest parameters = Map( "expression" -> "'111111111111xe23'", "sourceType" -> "\"STRING\"", - "targetType" -> "\"DOUBLE\""), + "targetType" -> "\"DOUBLE\"", + "ansiConfig" -> ansiConf), context = ExpectedContext( fragment = "CAST('111111111111xe23' AS DOUBLE)", start = 7, @@ -228,7 +230,8 @@ class QueryExecutionAnsiErrorsSuite extends QueryTest parameters = Map( "expression" -> "'111111111111xe23'", "sourceType" -> "\"STRING\"", - "targetType" -> "\"DOUBLE\""), + "targetType" -> "\"DOUBLE\"", + "ansiConfig" -> ansiConf), context = ExpectedContext( fragment = "cast", callSitePattern = getCurrentClassCallSitePattern)) @@ -275,7 +278,8 @@ class QueryExecutionAnsiErrorsSuite extends QueryTest condition = "CAST_OVERFLOW", parameters = Map("value" -> "1.2345678901234567E19D", "sourceType" -> "\"DOUBLE\"", - "targetType" -> ("\"TINYINT\"")) + "targetType" -> ("\"TINYINT\""), + "ansiConfig" -> ansiConf) ) } @@ -293,7 +297,8 @@ class QueryExecutionAnsiErrorsSuite extends QueryTest condition = "CAST_OVERFLOW", parameters = Map("value" -> "-1.2345678901234567E19D", "sourceType" -> "\"DOUBLE\"", - "targetType" -> "\"TINYINT\""), + "targetType" -> "\"TINYINT\"", + "ansiConfig" -> ansiConf), sqlState = "22003") } } diff --git a/sql/core/src/test/scala/org/apache/spark/sql/errors/QueryExecutionErrorsSuite.scala b/sql/core/src/test/scala/org/apache/spark/sql/errors/QueryExecutionErrorsSuite.scala index 17c3c1e1e2a7..46f2502582c3 100644 --- a/sql/core/src/test/scala/org/apache/spark/sql/errors/QueryExecutionErrorsSuite.scala +++ b/sql/core/src/test/scala/org/apache/spark/sql/errors/QueryExecutionErrorsSuite.scala @@ -735,7 +735,8 @@ class QueryExecutionErrorsSuite parameters = Map( "value" -> sourceValue, "sourceType" -> s""""${sourceType.sql}"""", - "targetType" -> s""""$it""""), + "targetType" -> s""""$it"""", + "ansiConfig" -> s""""${SQLConf.ANSI_ENABLED.key}""""), sqlState = "22003") } } diff --git a/sql/core/src/test/scala/org/apache/spark/sql/execution/command/AlterTableAddPartitionSuiteBase.scala b/sql/core/src/test/scala/org/apache/spark/sql/execution/command/AlterTableAddPartitionSuiteBase.scala index 13ea6f5a3053..e75a1b1ff0f9 100644 --- a/sql/core/src/test/scala/org/apache/spark/sql/execution/command/AlterTableAddPartitionSuiteBase.scala +++ b/sql/core/src/test/scala/org/apache/spark/sql/execution/command/AlterTableAddPartitionSuiteBase.scala @@ -249,7 +249,8 @@ trait AlterTableAddPartitionSuiteBase extends QueryTest with DDLCommandTestUtils parameters = Map( "expression" -> "'aaa'", "sourceType" -> "\"STRING\"", - "targetType" -> "\"INT\""), + "targetType" -> "\"INT\"", + "ansiConfig" -> "\"spark.sql.ansi.enabled\""), context = ExpectedContext( fragment = s"ALTER TABLE $t ADD PARTITION (p='aaa')", start = 0, diff --git a/sql/core/src/test/scala/org/apache/spark/sql/execution/command/v2/AlterTableAddPartitionSuite.scala b/sql/core/src/test/scala/org/apache/spark/sql/execution/command/v2/AlterTableAddPartitionSuite.scala index e7a2b194fa0b..cf3415f3d978 100644 --- a/sql/core/src/test/scala/org/apache/spark/sql/execution/command/v2/AlterTableAddPartitionSuite.scala +++ b/sql/core/src/test/scala/org/apache/spark/sql/execution/command/v2/AlterTableAddPartitionSuite.scala @@ -149,6 +149,7 @@ class AlterTableAddPartitionSuite }, condition = "CAST_INVALID_INPUT", parameters = Map( + "ansiConfig" -> "\"spark.sql.ansi.enabled\"", "expression" -> "'aaa'", "sourceType" -> "\"STRING\"", "targetType" -> "\"INT\""), diff --git a/sql/core/src/test/scala/org/apache/spark/sql/scripting/SqlScriptingInterpreterSuite.scala b/sql/core/src/test/scala/org/apache/spark/sql/scripting/SqlScriptingInterpreterSuite.scala index 7a1b6b1c5e46..885c9e51ffac 100644 --- a/sql/core/src/test/scala/org/apache/spark/sql/scripting/SqlScriptingInterpreterSuite.scala +++ b/sql/core/src/test/scala/org/apache/spark/sql/scripting/SqlScriptingInterpreterSuite.scala @@ -875,7 +875,8 @@ class SqlScriptingInterpreterSuite extends QueryTest with SharedSparkSession { parameters = Map( "expression" -> "'one'", "sourceType" -> "\"STRING\"", - "targetType" -> "\"BIGINT\""), + "targetType" -> "\"BIGINT\"", + "ansiConfig" -> f"\"${SQLConf.ANSI_ENABLED.key}\""), context = ExpectedContext(fragment = "", start = -1, stop = -1)) } withSQLConf(SQLConf.ANSI_ENABLED.key -> "false") {