From 68dbc238cb780ec594f233cc95655242841314b3 Mon Sep 17 00:00:00 2001 From: Max Gekk Date: Wed, 14 Sep 2022 08:38:47 +0300 Subject: [PATCH 1/3] Sort message parameters --- core/src/main/scala/org/apache/spark/SparkThrowableHelper.scala | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/core/src/main/scala/org/apache/spark/SparkThrowableHelper.scala b/core/src/main/scala/org/apache/spark/SparkThrowableHelper.scala index 93e88261cb672..81c5c6cb043b3 100644 --- a/core/src/main/scala/org/apache/spark/SparkThrowableHelper.scala +++ b/core/src/main/scala/org/apache/spark/SparkThrowableHelper.scala @@ -203,7 +203,7 @@ private[spark] object SparkThrowableHelper { val parameterNames = e.getParameterNames if (!parameterNames.isEmpty) { g.writeObjectFieldStart("messageParameters") - (parameterNames zip e.getMessageParameters).foreach { case (name, value) => + (parameterNames zip e.getMessageParameters).sortBy(_._1).foreach { case (name, value) => g.writeStringField(name, value) } g.writeEndObject() From 980cd43ed1f44e405a7d8d4109cf41c2b7c25b1d Mon Sep 17 00:00:00 2001 From: Max Gekk Date: Wed, 14 Sep 2022 08:55:06 +0300 Subject: [PATCH 2/3] Re-gen sql.out --- .../sql-tests/results/ansi/array.sql.out | 28 +- .../sql-tests/results/ansi/cast.sql.out | 164 +-- .../sql-tests/results/ansi/date.sql.out | 28 +- .../ansi/datetime-parsing-invalid.sql.out | 108 +- .../ansi/decimalArithmeticOperations.sql.out | 40 +- .../sql-tests/results/ansi/interval.sql.out | 184 +-- .../results/ansi/string-functions.sql.out | 82 +- .../sql-tests/results/ansi/timestamp.sql.out | 64 +- .../results/ansi/try_aggregates.sql.out | 8 +- .../results/ansi/try_arithmetic.sql.out | 32 +- .../ansi/try_datetime_functions.sql.out | 4 +- .../resources/sql-tests/results/cast.sql.out | 24 +- .../sql-tests/results/change-column.sql.out | 12 +- .../resources/sql-tests/results/date.sql.out | 12 +- .../datetime-formatting-invalid.sql.out | 88 +- .../sql-tests/results/datetime-legacy.sql.out | 16 +- .../results/datetime-parsing-invalid.sql.out | 32 +- .../sql-tests/results/describe.sql.out | 8 +- .../results/group-by-ordinal.sql.out | 16 +- .../sql-tests/results/having.sql.out | 4 +- .../sql-tests/results/interval.sql.out | 140 +- .../sql-tests/results/json-functions.sql.out | 8 +- .../resources/sql-tests/results/pivot.sql.out | 4 +- .../results/postgreSQL/boolean.sql.out | 60 +- .../results/postgreSQL/float4.sql.out | 28 +- .../results/postgreSQL/float8.sql.out | 20 +- .../sql-tests/results/postgreSQL/int4.sql.out | 24 +- .../sql-tests/results/postgreSQL/int8.sql.out | 32 +- .../sql-tests/results/postgreSQL/text.sql.out | 12 +- .../results/postgreSQL/window_part2.sql.out | 12 +- .../results/regexp-functions.sql.out | 12 +- .../results/string-functions.sql.out | 66 +- .../sql-tests/results/timestamp.sql.out | 40 +- .../timestampNTZ/timestamp-ansi.sql.out | 48 +- .../results/timestampNTZ/timestamp.sql.out | 20 +- .../results/try_datetime_functions.sql.out | 4 +- .../native/booleanEquality.sql.out | 96 +- .../native/decimalPrecision.sql.out | 1312 ++++++++--------- .../typeCoercion/native/division.sql.out | 320 ++-- .../native/promoteStrings.sql.out | 168 +-- .../native/stringCastAndExpressions.sql.out | 12 +- .../sql-tests/results/udf/udf-pivot.sql.out | 4 +- 42 files changed, 1698 insertions(+), 1698 deletions(-) diff --git a/sql/core/src/test/resources/sql-tests/results/ansi/array.sql.out b/sql/core/src/test/resources/sql-tests/results/ansi/array.sql.out index b0b220bba5960..f11f987423660 100644 --- a/sql/core/src/test/resources/sql-tests/results/ansi/array.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/ansi/array.sql.out @@ -168,9 +168,9 @@ org.apache.spark.SparkArrayIndexOutOfBoundsException { "errorClass" : "INVALID_ARRAY_INDEX_IN_ELEMENT_AT", "messageParameters" : { - "indexValue" : "5", + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "arraySize" : "3", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "indexValue" : "5" }, "queryContext" : [ { "objectType" : "", @@ -191,9 +191,9 @@ org.apache.spark.SparkArrayIndexOutOfBoundsException { "errorClass" : "INVALID_ARRAY_INDEX_IN_ELEMENT_AT", "messageParameters" : { - "indexValue" : "-5", + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "arraySize" : "3", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "indexValue" : "-5" }, "queryContext" : [ { "objectType" : "", @@ -232,9 +232,9 @@ org.apache.spark.SparkArrayIndexOutOfBoundsException { "errorClass" : "INVALID_ARRAY_INDEX", "messageParameters" : { - "indexValue" : "4", + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "arraySize" : "2", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "indexValue" : "4" }, "queryContext" : [ { "objectType" : "", @@ -255,9 +255,9 @@ org.apache.spark.SparkArrayIndexOutOfBoundsException { "errorClass" : "INVALID_ARRAY_INDEX", "messageParameters" : { - "indexValue" : "0", + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "arraySize" : "2", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "indexValue" : "0" }, "queryContext" : [ { "objectType" : "", @@ -278,9 +278,9 @@ org.apache.spark.SparkArrayIndexOutOfBoundsException { "errorClass" : "INVALID_ARRAY_INDEX", "messageParameters" : { - "indexValue" : "-1", + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "arraySize" : "2", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "indexValue" : "-1" }, "queryContext" : [ { "objectType" : "", @@ -333,9 +333,9 @@ org.apache.spark.SparkArrayIndexOutOfBoundsException { "errorClass" : "INVALID_ARRAY_INDEX", "messageParameters" : { - "indexValue" : "5", + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "arraySize" : "3", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "indexValue" : "5" }, "queryContext" : [ { "objectType" : "", @@ -356,9 +356,9 @@ org.apache.spark.SparkArrayIndexOutOfBoundsException { "errorClass" : "INVALID_ARRAY_INDEX", "messageParameters" : { - "indexValue" : "-1", + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "arraySize" : "3", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "indexValue" : "-1" }, "queryContext" : [ { "objectType" : "", diff --git a/sql/core/src/test/resources/sql-tests/results/ansi/cast.sql.out b/sql/core/src/test/resources/sql-tests/results/ansi/cast.sql.out index 378f817f7d9a6..deaece6e7e1ac 100644 --- a/sql/core/src/test/resources/sql-tests/results/ansi/cast.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/ansi/cast.sql.out @@ -9,10 +9,10 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "42000", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'1.23'", "sourceType" : "\"STRING\"", - "targetType" : "\"INT\"", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "targetType" : "\"INT\"" }, "queryContext" : [ { "objectType" : "", @@ -34,10 +34,10 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "42000", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'1.23'", "sourceType" : "\"STRING\"", - "targetType" : "\"BIGINT\"", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "targetType" : "\"BIGINT\"" }, "queryContext" : [ { "objectType" : "", @@ -59,10 +59,10 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "42000", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'-4.56'", "sourceType" : "\"STRING\"", - "targetType" : "\"INT\"", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "targetType" : "\"INT\"" }, "queryContext" : [ { "objectType" : "", @@ -84,10 +84,10 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "42000", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'-4.56'", "sourceType" : "\"STRING\"", - "targetType" : "\"BIGINT\"", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "targetType" : "\"BIGINT\"" }, "queryContext" : [ { "objectType" : "", @@ -109,10 +109,10 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "42000", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'abc'", "sourceType" : "\"STRING\"", - "targetType" : "\"INT\"", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "targetType" : "\"INT\"" }, "queryContext" : [ { "objectType" : "", @@ -134,10 +134,10 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "42000", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'abc'", "sourceType" : "\"STRING\"", - "targetType" : "\"BIGINT\"", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "targetType" : "\"BIGINT\"" }, "queryContext" : [ { "objectType" : "", @@ -159,10 +159,10 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "42000", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'abc'", "sourceType" : "\"STRING\"", - "targetType" : "\"FLOAT\"", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "targetType" : "\"FLOAT\"" }, "queryContext" : [ { "objectType" : "", @@ -184,10 +184,10 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "42000", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'abc'", "sourceType" : "\"STRING\"", - "targetType" : "\"DOUBLE\"", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "targetType" : "\"DOUBLE\"" }, "queryContext" : [ { "objectType" : "", @@ -209,10 +209,10 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "42000", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'1234567890123'", "sourceType" : "\"STRING\"", - "targetType" : "\"INT\"", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "targetType" : "\"INT\"" }, "queryContext" : [ { "objectType" : "", @@ -234,10 +234,10 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "42000", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'12345678901234567890123'", "sourceType" : "\"STRING\"", - "targetType" : "\"BIGINT\"", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "targetType" : "\"BIGINT\"" }, "queryContext" : [ { "objectType" : "", @@ -259,10 +259,10 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "42000", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "''", "sourceType" : "\"STRING\"", - "targetType" : "\"INT\"", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "targetType" : "\"INT\"" }, "queryContext" : [ { "objectType" : "", @@ -284,10 +284,10 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "42000", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "''", "sourceType" : "\"STRING\"", - "targetType" : "\"BIGINT\"", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "targetType" : "\"BIGINT\"" }, "queryContext" : [ { "objectType" : "", @@ -309,10 +309,10 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "42000", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "''", "sourceType" : "\"STRING\"", - "targetType" : "\"FLOAT\"", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "targetType" : "\"FLOAT\"" }, "queryContext" : [ { "objectType" : "", @@ -334,10 +334,10 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "42000", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "''", "sourceType" : "\"STRING\"", - "targetType" : "\"DOUBLE\"", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "targetType" : "\"DOUBLE\"" }, "queryContext" : [ { "objectType" : "", @@ -375,10 +375,10 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "42000", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'123.a'", "sourceType" : "\"STRING\"", - "targetType" : "\"INT\"", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "targetType" : "\"INT\"" }, "queryContext" : [ { "objectType" : "", @@ -400,10 +400,10 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "42000", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'123.a'", "sourceType" : "\"STRING\"", - "targetType" : "\"BIGINT\"", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "targetType" : "\"BIGINT\"" }, "queryContext" : [ { "objectType" : "", @@ -425,10 +425,10 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "42000", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'123.a'", "sourceType" : "\"STRING\"", - "targetType" : "\"FLOAT\"", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "targetType" : "\"FLOAT\"" }, "queryContext" : [ { "objectType" : "", @@ -450,10 +450,10 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "42000", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'123.a'", "sourceType" : "\"STRING\"", - "targetType" : "\"DOUBLE\"", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "targetType" : "\"DOUBLE\"" }, "queryContext" : [ { "objectType" : "", @@ -483,10 +483,10 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "42000", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'-2147483649'", "sourceType" : "\"STRING\"", - "targetType" : "\"INT\"", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "targetType" : "\"INT\"" }, "queryContext" : [ { "objectType" : "", @@ -516,10 +516,10 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "42000", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'2147483648'", "sourceType" : "\"STRING\"", - "targetType" : "\"INT\"", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "targetType" : "\"INT\"" }, "queryContext" : [ { "objectType" : "", @@ -549,10 +549,10 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "42000", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'-9223372036854775809'", "sourceType" : "\"STRING\"", - "targetType" : "\"BIGINT\"", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "targetType" : "\"BIGINT\"" }, "queryContext" : [ { "objectType" : "", @@ -582,10 +582,10 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "42000", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'9223372036854775808'", "sourceType" : "\"STRING\"", - "targetType" : "\"BIGINT\"", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "targetType" : "\"BIGINT\"" }, "queryContext" : [ { "objectType" : "", @@ -854,10 +854,10 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "42000", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'1中文'", "sourceType" : "\"STRING\"", - "targetType" : "\"TINYINT\"", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "targetType" : "\"TINYINT\"" }, "queryContext" : [ { "objectType" : "", @@ -879,10 +879,10 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "42000", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'1中文'", "sourceType" : "\"STRING\"", - "targetType" : "\"SMALLINT\"", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "targetType" : "\"SMALLINT\"" }, "queryContext" : [ { "objectType" : "", @@ -904,10 +904,10 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "42000", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'1中文'", "sourceType" : "\"STRING\"", - "targetType" : "\"INT\"", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "targetType" : "\"INT\"" }, "queryContext" : [ { "objectType" : "", @@ -929,10 +929,10 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "42000", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'中文1'", "sourceType" : "\"STRING\"", - "targetType" : "\"BIGINT\"", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "targetType" : "\"BIGINT\"" }, "queryContext" : [ { "objectType" : "", @@ -954,10 +954,10 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "42000", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'1中文'", "sourceType" : "\"STRING\"", - "targetType" : "\"BIGINT\"", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "targetType" : "\"BIGINT\"" }, "queryContext" : [ { "objectType" : "", @@ -997,10 +997,10 @@ org.apache.spark.SparkRuntimeException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "42000", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'\t\n xyz \t\r'", "sourceType" : "\"STRING\"", - "targetType" : "\"BOOLEAN\"", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "targetType" : "\"BOOLEAN\"" }, "queryContext" : [ { "objectType" : "", @@ -1030,10 +1030,10 @@ org.apache.spark.SparkArithmeticException "errorClass" : "NUMERIC_VALUE_OUT_OF_RANGE", "sqlState" : "22005", "messageParameters" : { - "value" : "123.45", + "config" : "\"spark.sql.ansi.enabled\"", "precision" : "4", "scale" : "2", - "config" : "\"spark.sql.ansi.enabled\"" + "value" : "123.45" }, "queryContext" : [ { "objectType" : "", @@ -1055,10 +1055,10 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "42000", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'xyz'", "sourceType" : "\"STRING\"", - "targetType" : "\"DECIMAL(4,2)\"", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "targetType" : "\"DECIMAL(4,2)\"" }, "queryContext" : [ { "objectType" : "", @@ -1088,10 +1088,10 @@ org.apache.spark.SparkDateTimeException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "42000", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'a'", "sourceType" : "\"STRING\"", - "targetType" : "\"DATE\"", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "targetType" : "\"DATE\"" }, "queryContext" : [ { "objectType" : "", @@ -1121,10 +1121,10 @@ org.apache.spark.SparkDateTimeException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "42000", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'a'", "sourceType" : "\"STRING\"", - "targetType" : "\"TIMESTAMP\"", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "targetType" : "\"TIMESTAMP\"" }, "queryContext" : [ { "objectType" : "", @@ -1154,10 +1154,10 @@ org.apache.spark.SparkDateTimeException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "42000", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'a'", "sourceType" : "\"STRING\"", - "targetType" : "\"TIMESTAMP_NTZ\"", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "targetType" : "\"TIMESTAMP_NTZ\"" }, "queryContext" : [ { "objectType" : "", @@ -1179,10 +1179,10 @@ org.apache.spark.SparkDateTimeException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "42000", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "Infinity", "sourceType" : "\"DOUBLE\"", - "targetType" : "\"TIMESTAMP\"", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "targetType" : "\"TIMESTAMP\"" }, "queryContext" : [ { "objectType" : "", @@ -1204,10 +1204,10 @@ org.apache.spark.SparkDateTimeException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "42000", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "Infinity", "sourceType" : "\"DOUBLE\"", - "targetType" : "\"TIMESTAMP\"", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "targetType" : "\"TIMESTAMP\"" }, "queryContext" : [ { "objectType" : "", @@ -1261,10 +1261,10 @@ org.apache.spark.SparkArithmeticException "errorClass" : "CAST_OVERFLOW", "sqlState" : "22005", "messageParameters" : { - "value" : "INTERVAL '23:59:59' HOUR TO SECOND", + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "sourceType" : "\"INTERVAL HOUR TO SECOND\"", "targetType" : "\"SMALLINT\"", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "value" : "INTERVAL '23:59:59' HOUR TO SECOND" } } @@ -1295,10 +1295,10 @@ org.apache.spark.SparkArithmeticException "errorClass" : "CAST_OVERFLOW", "sqlState" : "22005", "messageParameters" : { - "value" : "INTERVAL '-1000' MONTH", + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "sourceType" : "\"INTERVAL MONTH\"", "targetType" : "\"TINYINT\"", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "value" : "INTERVAL '-1000' MONTH" } } @@ -1313,10 +1313,10 @@ org.apache.spark.SparkArithmeticException "errorClass" : "CAST_OVERFLOW", "sqlState" : "22005", "messageParameters" : { - "value" : "INTERVAL '1000000' SECOND", + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "sourceType" : "\"INTERVAL SECOND\"", "targetType" : "\"SMALLINT\"", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "value" : "INTERVAL '1000000' SECOND" } } @@ -1403,10 +1403,10 @@ org.apache.spark.SparkArithmeticException "errorClass" : "CAST_OVERFLOW", "sqlState" : "22005", "messageParameters" : { - "value" : "2147483647", + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "sourceType" : "\"INT\"", "targetType" : "\"INTERVAL YEAR\"", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "value" : "2147483647" } } @@ -1421,10 +1421,10 @@ org.apache.spark.SparkArithmeticException "errorClass" : "CAST_OVERFLOW", "sqlState" : "22005", "messageParameters" : { - "value" : "-9223372036854775808L", + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "sourceType" : "\"BIGINT\"", "targetType" : "\"INTERVAL DAY\"", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "value" : "-9223372036854775808L" } } @@ -1495,10 +1495,10 @@ org.apache.spark.SparkArithmeticException "errorClass" : "NUMERIC_VALUE_OUT_OF_RANGE", "sqlState" : "22005", "messageParameters" : { - "value" : "10.123000", + "config" : "\"spark.sql.ansi.enabled\"", "precision" : "1", "scale" : "0", - "config" : "\"spark.sql.ansi.enabled\"" + "value" : "10.123000" }, "queryContext" : [ { "objectType" : "", diff --git a/sql/core/src/test/resources/sql-tests/results/ansi/date.sql.out b/sql/core/src/test/resources/sql-tests/results/ansi/date.sql.out index 83ff07bc43fac..73b9768b5fcb2 100644 --- a/sql/core/src/test/resources/sql-tests/results/ansi/date.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/ansi/date.sql.out @@ -141,8 +141,8 @@ org.apache.spark.SparkDateTimeException "errorClass" : "CANNOT_PARSE_TIMESTAMP", "sqlState" : "42000", "messageParameters" : { - "message" : "Invalid date 'February 29' as '1970' is not a leap year", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "ansiConfig" : "\"spark.sql.ansi.enabled\"", + "message" : "Invalid date 'February 29' as '1970' is not a leap year" } } @@ -240,10 +240,10 @@ org.apache.spark.SparkDateTimeException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "42000", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'xx'", "sourceType" : "\"STRING\"", - "targetType" : "\"DATE\"", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "targetType" : "\"DATE\"" }, "queryContext" : [ { "objectType" : "", @@ -348,10 +348,10 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "42000", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'1.2'", "sourceType" : "\"STRING\"", - "targetType" : "\"INT\"", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "targetType" : "\"INT\"" }, "queryContext" : [ { "objectType" : "", @@ -472,10 +472,10 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "42000", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'1.2'", "sourceType" : "\"STRING\"", - "targetType" : "\"INT\"", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "targetType" : "\"INT\"" }, "queryContext" : [ { "objectType" : "", @@ -697,8 +697,8 @@ org.apache.spark.SparkUpgradeException "errorClass" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION", "errorSubClass" : "DATETIME_PATTERN_RECOGNITION", "messageParameters" : { - "pattern" : "'dd/MMMMM/yyyy'", - "config" : "\"spark.sql.legacy.timeParserPolicy\"" + "config" : "\"spark.sql.legacy.timeParserPolicy\"", + "pattern" : "'dd/MMMMM/yyyy'" } } @@ -713,8 +713,8 @@ org.apache.spark.SparkUpgradeException "errorClass" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION", "errorSubClass" : "DATETIME_PATTERN_RECOGNITION", "messageParameters" : { - "pattern" : "'dd/MMMMM/yyyy'", - "config" : "\"spark.sql.legacy.timeParserPolicy\"" + "config" : "\"spark.sql.legacy.timeParserPolicy\"", + "pattern" : "'dd/MMMMM/yyyy'" } } @@ -729,8 +729,8 @@ org.apache.spark.SparkUpgradeException "errorClass" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION", "errorSubClass" : "DATETIME_PATTERN_RECOGNITION", "messageParameters" : { - "pattern" : "'dd/MMMMM/yyyy'", - "config" : "\"spark.sql.legacy.timeParserPolicy\"" + "config" : "\"spark.sql.legacy.timeParserPolicy\"", + "pattern" : "'dd/MMMMM/yyyy'" } } diff --git a/sql/core/src/test/resources/sql-tests/results/ansi/datetime-parsing-invalid.sql.out b/sql/core/src/test/resources/sql-tests/results/ansi/datetime-parsing-invalid.sql.out index 507ce620b8995..464f055b35ca0 100644 --- a/sql/core/src/test/resources/sql-tests/results/ansi/datetime-parsing-invalid.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/ansi/datetime-parsing-invalid.sql.out @@ -18,8 +18,8 @@ org.apache.spark.SparkUpgradeException "errorClass" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION", "errorSubClass" : "PARSE_DATETIME_BY_NEW_PARSER", "messageParameters" : { - "datetime" : "'1'", - "config" : "\"spark.sql.legacy.timeParserPolicy\"" + "config" : "\"spark.sql.legacy.timeParserPolicy\"", + "datetime" : "'1'" } } @@ -34,8 +34,8 @@ org.apache.spark.SparkDateTimeException "errorClass" : "CANNOT_PARSE_TIMESTAMP", "sqlState" : "42000", "messageParameters" : { - "message" : "Text '-12' could not be parsed at index 0", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "ansiConfig" : "\"spark.sql.ansi.enabled\"", + "message" : "Text '-12' could not be parsed at index 0" } } @@ -50,8 +50,8 @@ org.apache.spark.SparkUpgradeException "errorClass" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION", "errorSubClass" : "PARSE_DATETIME_BY_NEW_PARSER", "messageParameters" : { - "datetime" : "'123'", - "config" : "\"spark.sql.legacy.timeParserPolicy\"" + "config" : "\"spark.sql.legacy.timeParserPolicy\"", + "datetime" : "'123'" } } @@ -66,8 +66,8 @@ org.apache.spark.SparkUpgradeException "errorClass" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION", "errorSubClass" : "PARSE_DATETIME_BY_NEW_PARSER", "messageParameters" : { - "datetime" : "'1'", - "config" : "\"spark.sql.legacy.timeParserPolicy\"" + "config" : "\"spark.sql.legacy.timeParserPolicy\"", + "datetime" : "'1'" } } @@ -82,8 +82,8 @@ org.apache.spark.SparkUpgradeException "errorClass" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION", "errorSubClass" : "DATETIME_PATTERN_RECOGNITION", "messageParameters" : { - "pattern" : "'yyyyyyy'", - "config" : "\"spark.sql.legacy.timeParserPolicy\"" + "config" : "\"spark.sql.legacy.timeParserPolicy\"", + "pattern" : "'yyyyyyy'" } } @@ -98,8 +98,8 @@ org.apache.spark.SparkDateTimeException "errorClass" : "CANNOT_PARSE_TIMESTAMP", "sqlState" : "42000", "messageParameters" : { - "message" : "Invalid date 'DayOfYear 366' as '1970' is not a leap year", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "ansiConfig" : "\"spark.sql.ansi.enabled\"", + "message" : "Invalid date 'DayOfYear 366' as '1970' is not a leap year" } } @@ -114,8 +114,8 @@ org.apache.spark.SparkUpgradeException "errorClass" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION", "errorSubClass" : "PARSE_DATETIME_BY_NEW_PARSER", "messageParameters" : { - "datetime" : "'9'", - "config" : "\"spark.sql.legacy.timeParserPolicy\"" + "config" : "\"spark.sql.legacy.timeParserPolicy\"", + "datetime" : "'9'" } } @@ -130,8 +130,8 @@ org.apache.spark.SparkUpgradeException "errorClass" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION", "errorSubClass" : "PARSE_DATETIME_BY_NEW_PARSER", "messageParameters" : { - "datetime" : "'9'", - "config" : "\"spark.sql.legacy.timeParserPolicy\"" + "config" : "\"spark.sql.legacy.timeParserPolicy\"", + "datetime" : "'9'" } } @@ -146,8 +146,8 @@ org.apache.spark.SparkUpgradeException "errorClass" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION", "errorSubClass" : "PARSE_DATETIME_BY_NEW_PARSER", "messageParameters" : { - "datetime" : "'99'", - "config" : "\"spark.sql.legacy.timeParserPolicy\"" + "config" : "\"spark.sql.legacy.timeParserPolicy\"", + "datetime" : "'99'" } } @@ -162,8 +162,8 @@ org.apache.spark.SparkDateTimeException "errorClass" : "CANNOT_PARSE_TIMESTAMP", "sqlState" : "42000", "messageParameters" : { - "message" : "Conflict found: Field DayOfMonth 30 differs from DayOfMonth 31 derived from 1970-12-31", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "ansiConfig" : "\"spark.sql.ansi.enabled\"", + "message" : "Conflict found: Field DayOfMonth 30 differs from DayOfMonth 31 derived from 1970-12-31" } } @@ -178,8 +178,8 @@ org.apache.spark.SparkDateTimeException "errorClass" : "CANNOT_PARSE_TIMESTAMP", "sqlState" : "42000", "messageParameters" : { - "message" : "Conflict found: Field MonthOfYear 11 differs from MonthOfYear 12 derived from 1970-12-31", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "ansiConfig" : "\"spark.sql.ansi.enabled\"", + "message" : "Conflict found: Field MonthOfYear 11 differs from MonthOfYear 12 derived from 1970-12-31" } } @@ -194,8 +194,8 @@ org.apache.spark.SparkDateTimeException "errorClass" : "CANNOT_PARSE_TIMESTAMP", "sqlState" : "42000", "messageParameters" : { - "message" : "Text '2019-366' could not be parsed: Invalid date 'DayOfYear 366' as '2019' is not a leap year", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "ansiConfig" : "\"spark.sql.ansi.enabled\"", + "message" : "Text '2019-366' could not be parsed: Invalid date 'DayOfYear 366' as '2019' is not a leap year" } } @@ -210,8 +210,8 @@ org.apache.spark.SparkDateTimeException "errorClass" : "CANNOT_PARSE_TIMESTAMP", "sqlState" : "42000", "messageParameters" : { - "message" : "Conflict found: Field DayOfMonth 30 differs from DayOfMonth 31 derived from 1970-12-31", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "ansiConfig" : "\"spark.sql.ansi.enabled\"", + "message" : "Conflict found: Field DayOfMonth 30 differs from DayOfMonth 31 derived from 1970-12-31" } } @@ -226,8 +226,8 @@ org.apache.spark.SparkDateTimeException "errorClass" : "CANNOT_PARSE_TIMESTAMP", "sqlState" : "42000", "messageParameters" : { - "message" : "Text '2020-01-365' could not be parsed: Conflict found: Field DayOfMonth 30 differs from DayOfMonth 1 derived from 2020-12-30", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "ansiConfig" : "\"spark.sql.ansi.enabled\"", + "message" : "Text '2020-01-365' could not be parsed: Conflict found: Field DayOfMonth 30 differs from DayOfMonth 1 derived from 2020-12-30" } } @@ -242,8 +242,8 @@ org.apache.spark.SparkDateTimeException "errorClass" : "CANNOT_PARSE_TIMESTAMP", "sqlState" : "42000", "messageParameters" : { - "message" : "Text '2020-10-350' could not be parsed: Conflict found: Field MonthOfYear 12 differs from MonthOfYear 10 derived from 2020-12-15", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "ansiConfig" : "\"spark.sql.ansi.enabled\"", + "message" : "Text '2020-10-350' could not be parsed: Conflict found: Field MonthOfYear 12 differs from MonthOfYear 10 derived from 2020-12-15" } } @@ -258,8 +258,8 @@ org.apache.spark.SparkDateTimeException "errorClass" : "CANNOT_PARSE_TIMESTAMP", "sqlState" : "42000", "messageParameters" : { - "message" : "Text '2020-11-31-366' could not be parsed: Invalid date 'NOVEMBER 31'", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "ansiConfig" : "\"spark.sql.ansi.enabled\"", + "message" : "Text '2020-11-31-366' could not be parsed: Invalid date 'NOVEMBER 31'" } } @@ -274,8 +274,8 @@ org.apache.spark.SparkUpgradeException "errorClass" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION", "errorSubClass" : "PARSE_DATETIME_BY_NEW_PARSER", "messageParameters" : { - "datetime" : "'2018-366'", - "config" : "\"spark.sql.legacy.timeParserPolicy\"" + "config" : "\"spark.sql.legacy.timeParserPolicy\"", + "datetime" : "'2018-366'" } } @@ -290,8 +290,8 @@ org.apache.spark.SparkDateTimeException "errorClass" : "CANNOT_PARSE_TIMESTAMP", "sqlState" : "42000", "messageParameters" : { - "message" : "Text '2020-01-27T20:06:11.847' could not be parsed at index 10", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "ansiConfig" : "\"spark.sql.ansi.enabled\"", + "message" : "Text '2020-01-27T20:06:11.847' could not be parsed at index 10" } } @@ -306,8 +306,8 @@ org.apache.spark.SparkDateTimeException "errorClass" : "CANNOT_PARSE_TIMESTAMP", "sqlState" : "42000", "messageParameters" : { - "message" : "Text 'Unparseable' could not be parsed at index 0", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "ansiConfig" : "\"spark.sql.ansi.enabled\"", + "message" : "Text 'Unparseable' could not be parsed at index 0" } } @@ -322,8 +322,8 @@ org.apache.spark.SparkDateTimeException "errorClass" : "CANNOT_PARSE_TIMESTAMP", "sqlState" : "42000", "messageParameters" : { - "message" : "Text '2020-01-27T20:06:11.847' could not be parsed at index 10", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "ansiConfig" : "\"spark.sql.ansi.enabled\"", + "message" : "Text '2020-01-27T20:06:11.847' could not be parsed at index 10" } } @@ -338,8 +338,8 @@ org.apache.spark.SparkDateTimeException "errorClass" : "CANNOT_PARSE_TIMESTAMP", "sqlState" : "42000", "messageParameters" : { - "message" : "Text 'Unparseable' could not be parsed at index 0", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "ansiConfig" : "\"spark.sql.ansi.enabled\"", + "message" : "Text 'Unparseable' could not be parsed at index 0" } } @@ -354,8 +354,8 @@ org.apache.spark.SparkDateTimeException "errorClass" : "CANNOT_PARSE_TIMESTAMP", "sqlState" : "42000", "messageParameters" : { - "message" : "Text '2020-01-27T20:06:11.847' could not be parsed at index 10", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "ansiConfig" : "\"spark.sql.ansi.enabled\"", + "message" : "Text '2020-01-27T20:06:11.847' could not be parsed at index 10" } } @@ -370,8 +370,8 @@ org.apache.spark.SparkDateTimeException "errorClass" : "CANNOT_PARSE_TIMESTAMP", "sqlState" : "42000", "messageParameters" : { - "message" : "Text 'Unparseable' could not be parsed at index 0", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "ansiConfig" : "\"spark.sql.ansi.enabled\"", + "message" : "Text 'Unparseable' could not be parsed at index 0" } } @@ -386,8 +386,8 @@ org.apache.spark.SparkDateTimeException "errorClass" : "CANNOT_PARSE_TIMESTAMP", "sqlState" : "42000", "messageParameters" : { - "message" : "Text '2020-01-27T20:06:11.847' could not be parsed at index 10", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "ansiConfig" : "\"spark.sql.ansi.enabled\"", + "message" : "Text '2020-01-27T20:06:11.847' could not be parsed at index 10" } } @@ -402,8 +402,8 @@ org.apache.spark.SparkDateTimeException "errorClass" : "CANNOT_PARSE_TIMESTAMP", "sqlState" : "42000", "messageParameters" : { - "message" : "Text 'Unparseable' could not be parsed at index 0", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "ansiConfig" : "\"spark.sql.ansi.enabled\"", + "message" : "Text 'Unparseable' could not be parsed at index 0" } } @@ -418,10 +418,10 @@ org.apache.spark.SparkDateTimeException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "42000", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'Unparseable'", "sourceType" : "\"STRING\"", - "targetType" : "\"TIMESTAMP\"", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "targetType" : "\"TIMESTAMP\"" }, "queryContext" : [ { "objectType" : "", @@ -443,10 +443,10 @@ org.apache.spark.SparkDateTimeException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "42000", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'Unparseable'", "sourceType" : "\"STRING\"", - "targetType" : "\"DATE\"", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "targetType" : "\"DATE\"" }, "queryContext" : [ { "objectType" : "", diff --git a/sql/core/src/test/resources/sql-tests/results/ansi/decimalArithmeticOperations.sql.out b/sql/core/src/test/resources/sql-tests/results/ansi/decimalArithmeticOperations.sql.out index ff799d199d92a..a0a40e06432b2 100644 --- a/sql/core/src/test/resources/sql-tests/results/ansi/decimalArithmeticOperations.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/ansi/decimalArithmeticOperations.sql.out @@ -77,10 +77,10 @@ org.apache.spark.SparkArithmeticException "errorClass" : "NUMERIC_VALUE_OUT_OF_RANGE", "sqlState" : "22005", "messageParameters" : { - "value" : "10000000000000000000000000000000000000.1", + "config" : "\"spark.sql.ansi.enabled\"", "precision" : "38", "scale" : "1", - "config" : "\"spark.sql.ansi.enabled\"" + "value" : "10000000000000000000000000000000000000.1" }, "queryContext" : [ { "objectType" : "", @@ -102,10 +102,10 @@ org.apache.spark.SparkArithmeticException "errorClass" : "NUMERIC_VALUE_OUT_OF_RANGE", "sqlState" : "22005", "messageParameters" : { - "value" : "-11000000000000000000000000000000000000.1", + "config" : "\"spark.sql.ansi.enabled\"", "precision" : "38", "scale" : "1", - "config" : "\"spark.sql.ansi.enabled\"" + "value" : "-11000000000000000000000000000000000000.1" }, "queryContext" : [ { "objectType" : "", @@ -127,10 +127,10 @@ org.apache.spark.SparkArithmeticException "errorClass" : "NUMERIC_VALUE_OUT_OF_RANGE", "sqlState" : "22005", "messageParameters" : { - "value" : "152415787532388367501905199875019052100", + "config" : "\"spark.sql.ansi.enabled\"", "precision" : "38", "scale" : "2", - "config" : "\"spark.sql.ansi.enabled\"" + "value" : "152415787532388367501905199875019052100" }, "queryContext" : [ { "objectType" : "", @@ -152,10 +152,10 @@ org.apache.spark.SparkArithmeticException "errorClass" : "NUMERIC_VALUE_OUT_OF_RANGE", "sqlState" : "22005", "messageParameters" : { - "value" : "1000000000000000000000000000000000000.00000000000000000000000000000000000000", + "config" : "\"spark.sql.ansi.enabled\"", "precision" : "38", "scale" : "6", - "config" : "\"spark.sql.ansi.enabled\"" + "value" : "1000000000000000000000000000000000000.00000000000000000000000000000000000000" }, "queryContext" : [ { "objectType" : "", @@ -201,10 +201,10 @@ org.apache.spark.SparkArithmeticException "errorClass" : "NUMERIC_VALUE_OUT_OF_RANGE", "sqlState" : "22005", "messageParameters" : { - "value" : "10123456789012345678901234567890123456.00000000000000000000000000000000000000", + "config" : "\"spark.sql.ansi.enabled\"", "precision" : "38", "scale" : "6", - "config" : "\"spark.sql.ansi.enabled\"" + "value" : "10123456789012345678901234567890123456.00000000000000000000000000000000000000" }, "queryContext" : [ { "objectType" : "", @@ -226,10 +226,10 @@ org.apache.spark.SparkArithmeticException "errorClass" : "NUMERIC_VALUE_OUT_OF_RANGE", "sqlState" : "22005", "messageParameters" : { - "value" : "101234567890123456789012345678901234.56000000000000000000000000000000000000", + "config" : "\"spark.sql.ansi.enabled\"", "precision" : "38", "scale" : "6", - "config" : "\"spark.sql.ansi.enabled\"" + "value" : "101234567890123456789012345678901234.56000000000000000000000000000000000000" }, "queryContext" : [ { "objectType" : "", @@ -251,10 +251,10 @@ org.apache.spark.SparkArithmeticException "errorClass" : "NUMERIC_VALUE_OUT_OF_RANGE", "sqlState" : "22005", "messageParameters" : { - "value" : "10123456789012345678901234567890123.45600000000000000000000000000000000000", + "config" : "\"spark.sql.ansi.enabled\"", "precision" : "38", "scale" : "6", - "config" : "\"spark.sql.ansi.enabled\"" + "value" : "10123456789012345678901234567890123.45600000000000000000000000000000000000" }, "queryContext" : [ { "objectType" : "", @@ -276,10 +276,10 @@ org.apache.spark.SparkArithmeticException "errorClass" : "NUMERIC_VALUE_OUT_OF_RANGE", "sqlState" : "22005", "messageParameters" : { - "value" : "1012345678901234567890123456789012.34560000000000000000000000000000000000", + "config" : "\"spark.sql.ansi.enabled\"", "precision" : "38", "scale" : "6", - "config" : "\"spark.sql.ansi.enabled\"" + "value" : "1012345678901234567890123456789012.34560000000000000000000000000000000000" }, "queryContext" : [ { "objectType" : "", @@ -301,10 +301,10 @@ org.apache.spark.SparkArithmeticException "errorClass" : "NUMERIC_VALUE_OUT_OF_RANGE", "sqlState" : "22005", "messageParameters" : { - "value" : "101234567890123456789012345678901.23456000000000000000000000000000000000", + "config" : "\"spark.sql.ansi.enabled\"", "precision" : "38", "scale" : "6", - "config" : "\"spark.sql.ansi.enabled\"" + "value" : "101234567890123456789012345678901.23456000000000000000000000000000000000" }, "queryContext" : [ { "objectType" : "", @@ -334,10 +334,10 @@ org.apache.spark.SparkArithmeticException "errorClass" : "NUMERIC_VALUE_OUT_OF_RANGE", "sqlState" : "22005", "messageParameters" : { - "value" : "101234567890123456789012345678901.23456000000000000000000000000000000000", + "config" : "\"spark.sql.ansi.enabled\"", "precision" : "38", "scale" : "6", - "config" : "\"spark.sql.ansi.enabled\"" + "value" : "101234567890123456789012345678901.23456000000000000000000000000000000000" }, "queryContext" : [ { "objectType" : "", diff --git a/sql/core/src/test/resources/sql-tests/results/ansi/interval.sql.out b/sql/core/src/test/resources/sql-tests/results/ansi/interval.sql.out index fe601ed66b8fd..4c1e840f8f385 100644 --- a/sql/core/src/test/resources/sql-tests/results/ansi/interval.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/ansi/interval.sql.out @@ -123,10 +123,10 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "42000", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'a'", "sourceType" : "\"STRING\"", - "targetType" : "\"DOUBLE\"", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "targetType" : "\"DOUBLE\"" }, "queryContext" : [ { "objectType" : "", @@ -148,10 +148,10 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "42000", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'a'", "sourceType" : "\"STRING\"", - "targetType" : "\"DOUBLE\"", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "targetType" : "\"DOUBLE\"" }, "queryContext" : [ { "objectType" : "", @@ -173,10 +173,10 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "42000", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'a'", "sourceType" : "\"STRING\"", - "targetType" : "\"DOUBLE\"", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "targetType" : "\"DOUBLE\"" }, "queryContext" : [ { "objectType" : "", @@ -198,10 +198,10 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "42000", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'a'", "sourceType" : "\"STRING\"", - "targetType" : "\"DOUBLE\"", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "targetType" : "\"DOUBLE\"" }, "queryContext" : [ { "objectType" : "", @@ -239,10 +239,10 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "42000", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'a'", "sourceType" : "\"STRING\"", - "targetType" : "\"DOUBLE\"", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "targetType" : "\"DOUBLE\"" }, "queryContext" : [ { "objectType" : "", @@ -264,10 +264,10 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "42000", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'a'", "sourceType" : "\"STRING\"", - "targetType" : "\"DOUBLE\"", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "targetType" : "\"DOUBLE\"" }, "queryContext" : [ { "objectType" : "", @@ -289,9 +289,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(2 / INTERVAL '02' SECOND)\"", "left" : "\"STRING\"", - "right" : "\"INTERVAL SECOND\"" + "right" : "\"INTERVAL SECOND\"", + "sqlExpr" : "\"(2 / INTERVAL '02' SECOND)\"" }, "queryContext" : [ { "objectType" : "", @@ -313,9 +313,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(2 / INTERVAL '2' YEAR)\"", "left" : "\"STRING\"", - "right" : "\"INTERVAL YEAR\"" + "right" : "\"INTERVAL YEAR\"", + "sqlExpr" : "\"(2 / INTERVAL '2' YEAR)\"" }, "queryContext" : [ { "objectType" : "", @@ -423,9 +423,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(2 / INTERVAL '2' YEAR)\"", "left" : "\"INT\"", - "right" : "\"INTERVAL YEAR\"" + "right" : "\"INTERVAL YEAR\"", + "sqlExpr" : "\"(2 / INTERVAL '2' YEAR)\"" }, "queryContext" : [ { "objectType" : "", @@ -447,9 +447,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(2 / INTERVAL '02' HOUR)\"", "left" : "\"INT\"", - "right" : "\"INTERVAL HOUR\"" + "right" : "\"INTERVAL HOUR\"", + "sqlExpr" : "\"(2 / INTERVAL '02' HOUR)\"" }, "queryContext" : [ { "objectType" : "", @@ -471,9 +471,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(NULL / INTERVAL '2' YEAR)\"", "left" : "\"VOID\"", - "right" : "\"INTERVAL YEAR\"" + "right" : "\"INTERVAL YEAR\"", + "sqlExpr" : "\"(NULL / INTERVAL '2' YEAR)\"" }, "queryContext" : [ { "objectType" : "", @@ -495,9 +495,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(NULL / INTERVAL '02' HOUR)\"", "left" : "\"VOID\"", - "right" : "\"INTERVAL HOUR\"" + "right" : "\"INTERVAL HOUR\"", + "sqlExpr" : "\"(NULL / INTERVAL '02' HOUR)\"" }, "queryContext" : [ { "objectType" : "", @@ -847,10 +847,10 @@ org.apache.spark.SparkArithmeticException "errorClass" : "NUMERIC_VALUE_OUT_OF_RANGE", "sqlState" : "22005", "messageParameters" : { - "value" : "1234567890123456789", + "config" : "\"spark.sql.ansi.enabled\"", "precision" : "18", "scale" : "6", - "config" : "\"spark.sql.ansi.enabled\"" + "value" : "1234567890123456789" }, "queryContext" : [ { "objectType" : "", @@ -1673,9 +1673,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(INTERVAL '2' YEAR + 3-3 year to month)\"", "left" : "\"INTERVAL YEAR\"", - "right" : "\"STRING\"" + "right" : "\"STRING\"", + "sqlExpr" : "\"(INTERVAL '2' YEAR + 3-3 year to month)\"" }, "queryContext" : [ { "objectType" : "", @@ -1713,9 +1713,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(INTERVAL '2' YEAR + 3-3)\"", "left" : "\"INTERVAL YEAR\"", - "right" : "\"STRING\"" + "right" : "\"STRING\"", + "sqlExpr" : "\"(INTERVAL '2' YEAR + 3-3)\"" }, "queryContext" : [ { "objectType" : "", @@ -1737,9 +1737,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(INTERVAL '2' YEAR - 4)\"", "left" : "\"INTERVAL YEAR\"", - "right" : "\"STRING\"" + "right" : "\"STRING\"", + "sqlExpr" : "\"(INTERVAL '2' YEAR - 4)\"" }, "queryContext" : [ { "objectType" : "", @@ -1761,10 +1761,10 @@ org.apache.spark.SparkDateTimeException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "42000", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'4 11:11'", "sourceType" : "\"STRING\"", - "targetType" : "\"TIMESTAMP\"", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "targetType" : "\"TIMESTAMP\"" }, "queryContext" : [ { "objectType" : "", @@ -1786,10 +1786,10 @@ org.apache.spark.SparkDateTimeException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "42000", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'4 12:12:12'", "sourceType" : "\"STRING\"", - "targetType" : "\"TIMESTAMP\"", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "targetType" : "\"TIMESTAMP\"" }, "queryContext" : [ { "objectType" : "", @@ -1819,9 +1819,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(INTERVAL '2' YEAR + str)\"", "left" : "\"INTERVAL YEAR\"", - "right" : "\"STRING\"" + "right" : "\"STRING\"", + "sqlExpr" : "\"(INTERVAL '2' YEAR + str)\"" }, "queryContext" : [ { "objectType" : "", @@ -1843,9 +1843,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(INTERVAL '2' YEAR - str)\"", "left" : "\"INTERVAL YEAR\"", - "right" : "\"STRING\"" + "right" : "\"STRING\"", + "sqlExpr" : "\"(INTERVAL '2' YEAR - str)\"" }, "queryContext" : [ { "objectType" : "", @@ -1867,10 +1867,10 @@ org.apache.spark.SparkDateTimeException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "42000", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'1'", "sourceType" : "\"STRING\"", - "targetType" : "\"TIMESTAMP\"", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "targetType" : "\"TIMESTAMP\"" }, "queryContext" : [ { "objectType" : "", @@ -1892,10 +1892,10 @@ org.apache.spark.SparkDateTimeException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "42000", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'1'", "sourceType" : "\"STRING\"", - "targetType" : "\"TIMESTAMP\"", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "targetType" : "\"TIMESTAMP\"" }, "queryContext" : [ { "objectType" : "", @@ -1944,9 +1944,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(INTERVAL '3' DAY - INTERVAL '2-2' YEAR TO MONTH)\"", "left" : "\"INTERVAL DAY\"", - "right" : "\"INTERVAL YEAR TO MONTH\"" + "right" : "\"INTERVAL YEAR TO MONTH\"", + "sqlExpr" : "\"(INTERVAL '3' DAY - INTERVAL '2-2' YEAR TO MONTH)\"" }, "queryContext" : [ { "objectType" : "", @@ -1977,9 +1977,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(1 + INTERVAL '2' MONTH)\"", "left" : "\"INT\"", - "right" : "\"INTERVAL MONTH\"" + "right" : "\"INTERVAL MONTH\"", + "sqlExpr" : "\"(1 + INTERVAL '2' MONTH)\"" }, "queryContext" : [ { "objectType" : "", @@ -2010,9 +2010,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(INTERVAL '2' MONTH - 1)\"", "left" : "\"INTERVAL MONTH\"", - "right" : "\"INT\"" + "right" : "\"INT\"", + "sqlExpr" : "\"(INTERVAL '2' MONTH - 1)\"" }, "queryContext" : [ { "objectType" : "", @@ -2161,8 +2161,8 @@ org.apache.spark.SparkArithmeticException "errorClass" : "INTERVAL_ARITHMETIC_OVERFLOW", "sqlState" : "22003", "messageParameters" : { - "message" : "integer overflow", - "alternative" : "" + "alternative" : "", + "message" : "integer overflow" } } @@ -2177,8 +2177,8 @@ org.apache.spark.SparkArithmeticException "errorClass" : "INTERVAL_ARITHMETIC_OVERFLOW", "sqlState" : "22003", "messageParameters" : { - "message" : "integer overflow", - "alternative" : " Use 'try_subtract' to tolerate overflow and return NULL instead." + "alternative" : " Use 'try_subtract' to tolerate overflow and return NULL instead.", + "message" : "integer overflow" } } @@ -2193,8 +2193,8 @@ org.apache.spark.SparkArithmeticException "errorClass" : "INTERVAL_ARITHMETIC_OVERFLOW", "sqlState" : "22003", "messageParameters" : { - "message" : "integer overflow", - "alternative" : " Use 'try_add' to tolerate overflow and return NULL instead." + "alternative" : " Use 'try_add' to tolerate overflow and return NULL instead.", + "message" : "integer overflow" } } @@ -2429,8 +2429,8 @@ org.apache.spark.SparkArithmeticException "errorClass" : "INTERVAL_ARITHMETIC_OVERFLOW", "sqlState" : "22003", "messageParameters" : { - "message" : "Interval value overflows after being divided by -1", - "alternative" : " Use 'try_divide' to tolerate overflow and return NULL instead." + "alternative" : " Use 'try_divide' to tolerate overflow and return NULL instead.", + "message" : "Interval value overflows after being divided by -1" }, "queryContext" : [ { "objectType" : "", @@ -2452,8 +2452,8 @@ org.apache.spark.SparkArithmeticException "errorClass" : "INTERVAL_ARITHMETIC_OVERFLOW", "sqlState" : "22003", "messageParameters" : { - "message" : "Interval value overflows after being divided by -1", - "alternative" : " Use 'try_divide' to tolerate overflow and return NULL instead." + "alternative" : " Use 'try_divide' to tolerate overflow and return NULL instead.", + "message" : "Interval value overflows after being divided by -1" }, "queryContext" : [ { "objectType" : "", @@ -2509,8 +2509,8 @@ org.apache.spark.SparkArithmeticException "errorClass" : "INTERVAL_ARITHMETIC_OVERFLOW", "sqlState" : "22003", "messageParameters" : { - "message" : "Interval value overflows after being divided by -1", - "alternative" : " Use 'try_divide' to tolerate overflow and return NULL instead." + "alternative" : " Use 'try_divide' to tolerate overflow and return NULL instead.", + "message" : "Interval value overflows after being divided by -1" }, "queryContext" : [ { "objectType" : "", @@ -2532,8 +2532,8 @@ org.apache.spark.SparkArithmeticException "errorClass" : "INTERVAL_ARITHMETIC_OVERFLOW", "sqlState" : "22003", "messageParameters" : { - "message" : "Interval value overflows after being divided by -1", - "alternative" : " Use 'try_divide' to tolerate overflow and return NULL instead." + "alternative" : " Use 'try_divide' to tolerate overflow and return NULL instead.", + "message" : "Interval value overflows after being divided by -1" }, "queryContext" : [ { "objectType" : "", @@ -2881,9 +2881,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(INTERVAL '1' MONTH > INTERVAL '20' DAY)\"", "left" : "\"INTERVAL MONTH\"", - "right" : "\"INTERVAL DAY\"" + "right" : "\"INTERVAL DAY\"", + "sqlExpr" : "\"(INTERVAL '1' MONTH > INTERVAL '20' DAY)\"" }, "queryContext" : [ { "objectType" : "", @@ -2905,9 +2905,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(INTERVAL '1' DAY < 1)\"", "left" : "\"INTERVAL DAY\"", - "right" : "\"STRING\"" + "right" : "\"STRING\"", + "sqlExpr" : "\"(INTERVAL '1' DAY < 1)\"" }, "queryContext" : [ { "objectType" : "", @@ -2929,9 +2929,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(INTERVAL '1' DAY = 1)\"", "left" : "\"INTERVAL DAY\"", - "right" : "\"STRING\"" + "right" : "\"STRING\"", + "sqlExpr" : "\"(INTERVAL '1' DAY = 1)\"" }, "queryContext" : [ { "objectType" : "", @@ -2953,9 +2953,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(INTERVAL '1' DAY > 1)\"", "left" : "\"INTERVAL DAY\"", - "right" : "\"STRING\"" + "right" : "\"STRING\"", + "sqlExpr" : "\"(INTERVAL '1' DAY > 1)\"" }, "queryContext" : [ { "objectType" : "", @@ -2977,9 +2977,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(1 < INTERVAL '1' DAY)\"", "left" : "\"STRING\"", - "right" : "\"INTERVAL DAY\"" + "right" : "\"INTERVAL DAY\"", + "sqlExpr" : "\"(1 < INTERVAL '1' DAY)\"" }, "queryContext" : [ { "objectType" : "", @@ -3001,9 +3001,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(1 = INTERVAL '1' DAY)\"", "left" : "\"STRING\"", - "right" : "\"INTERVAL DAY\"" + "right" : "\"INTERVAL DAY\"", + "sqlExpr" : "\"(1 = INTERVAL '1' DAY)\"" }, "queryContext" : [ { "objectType" : "", @@ -3025,9 +3025,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(1 > INTERVAL '1' DAY)\"", "left" : "\"STRING\"", - "right" : "\"INTERVAL DAY\"" + "right" : "\"INTERVAL DAY\"", + "sqlExpr" : "\"(1 > INTERVAL '1' DAY)\"" }, "queryContext" : [ { "objectType" : "", @@ -3049,9 +3049,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(INTERVAL '1' YEAR < 1)\"", "left" : "\"INTERVAL YEAR\"", - "right" : "\"STRING\"" + "right" : "\"STRING\"", + "sqlExpr" : "\"(INTERVAL '1' YEAR < 1)\"" }, "queryContext" : [ { "objectType" : "", @@ -3073,9 +3073,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(INTERVAL '1' YEAR = 1)\"", "left" : "\"INTERVAL YEAR\"", - "right" : "\"STRING\"" + "right" : "\"STRING\"", + "sqlExpr" : "\"(INTERVAL '1' YEAR = 1)\"" }, "queryContext" : [ { "objectType" : "", @@ -3097,9 +3097,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(INTERVAL '1' YEAR > 1)\"", "left" : "\"INTERVAL YEAR\"", - "right" : "\"STRING\"" + "right" : "\"STRING\"", + "sqlExpr" : "\"(INTERVAL '1' YEAR > 1)\"" }, "queryContext" : [ { "objectType" : "", @@ -3121,9 +3121,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(1 < INTERVAL '1' YEAR)\"", "left" : "\"STRING\"", - "right" : "\"INTERVAL YEAR\"" + "right" : "\"INTERVAL YEAR\"", + "sqlExpr" : "\"(1 < INTERVAL '1' YEAR)\"" }, "queryContext" : [ { "objectType" : "", @@ -3145,9 +3145,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(1 = INTERVAL '1' YEAR)\"", "left" : "\"STRING\"", - "right" : "\"INTERVAL YEAR\"" + "right" : "\"INTERVAL YEAR\"", + "sqlExpr" : "\"(1 = INTERVAL '1' YEAR)\"" }, "queryContext" : [ { "objectType" : "", @@ -3169,9 +3169,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(1 > INTERVAL '1' YEAR)\"", "left" : "\"STRING\"", - "right" : "\"INTERVAL YEAR\"" + "right" : "\"INTERVAL YEAR\"", + "sqlExpr" : "\"(1 > INTERVAL '1' YEAR)\"" }, "queryContext" : [ { "objectType" : "", @@ -3291,9 +3291,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(INTERVAL '1' MONTH div INTERVAL '-1' DAY)\"", "left" : "\"INTERVAL MONTH\"", - "right" : "\"INTERVAL DAY\"" + "right" : "\"INTERVAL DAY\"", + "sqlExpr" : "\"(INTERVAL '1' MONTH div INTERVAL '-1' DAY)\"" }, "queryContext" : [ { "objectType" : "", diff --git a/sql/core/src/test/resources/sql-tests/results/ansi/string-functions.sql.out b/sql/core/src/test/resources/sql-tests/results/ansi/string-functions.sql.out index 810f1942be233..4e04a29a98cf1 100644 --- a/sql/core/src/test/resources/sql-tests/results/ansi/string-functions.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/ansi/string-functions.sql.out @@ -83,10 +83,10 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "42000", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'a'", "sourceType" : "\"STRING\"", - "targetType" : "\"INT\"", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "targetType" : "\"INT\"" }, "queryContext" : [ { "objectType" : "", @@ -124,10 +124,10 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "42000", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'a'", "sourceType" : "\"STRING\"", - "targetType" : "\"INT\"", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "targetType" : "\"INT\"" }, "queryContext" : [ { "objectType" : "", @@ -488,10 +488,10 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "42000", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'invalid_length'", "sourceType" : "\"STRING\"", - "targetType" : "\"INT\"", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "targetType" : "\"INT\"" }, "queryContext" : [ { "objectType" : "", @@ -513,10 +513,10 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "42000", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'invalid_length'", "sourceType" : "\"STRING\"", - "targetType" : "\"INT\"", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "targetType" : "\"INT\"" }, "queryContext" : [ { "objectType" : "", @@ -1187,10 +1187,10 @@ org.apache.spark.SparkIllegalArgumentException { "errorClass" : "CONVERSION_INVALID_INPUT", "messageParameters" : { - "str" : "' ab cdef= = '", "fmt" : "'BASE64'", - "targetType" : "\"BINARY\"", - "suggestion" : "`try_to_binary`" + "str" : "' ab cdef= = '", + "suggestion" : "`try_to_binary`", + "targetType" : "\"BINARY\"" } } @@ -1214,10 +1214,10 @@ org.apache.spark.SparkIllegalArgumentException { "errorClass" : "CONVERSION_INVALID_INPUT", "messageParameters" : { - "str" : "'a'", "fmt" : "'BASE64'", - "targetType" : "\"BINARY\"", - "suggestion" : "`try_to_binary`" + "str" : "'a'", + "suggestion" : "`try_to_binary`", + "targetType" : "\"BINARY\"" } } @@ -1231,10 +1231,10 @@ org.apache.spark.SparkIllegalArgumentException { "errorClass" : "CONVERSION_INVALID_INPUT", "messageParameters" : { - "str" : "'a?'", "fmt" : "'BASE64'", - "targetType" : "\"BINARY\"", - "suggestion" : "`try_to_binary`" + "str" : "'a?'", + "suggestion" : "`try_to_binary`", + "targetType" : "\"BINARY\"" } } @@ -1248,10 +1248,10 @@ org.apache.spark.SparkIllegalArgumentException { "errorClass" : "CONVERSION_INVALID_INPUT", "messageParameters" : { - "str" : "'abcde'", "fmt" : "'BASE64'", - "targetType" : "\"BINARY\"", - "suggestion" : "`try_to_binary`" + "str" : "'abcde'", + "suggestion" : "`try_to_binary`", + "targetType" : "\"BINARY\"" } } @@ -1265,10 +1265,10 @@ org.apache.spark.SparkIllegalArgumentException { "errorClass" : "CONVERSION_INVALID_INPUT", "messageParameters" : { - "str" : "'abcd='", "fmt" : "'BASE64'", - "targetType" : "\"BINARY\"", - "suggestion" : "`try_to_binary`" + "str" : "'abcd='", + "suggestion" : "`try_to_binary`", + "targetType" : "\"BINARY\"" } } @@ -1282,10 +1282,10 @@ org.apache.spark.SparkIllegalArgumentException { "errorClass" : "CONVERSION_INVALID_INPUT", "messageParameters" : { - "str" : "'a==='", "fmt" : "'BASE64'", - "targetType" : "\"BINARY\"", - "suggestion" : "`try_to_binary`" + "str" : "'a==='", + "suggestion" : "`try_to_binary`", + "targetType" : "\"BINARY\"" } } @@ -1299,10 +1299,10 @@ org.apache.spark.SparkIllegalArgumentException { "errorClass" : "CONVERSION_INVALID_INPUT", "messageParameters" : { - "str" : "'ab==f'", "fmt" : "'BASE64'", - "targetType" : "\"BINARY\"", - "suggestion" : "`try_to_binary`" + "str" : "'ab==f'", + "suggestion" : "`try_to_binary`", + "targetType" : "\"BINARY\"" } } @@ -1389,10 +1389,10 @@ org.apache.spark.SparkIllegalArgumentException { "errorClass" : "CONVERSION_INVALID_INPUT", "messageParameters" : { - "str" : "'GG'", "fmt" : "'HEX'", - "targetType" : "\"BINARY\"", - "suggestion" : "`try_to_binary`" + "str" : "'GG'", + "suggestion" : "`try_to_binary`", + "targetType" : "\"BINARY\"" } } @@ -1406,10 +1406,10 @@ org.apache.spark.SparkIllegalArgumentException { "errorClass" : "CONVERSION_INVALID_INPUT", "messageParameters" : { - "str" : "'01 AF'", "fmt" : "'HEX'", - "targetType" : "\"BINARY\"", - "suggestion" : "`try_to_binary`" + "str" : "'01 AF'", + "suggestion" : "`try_to_binary`", + "targetType" : "\"BINARY\"" } } @@ -1431,10 +1431,10 @@ org.apache.spark.SparkIllegalArgumentException { "errorClass" : "CONVERSION_INVALID_INPUT", "messageParameters" : { - "str" : "' ab cdef= = '", "fmt" : "'BASE64'", - "targetType" : "\"BINARY\"", - "suggestion" : "`try_to_binary`" + "str" : "' ab cdef= = '", + "suggestion" : "`try_to_binary`", + "targetType" : "\"BINARY\"" } } @@ -1448,10 +1448,10 @@ org.apache.spark.SparkIllegalArgumentException { "errorClass" : "CONVERSION_INVALID_INPUT", "messageParameters" : { - "str" : "' ab cdef= = '", "fmt" : "'HEX'", - "targetType" : "\"BINARY\"", - "suggestion" : "`try_to_binary`" + "str" : "' ab cdef= = '", + "suggestion" : "`try_to_binary`", + "targetType" : "\"BINARY\"" } } diff --git a/sql/core/src/test/resources/sql-tests/results/ansi/timestamp.sql.out b/sql/core/src/test/resources/sql-tests/results/ansi/timestamp.sql.out index bd2bed273180d..bd675df539e77 100644 --- a/sql/core/src/test/resources/sql-tests/results/ansi/timestamp.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/ansi/timestamp.sql.out @@ -347,8 +347,8 @@ org.apache.spark.SparkDateTimeException "errorClass" : "CANNOT_PARSE_TIMESTAMP", "sqlState" : "42000", "messageParameters" : { - "message" : "Text '2019-10-06 10:11:12.' could not be parsed at index 20", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "ansiConfig" : "\"spark.sql.ansi.enabled\"", + "message" : "Text '2019-10-06 10:11:12.' could not be parsed at index 20" } } @@ -419,8 +419,8 @@ org.apache.spark.SparkDateTimeException "errorClass" : "CANNOT_PARSE_TIMESTAMP", "sqlState" : "42000", "messageParameters" : { - "message" : "Text '2019-10-06 10:11:12.1234567PST' could not be parsed, unparsed text found at index 26", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "ansiConfig" : "\"spark.sql.ansi.enabled\"", + "message" : "Text '2019-10-06 10:11:12.1234567PST' could not be parsed, unparsed text found at index 26" } } @@ -443,8 +443,8 @@ org.apache.spark.SparkDateTimeException "errorClass" : "CANNOT_PARSE_TIMESTAMP", "sqlState" : "42000", "messageParameters" : { - "message" : "Text '223456 2019-10-06 10:11:12.123456PST' could not be parsed at index 27", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "ansiConfig" : "\"spark.sql.ansi.enabled\"", + "message" : "Text '223456 2019-10-06 10:11:12.123456PST' could not be parsed at index 27" } } @@ -515,8 +515,8 @@ org.apache.spark.SparkDateTimeException "errorClass" : "CANNOT_PARSE_TIMESTAMP", "sqlState" : "42000", "messageParameters" : { - "message" : "Text '12.1232019-10-06S10:11' could not be parsed at index 7", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "ansiConfig" : "\"spark.sql.ansi.enabled\"", + "message" : "Text '12.1232019-10-06S10:11' could not be parsed at index 7" } } @@ -531,8 +531,8 @@ org.apache.spark.SparkDateTimeException "errorClass" : "CANNOT_PARSE_TIMESTAMP", "sqlState" : "42000", "messageParameters" : { - "message" : "Text '12.1232019-10-06S10:11' could not be parsed at index 9", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "ansiConfig" : "\"spark.sql.ansi.enabled\"", + "message" : "Text '12.1232019-10-06S10:11' could not be parsed at index 9" } } @@ -611,8 +611,8 @@ org.apache.spark.SparkDateTimeException "errorClass" : "CANNOT_PARSE_TIMESTAMP", "sqlState" : "42000", "messageParameters" : { - "message" : "Invalid date 'February 29' as '1970' is not a leap year", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "ansiConfig" : "\"spark.sql.ansi.enabled\"", + "message" : "Invalid date 'February 29' as '1970' is not a leap year" } } @@ -723,9 +723,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_WRONG_TYPE", "messageParameters" : { - "sqlExpr" : "\"(TIMESTAMP '2011-11-11 11:11:11' + 1)\"", + "actualDataType" : "\"TIMESTAMP\"", "inputType" : "(\"NUMERIC\" or \"INTERVAL DAY TO SECOND\" or \"INTERVAL YEAR TO MONTH\" or \"INTERVAL\")", - "actualDataType" : "\"TIMESTAMP\"" + "sqlExpr" : "\"(TIMESTAMP '2011-11-11 11:11:11' + 1)\"" }, "queryContext" : [ { "objectType" : "", @@ -747,9 +747,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_WRONG_TYPE", "messageParameters" : { - "sqlExpr" : "\"(1 + TIMESTAMP '2011-11-11 11:11:11')\"", + "actualDataType" : "\"TIMESTAMP\"", "inputType" : "(\"NUMERIC\" or \"INTERVAL DAY TO SECOND\" or \"INTERVAL YEAR TO MONTH\" or \"INTERVAL\")", - "actualDataType" : "\"TIMESTAMP\"" + "sqlExpr" : "\"(1 + TIMESTAMP '2011-11-11 11:11:11')\"" }, "queryContext" : [ { "objectType" : "", @@ -771,9 +771,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(TIMESTAMP '2011-11-11 11:11:11' + NULL)\"", "left" : "\"TIMESTAMP\"", - "right" : "\"VOID\"" + "right" : "\"VOID\"", + "sqlExpr" : "\"(TIMESTAMP '2011-11-11 11:11:11' + NULL)\"" }, "queryContext" : [ { "objectType" : "", @@ -795,9 +795,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(NULL + TIMESTAMP '2011-11-11 11:11:11')\"", "left" : "\"VOID\"", - "right" : "\"TIMESTAMP\"" + "right" : "\"TIMESTAMP\"", + "sqlExpr" : "\"(NULL + TIMESTAMP '2011-11-11 11:11:11')\"" }, "queryContext" : [ { "objectType" : "", @@ -842,8 +842,8 @@ org.apache.spark.SparkUpgradeException "errorClass" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION", "errorSubClass" : "DATETIME_PATTERN_RECOGNITION", "messageParameters" : { - "pattern" : "'yyyy-MM-dd GGGGG'", - "config" : "\"spark.sql.legacy.timeParserPolicy\"" + "config" : "\"spark.sql.legacy.timeParserPolicy\"", + "pattern" : "'yyyy-MM-dd GGGGG'" } } @@ -858,8 +858,8 @@ org.apache.spark.SparkUpgradeException "errorClass" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION", "errorSubClass" : "DATETIME_PATTERN_RECOGNITION", "messageParameters" : { - "pattern" : "'dd MM yyyy EEEEEE'", - "config" : "\"spark.sql.legacy.timeParserPolicy\"" + "config" : "\"spark.sql.legacy.timeParserPolicy\"", + "pattern" : "'dd MM yyyy EEEEEE'" } } @@ -874,8 +874,8 @@ org.apache.spark.SparkUpgradeException "errorClass" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION", "errorSubClass" : "DATETIME_PATTERN_RECOGNITION", "messageParameters" : { - "pattern" : "'dd MM yyyy EEEEE'", - "config" : "\"spark.sql.legacy.timeParserPolicy\"" + "config" : "\"spark.sql.legacy.timeParserPolicy\"", + "pattern" : "'dd MM yyyy EEEEE'" } } @@ -890,8 +890,8 @@ org.apache.spark.SparkUpgradeException "errorClass" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION", "errorSubClass" : "DATETIME_PATTERN_RECOGNITION", "messageParameters" : { - "pattern" : "'dd MM yyyy EEEEE'", - "config" : "\"spark.sql.legacy.timeParserPolicy\"" + "config" : "\"spark.sql.legacy.timeParserPolicy\"", + "pattern" : "'dd MM yyyy EEEEE'" } } @@ -906,8 +906,8 @@ org.apache.spark.SparkUpgradeException "errorClass" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION", "errorSubClass" : "DATETIME_PATTERN_RECOGNITION", "messageParameters" : { - "pattern" : "'dd/MMMMM/yyyy'", - "config" : "\"spark.sql.legacy.timeParserPolicy\"" + "config" : "\"spark.sql.legacy.timeParserPolicy\"", + "pattern" : "'dd/MMMMM/yyyy'" } } @@ -922,8 +922,8 @@ org.apache.spark.SparkUpgradeException "errorClass" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION", "errorSubClass" : "DATETIME_PATTERN_RECOGNITION", "messageParameters" : { - "pattern" : "'dd/MMMMM/yyyy'", - "config" : "\"spark.sql.legacy.timeParserPolicy\"" + "config" : "\"spark.sql.legacy.timeParserPolicy\"", + "pattern" : "'dd/MMMMM/yyyy'" } } diff --git a/sql/core/src/test/resources/sql-tests/results/ansi/try_aggregates.sql.out b/sql/core/src/test/resources/sql-tests/results/ansi/try_aggregates.sql.out index 2aee862ad2272..94048ac8897bb 100644 --- a/sql/core/src/test/resources/sql-tests/results/ansi/try_aggregates.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/ansi/try_aggregates.sql.out @@ -155,9 +155,9 @@ org.apache.spark.SparkArithmeticException "errorClass" : "ARITHMETIC_OVERFLOW", "sqlState" : "22003", "messageParameters" : { - "message" : "long overflow", "alternative" : " Use 'try_add' to tolerate overflow and return NULL instead.", - "config" : "\"spark.sql.ansi.enabled\"" + "config" : "\"spark.sql.ansi.enabled\"", + "message" : "long overflow" }, "queryContext" : [ { "objectType" : "", @@ -363,9 +363,9 @@ org.apache.spark.SparkArithmeticException "errorClass" : "ARITHMETIC_OVERFLOW", "sqlState" : "22003", "messageParameters" : { - "message" : "long overflow", "alternative" : " Use 'try_add' to tolerate overflow and return NULL instead.", - "config" : "\"spark.sql.ansi.enabled\"" + "config" : "\"spark.sql.ansi.enabled\"", + "message" : "long overflow" }, "queryContext" : [ { "objectType" : "", diff --git a/sql/core/src/test/resources/sql-tests/results/ansi/try_arithmetic.sql.out b/sql/core/src/test/resources/sql-tests/results/ansi/try_arithmetic.sql.out index 3f317a0b982f1..9287ea356023f 100644 --- a/sql/core/src/test/resources/sql-tests/results/ansi/try_arithmetic.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/ansi/try_arithmetic.sql.out @@ -49,9 +49,9 @@ org.apache.spark.SparkArithmeticException "errorClass" : "ARITHMETIC_OVERFLOW", "sqlState" : "22003", "messageParameters" : { - "message" : "integer overflow", "alternative" : " Use 'try_add' to tolerate overflow and return NULL instead.", - "config" : "\"spark.sql.ansi.enabled\"" + "config" : "\"spark.sql.ansi.enabled\"", + "message" : "integer overflow" }, "queryContext" : [ { "objectType" : "", @@ -73,9 +73,9 @@ org.apache.spark.SparkArithmeticException "errorClass" : "ARITHMETIC_OVERFLOW", "sqlState" : "22003", "messageParameters" : { - "message" : "long overflow", "alternative" : " Use 'try_add' to tolerate overflow and return NULL instead.", - "config" : "\"spark.sql.ansi.enabled\"" + "config" : "\"spark.sql.ansi.enabled\"", + "message" : "long overflow" }, "queryContext" : [ { "objectType" : "", @@ -264,9 +264,9 @@ org.apache.spark.SparkArithmeticException "errorClass" : "ARITHMETIC_OVERFLOW", "sqlState" : "22003", "messageParameters" : { - "message" : "integer overflow", "alternative" : " Use 'try_add' to tolerate overflow and return NULL instead.", - "config" : "\"spark.sql.ansi.enabled\"" + "config" : "\"spark.sql.ansi.enabled\"", + "message" : "integer overflow" }, "queryContext" : [ { "objectType" : "", @@ -288,9 +288,9 @@ org.apache.spark.SparkArithmeticException "errorClass" : "ARITHMETIC_OVERFLOW", "sqlState" : "22003", "messageParameters" : { - "message" : "long overflow", "alternative" : " Use 'try_add' to tolerate overflow and return NULL instead.", - "config" : "\"spark.sql.ansi.enabled\"" + "config" : "\"spark.sql.ansi.enabled\"", + "message" : "long overflow" }, "queryContext" : [ { "objectType" : "", @@ -422,9 +422,9 @@ org.apache.spark.SparkArithmeticException "errorClass" : "ARITHMETIC_OVERFLOW", "sqlState" : "22003", "messageParameters" : { - "message" : "integer overflow", "alternative" : " Use 'try_add' to tolerate overflow and return NULL instead.", - "config" : "\"spark.sql.ansi.enabled\"" + "config" : "\"spark.sql.ansi.enabled\"", + "message" : "integer overflow" }, "queryContext" : [ { "objectType" : "", @@ -446,9 +446,9 @@ org.apache.spark.SparkArithmeticException "errorClass" : "ARITHMETIC_OVERFLOW", "sqlState" : "22003", "messageParameters" : { - "message" : "long overflow", "alternative" : " Use 'try_add' to tolerate overflow and return NULL instead.", - "config" : "\"spark.sql.ansi.enabled\"" + "config" : "\"spark.sql.ansi.enabled\"", + "message" : "long overflow" }, "queryContext" : [ { "objectType" : "", @@ -564,9 +564,9 @@ org.apache.spark.SparkArithmeticException "errorClass" : "ARITHMETIC_OVERFLOW", "sqlState" : "22003", "messageParameters" : { - "message" : "integer overflow", "alternative" : " Use 'try_add' to tolerate overflow and return NULL instead.", - "config" : "\"spark.sql.ansi.enabled\"" + "config" : "\"spark.sql.ansi.enabled\"", + "message" : "integer overflow" }, "queryContext" : [ { "objectType" : "", @@ -588,9 +588,9 @@ org.apache.spark.SparkArithmeticException "errorClass" : "ARITHMETIC_OVERFLOW", "sqlState" : "22003", "messageParameters" : { - "message" : "long overflow", "alternative" : " Use 'try_add' to tolerate overflow and return NULL instead.", - "config" : "\"spark.sql.ansi.enabled\"" + "config" : "\"spark.sql.ansi.enabled\"", + "message" : "long overflow" }, "queryContext" : [ { "objectType" : "", diff --git a/sql/core/src/test/resources/sql-tests/results/ansi/try_datetime_functions.sql.out b/sql/core/src/test/resources/sql-tests/results/ansi/try_datetime_functions.sql.out index ab2172f56de86..c49637869e1fc 100644 --- a/sql/core/src/test/resources/sql-tests/results/ansi/try_datetime_functions.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/ansi/try_datetime_functions.sql.out @@ -49,7 +49,7 @@ org.apache.spark.SparkUpgradeException "errorClass" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION", "errorSubClass" : "DATETIME_PATTERN_RECOGNITION", "messageParameters" : { - "pattern" : "'dd MM yyyy EEEEEE'", - "config" : "\"spark.sql.legacy.timeParserPolicy\"" + "config" : "\"spark.sql.legacy.timeParserPolicy\"", + "pattern" : "'dd MM yyyy EEEEEE'" } } diff --git a/sql/core/src/test/resources/sql-tests/results/cast.sql.out b/sql/core/src/test/resources/sql-tests/results/cast.sql.out index 9271fd6350e47..0b9912160e448 100644 --- a/sql/core/src/test/resources/sql-tests/results/cast.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/cast.sql.out @@ -635,10 +635,10 @@ org.apache.spark.SparkArithmeticException "errorClass" : "CAST_OVERFLOW", "sqlState" : "22005", "messageParameters" : { - "value" : "INTERVAL '23:59:59' HOUR TO SECOND", + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "sourceType" : "\"INTERVAL HOUR TO SECOND\"", "targetType" : "\"SMALLINT\"", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "value" : "INTERVAL '23:59:59' HOUR TO SECOND" } } @@ -669,10 +669,10 @@ org.apache.spark.SparkArithmeticException "errorClass" : "CAST_OVERFLOW", "sqlState" : "22005", "messageParameters" : { - "value" : "INTERVAL '-1000' MONTH", + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "sourceType" : "\"INTERVAL MONTH\"", "targetType" : "\"TINYINT\"", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "value" : "INTERVAL '-1000' MONTH" } } @@ -687,10 +687,10 @@ org.apache.spark.SparkArithmeticException "errorClass" : "CAST_OVERFLOW", "sqlState" : "22005", "messageParameters" : { - "value" : "INTERVAL '1000000' SECOND", + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "sourceType" : "\"INTERVAL SECOND\"", "targetType" : "\"SMALLINT\"", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "value" : "INTERVAL '1000000' SECOND" } } @@ -777,10 +777,10 @@ org.apache.spark.SparkArithmeticException "errorClass" : "CAST_OVERFLOW", "sqlState" : "22005", "messageParameters" : { - "value" : "2147483647", + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "sourceType" : "\"INT\"", "targetType" : "\"INTERVAL YEAR\"", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "value" : "2147483647" } } @@ -795,10 +795,10 @@ org.apache.spark.SparkArithmeticException "errorClass" : "CAST_OVERFLOW", "sqlState" : "22005", "messageParameters" : { - "value" : "-9223372036854775808L", + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "sourceType" : "\"BIGINT\"", "targetType" : "\"INTERVAL DAY\"", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "value" : "-9223372036854775808L" } } @@ -869,10 +869,10 @@ org.apache.spark.SparkArithmeticException "errorClass" : "NUMERIC_VALUE_OUT_OF_RANGE", "sqlState" : "22005", "messageParameters" : { - "value" : "10.123000", + "config" : "\"spark.sql.ansi.enabled\"", "precision" : "1", "scale" : "0", - "config" : "\"spark.sql.ansi.enabled\"" + "value" : "10.123000" } } diff --git a/sql/core/src/test/resources/sql-tests/results/change-column.sql.out b/sql/core/src/test/resources/sql-tests/results/change-column.sql.out index b987fe7c02ad6..68e827ad55cc0 100644 --- a/sql/core/src/test/resources/sql-tests/results/change-column.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/change-column.sql.out @@ -52,8 +52,8 @@ org.apache.spark.sql.AnalysisException "errorSubClass" : "TABLE_OPERATION", "sqlState" : "0A000", "messageParameters" : { - "tableName" : "`spark_catalog`.`default`.`test_change`", - "operation" : "RENAME COLUMN" + "operation" : "RENAME COLUMN", + "tableName" : "`spark_catalog`.`default`.`test_change`" } } @@ -98,8 +98,8 @@ org.apache.spark.sql.AnalysisException "errorSubClass" : "TABLE_OPERATION", "sqlState" : "0A000", "messageParameters" : { - "tableName" : "`spark_catalog`.`default`.`test_change`", - "operation" : "ALTER COLUMN ... FIRST | ALTER" + "operation" : "ALTER COLUMN ... FIRST | ALTER", + "tableName" : "`spark_catalog`.`default`.`test_change`" } } @@ -115,8 +115,8 @@ org.apache.spark.sql.AnalysisException "errorSubClass" : "TABLE_OPERATION", "sqlState" : "0A000", "messageParameters" : { - "tableName" : "`spark_catalog`.`default`.`test_change`", - "operation" : "ALTER COLUMN ... FIRST | ALTER" + "operation" : "ALTER COLUMN ... FIRST | ALTER", + "tableName" : "`spark_catalog`.`default`.`test_change`" } } diff --git a/sql/core/src/test/resources/sql-tests/results/date.sql.out b/sql/core/src/test/resources/sql-tests/results/date.sql.out index e1dbb7799e672..00c5f4c2d911b 100644 --- a/sql/core/src/test/resources/sql-tests/results/date.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/date.sql.out @@ -653,8 +653,8 @@ org.apache.spark.SparkUpgradeException "errorClass" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION", "errorSubClass" : "DATETIME_PATTERN_RECOGNITION", "messageParameters" : { - "pattern" : "'dd/MMMMM/yyyy'", - "config" : "\"spark.sql.legacy.timeParserPolicy\"" + "config" : "\"spark.sql.legacy.timeParserPolicy\"", + "pattern" : "'dd/MMMMM/yyyy'" } } @@ -669,8 +669,8 @@ org.apache.spark.SparkUpgradeException "errorClass" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION", "errorSubClass" : "DATETIME_PATTERN_RECOGNITION", "messageParameters" : { - "pattern" : "'dd/MMMMM/yyyy'", - "config" : "\"spark.sql.legacy.timeParserPolicy\"" + "config" : "\"spark.sql.legacy.timeParserPolicy\"", + "pattern" : "'dd/MMMMM/yyyy'" } } @@ -685,8 +685,8 @@ org.apache.spark.SparkUpgradeException "errorClass" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION", "errorSubClass" : "DATETIME_PATTERN_RECOGNITION", "messageParameters" : { - "pattern" : "'dd/MMMMM/yyyy'", - "config" : "\"spark.sql.legacy.timeParserPolicy\"" + "config" : "\"spark.sql.legacy.timeParserPolicy\"", + "pattern" : "'dd/MMMMM/yyyy'" } } diff --git a/sql/core/src/test/resources/sql-tests/results/datetime-formatting-invalid.sql.out b/sql/core/src/test/resources/sql-tests/results/datetime-formatting-invalid.sql.out index da0a801c4badc..4b9a60a41d847 100644 --- a/sql/core/src/test/resources/sql-tests/results/datetime-formatting-invalid.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/datetime-formatting-invalid.sql.out @@ -9,8 +9,8 @@ org.apache.spark.SparkUpgradeException "errorClass" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION", "errorSubClass" : "DATETIME_PATTERN_RECOGNITION", "messageParameters" : { - "pattern" : "'GGGGG'", - "config" : "\"spark.sql.legacy.timeParserPolicy\"" + "config" : "\"spark.sql.legacy.timeParserPolicy\"", + "pattern" : "'GGGGG'" } } @@ -25,8 +25,8 @@ org.apache.spark.SparkUpgradeException "errorClass" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION", "errorSubClass" : "DATETIME_PATTERN_RECOGNITION", "messageParameters" : { - "pattern" : "'yyyyyyy'", - "config" : "\"spark.sql.legacy.timeParserPolicy\"" + "config" : "\"spark.sql.legacy.timeParserPolicy\"", + "pattern" : "'yyyyyyy'" } } @@ -59,8 +59,8 @@ org.apache.spark.SparkUpgradeException "errorClass" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION", "errorSubClass" : "DATETIME_PATTERN_RECOGNITION", "messageParameters" : { - "pattern" : "'MMMMM'", - "config" : "\"spark.sql.legacy.timeParserPolicy\"" + "config" : "\"spark.sql.legacy.timeParserPolicy\"", + "pattern" : "'MMMMM'" } } @@ -75,8 +75,8 @@ org.apache.spark.SparkUpgradeException "errorClass" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION", "errorSubClass" : "DATETIME_PATTERN_RECOGNITION", "messageParameters" : { - "pattern" : "'LLLLL'", - "config" : "\"spark.sql.legacy.timeParserPolicy\"" + "config" : "\"spark.sql.legacy.timeParserPolicy\"", + "pattern" : "'LLLLL'" } } @@ -91,8 +91,8 @@ org.apache.spark.SparkUpgradeException "errorClass" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION", "errorSubClass" : "DATETIME_PATTERN_RECOGNITION", "messageParameters" : { - "pattern" : "'EEEEE'", - "config" : "\"spark.sql.legacy.timeParserPolicy\"" + "config" : "\"spark.sql.legacy.timeParserPolicy\"", + "pattern" : "'EEEEE'" } } @@ -107,8 +107,8 @@ org.apache.spark.SparkUpgradeException "errorClass" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION", "errorSubClass" : "DATETIME_PATTERN_RECOGNITION", "messageParameters" : { - "pattern" : "'FF'", - "config" : "\"spark.sql.legacy.timeParserPolicy\"" + "config" : "\"spark.sql.legacy.timeParserPolicy\"", + "pattern" : "'FF'" } } @@ -123,8 +123,8 @@ org.apache.spark.SparkUpgradeException "errorClass" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION", "errorSubClass" : "DATETIME_PATTERN_RECOGNITION", "messageParameters" : { - "pattern" : "'ddd'", - "config" : "\"spark.sql.legacy.timeParserPolicy\"" + "config" : "\"spark.sql.legacy.timeParserPolicy\"", + "pattern" : "'ddd'" } } @@ -139,8 +139,8 @@ org.apache.spark.SparkUpgradeException "errorClass" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION", "errorSubClass" : "DATETIME_PATTERN_RECOGNITION", "messageParameters" : { - "pattern" : "'DDDD'", - "config" : "\"spark.sql.legacy.timeParserPolicy\"" + "config" : "\"spark.sql.legacy.timeParserPolicy\"", + "pattern" : "'DDDD'" } } @@ -155,8 +155,8 @@ org.apache.spark.SparkUpgradeException "errorClass" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION", "errorSubClass" : "DATETIME_PATTERN_RECOGNITION", "messageParameters" : { - "pattern" : "'HHH'", - "config" : "\"spark.sql.legacy.timeParserPolicy\"" + "config" : "\"spark.sql.legacy.timeParserPolicy\"", + "pattern" : "'HHH'" } } @@ -171,8 +171,8 @@ org.apache.spark.SparkUpgradeException "errorClass" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION", "errorSubClass" : "DATETIME_PATTERN_RECOGNITION", "messageParameters" : { - "pattern" : "'hhh'", - "config" : "\"spark.sql.legacy.timeParserPolicy\"" + "config" : "\"spark.sql.legacy.timeParserPolicy\"", + "pattern" : "'hhh'" } } @@ -187,8 +187,8 @@ org.apache.spark.SparkUpgradeException "errorClass" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION", "errorSubClass" : "DATETIME_PATTERN_RECOGNITION", "messageParameters" : { - "pattern" : "'kkk'", - "config" : "\"spark.sql.legacy.timeParserPolicy\"" + "config" : "\"spark.sql.legacy.timeParserPolicy\"", + "pattern" : "'kkk'" } } @@ -203,8 +203,8 @@ org.apache.spark.SparkUpgradeException "errorClass" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION", "errorSubClass" : "DATETIME_PATTERN_RECOGNITION", "messageParameters" : { - "pattern" : "'KKK'", - "config" : "\"spark.sql.legacy.timeParserPolicy\"" + "config" : "\"spark.sql.legacy.timeParserPolicy\"", + "pattern" : "'KKK'" } } @@ -219,8 +219,8 @@ org.apache.spark.SparkUpgradeException "errorClass" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION", "errorSubClass" : "DATETIME_PATTERN_RECOGNITION", "messageParameters" : { - "pattern" : "'mmm'", - "config" : "\"spark.sql.legacy.timeParserPolicy\"" + "config" : "\"spark.sql.legacy.timeParserPolicy\"", + "pattern" : "'mmm'" } } @@ -235,8 +235,8 @@ org.apache.spark.SparkUpgradeException "errorClass" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION", "errorSubClass" : "DATETIME_PATTERN_RECOGNITION", "messageParameters" : { - "pattern" : "'sss'", - "config" : "\"spark.sql.legacy.timeParserPolicy\"" + "config" : "\"spark.sql.legacy.timeParserPolicy\"", + "pattern" : "'sss'" } } @@ -251,8 +251,8 @@ org.apache.spark.SparkUpgradeException "errorClass" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION", "errorSubClass" : "DATETIME_PATTERN_RECOGNITION", "messageParameters" : { - "pattern" : "'SSSSSSSSSS'", - "config" : "\"spark.sql.legacy.timeParserPolicy\"" + "config" : "\"spark.sql.legacy.timeParserPolicy\"", + "pattern" : "'SSSSSSSSSS'" } } @@ -267,8 +267,8 @@ org.apache.spark.SparkUpgradeException "errorClass" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION", "errorSubClass" : "DATETIME_PATTERN_RECOGNITION", "messageParameters" : { - "pattern" : "'aa'", - "config" : "\"spark.sql.legacy.timeParserPolicy\"" + "config" : "\"spark.sql.legacy.timeParserPolicy\"", + "pattern" : "'aa'" } } @@ -292,8 +292,8 @@ org.apache.spark.SparkUpgradeException "errorClass" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION", "errorSubClass" : "DATETIME_PATTERN_RECOGNITION", "messageParameters" : { - "pattern" : "'zzzzz'", - "config" : "\"spark.sql.legacy.timeParserPolicy\"" + "config" : "\"spark.sql.legacy.timeParserPolicy\"", + "pattern" : "'zzzzz'" } } @@ -317,8 +317,8 @@ org.apache.spark.SparkUpgradeException "errorClass" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION", "errorSubClass" : "DATETIME_PATTERN_RECOGNITION", "messageParameters" : { - "pattern" : "'ZZZZZZ'", - "config" : "\"spark.sql.legacy.timeParserPolicy\"" + "config" : "\"spark.sql.legacy.timeParserPolicy\"", + "pattern" : "'ZZZZZZ'" } } @@ -387,8 +387,8 @@ org.apache.spark.SparkUpgradeException "errorClass" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION", "errorSubClass" : "DATETIME_PATTERN_RECOGNITION", "messageParameters" : { - "pattern" : "'Y'", - "config" : "\"spark.sql.legacy.timeParserPolicy\"" + "config" : "\"spark.sql.legacy.timeParserPolicy\"", + "pattern" : "'Y'" } } @@ -403,8 +403,8 @@ org.apache.spark.SparkUpgradeException "errorClass" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION", "errorSubClass" : "DATETIME_PATTERN_RECOGNITION", "messageParameters" : { - "pattern" : "'w'", - "config" : "\"spark.sql.legacy.timeParserPolicy\"" + "config" : "\"spark.sql.legacy.timeParserPolicy\"", + "pattern" : "'w'" } } @@ -419,8 +419,8 @@ org.apache.spark.SparkUpgradeException "errorClass" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION", "errorSubClass" : "DATETIME_PATTERN_RECOGNITION", "messageParameters" : { - "pattern" : "'W'", - "config" : "\"spark.sql.legacy.timeParserPolicy\"" + "config" : "\"spark.sql.legacy.timeParserPolicy\"", + "pattern" : "'W'" } } @@ -435,8 +435,8 @@ org.apache.spark.SparkUpgradeException "errorClass" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION", "errorSubClass" : "DATETIME_PATTERN_RECOGNITION", "messageParameters" : { - "pattern" : "'u'", - "config" : "\"spark.sql.legacy.timeParserPolicy\"" + "config" : "\"spark.sql.legacy.timeParserPolicy\"", + "pattern" : "'u'" } } diff --git a/sql/core/src/test/resources/sql-tests/results/datetime-legacy.sql.out b/sql/core/src/test/resources/sql-tests/results/datetime-legacy.sql.out index 6091cb9b0f917..bb363d7e059a7 100644 --- a/sql/core/src/test/resources/sql-tests/results/datetime-legacy.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/datetime-legacy.sql.out @@ -1497,9 +1497,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(TIMESTAMP '2011-11-11 11:11:11' + 1)\"", "left" : "\"TIMESTAMP\"", - "right" : "\"DOUBLE\"" + "right" : "\"DOUBLE\"", + "sqlExpr" : "\"(TIMESTAMP '2011-11-11 11:11:11' + 1)\"" }, "queryContext" : [ { "objectType" : "", @@ -1521,9 +1521,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(1 + TIMESTAMP '2011-11-11 11:11:11')\"", "left" : "\"DOUBLE\"", - "right" : "\"TIMESTAMP\"" + "right" : "\"TIMESTAMP\"", + "sqlExpr" : "\"(1 + TIMESTAMP '2011-11-11 11:11:11')\"" }, "queryContext" : [ { "objectType" : "", @@ -1545,9 +1545,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(TIMESTAMP '2011-11-11 11:11:11' + NULL)\"", "left" : "\"TIMESTAMP\"", - "right" : "\"VOID\"" + "right" : "\"VOID\"", + "sqlExpr" : "\"(TIMESTAMP '2011-11-11 11:11:11' + NULL)\"" }, "queryContext" : [ { "objectType" : "", @@ -1569,9 +1569,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(NULL + TIMESTAMP '2011-11-11 11:11:11')\"", "left" : "\"VOID\"", - "right" : "\"TIMESTAMP\"" + "right" : "\"TIMESTAMP\"", + "sqlExpr" : "\"(NULL + TIMESTAMP '2011-11-11 11:11:11')\"" }, "queryContext" : [ { "objectType" : "", diff --git a/sql/core/src/test/resources/sql-tests/results/datetime-parsing-invalid.sql.out b/sql/core/src/test/resources/sql-tests/results/datetime-parsing-invalid.sql.out index 79cf395abbabb..ef7626260b9f1 100644 --- a/sql/core/src/test/resources/sql-tests/results/datetime-parsing-invalid.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/datetime-parsing-invalid.sql.out @@ -18,8 +18,8 @@ org.apache.spark.SparkUpgradeException "errorClass" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION", "errorSubClass" : "PARSE_DATETIME_BY_NEW_PARSER", "messageParameters" : { - "datetime" : "'1'", - "config" : "\"spark.sql.legacy.timeParserPolicy\"" + "config" : "\"spark.sql.legacy.timeParserPolicy\"", + "datetime" : "'1'" } } @@ -42,8 +42,8 @@ org.apache.spark.SparkUpgradeException "errorClass" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION", "errorSubClass" : "PARSE_DATETIME_BY_NEW_PARSER", "messageParameters" : { - "datetime" : "'123'", - "config" : "\"spark.sql.legacy.timeParserPolicy\"" + "config" : "\"spark.sql.legacy.timeParserPolicy\"", + "datetime" : "'123'" } } @@ -58,8 +58,8 @@ org.apache.spark.SparkUpgradeException "errorClass" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION", "errorSubClass" : "PARSE_DATETIME_BY_NEW_PARSER", "messageParameters" : { - "datetime" : "'1'", - "config" : "\"spark.sql.legacy.timeParserPolicy\"" + "config" : "\"spark.sql.legacy.timeParserPolicy\"", + "datetime" : "'1'" } } @@ -74,8 +74,8 @@ org.apache.spark.SparkUpgradeException "errorClass" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION", "errorSubClass" : "DATETIME_PATTERN_RECOGNITION", "messageParameters" : { - "pattern" : "'yyyyyyy'", - "config" : "\"spark.sql.legacy.timeParserPolicy\"" + "config" : "\"spark.sql.legacy.timeParserPolicy\"", + "pattern" : "'yyyyyyy'" } } @@ -98,8 +98,8 @@ org.apache.spark.SparkUpgradeException "errorClass" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION", "errorSubClass" : "PARSE_DATETIME_BY_NEW_PARSER", "messageParameters" : { - "datetime" : "'9'", - "config" : "\"spark.sql.legacy.timeParserPolicy\"" + "config" : "\"spark.sql.legacy.timeParserPolicy\"", + "datetime" : "'9'" } } @@ -114,8 +114,8 @@ org.apache.spark.SparkUpgradeException "errorClass" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION", "errorSubClass" : "PARSE_DATETIME_BY_NEW_PARSER", "messageParameters" : { - "datetime" : "'9'", - "config" : "\"spark.sql.legacy.timeParserPolicy\"" + "config" : "\"spark.sql.legacy.timeParserPolicy\"", + "datetime" : "'9'" } } @@ -130,8 +130,8 @@ org.apache.spark.SparkUpgradeException "errorClass" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION", "errorSubClass" : "PARSE_DATETIME_BY_NEW_PARSER", "messageParameters" : { - "datetime" : "'99'", - "config" : "\"spark.sql.legacy.timeParserPolicy\"" + "config" : "\"spark.sql.legacy.timeParserPolicy\"", + "datetime" : "'99'" } } @@ -202,8 +202,8 @@ org.apache.spark.SparkUpgradeException "errorClass" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION", "errorSubClass" : "PARSE_DATETIME_BY_NEW_PARSER", "messageParameters" : { - "datetime" : "'2018-366'", - "config" : "\"spark.sql.legacy.timeParserPolicy\"" + "config" : "\"spark.sql.legacy.timeParserPolicy\"", + "datetime" : "'2018-366'" } } diff --git a/sql/core/src/test/resources/sql-tests/results/describe.sql.out b/sql/core/src/test/resources/sql-tests/results/describe.sql.out index 45df378495ccb..5bb86a398c82a 100644 --- a/sql/core/src/test/resources/sql-tests/results/describe.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/describe.sql.out @@ -474,9 +474,9 @@ org.apache.spark.sql.AnalysisException { "errorClass" : "FORBIDDEN_OPERATION", "messageParameters" : { - "statement" : "DESC PARTITION", + "objectName" : "`temp_v`", "objectType" : "TEMPORARY VIEW", - "objectName" : "`temp_v`" + "statement" : "DESC PARTITION" } } @@ -560,9 +560,9 @@ org.apache.spark.sql.AnalysisException { "errorClass" : "FORBIDDEN_OPERATION", "messageParameters" : { - "statement" : "DESC PARTITION", + "objectName" : "`v`", "objectType" : "VIEW", - "objectName" : "`v`" + "statement" : "DESC PARTITION" } } diff --git a/sql/core/src/test/resources/sql-tests/results/group-by-ordinal.sql.out b/sql/core/src/test/resources/sql-tests/results/group-by-ordinal.sql.out index 58bdf32e02eea..e21b79eb04697 100644 --- a/sql/core/src/test/resources/sql-tests/results/group-by-ordinal.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/group-by-ordinal.sql.out @@ -165,8 +165,8 @@ org.apache.spark.sql.AnalysisException "errorClass" : "GROUP_BY_POS_REFERS_AGG_EXPR", "sqlState" : "42000", "messageParameters" : { - "index" : "3", - "aggExpr" : "sum(data.b) AS `sum(b)`" + "aggExpr" : "sum(data.b) AS `sum(b)`", + "index" : "3" }, "queryContext" : [ { "objectType" : "", @@ -188,8 +188,8 @@ org.apache.spark.sql.AnalysisException "errorClass" : "GROUP_BY_POS_REFERS_AGG_EXPR", "sqlState" : "42000", "messageParameters" : { - "index" : "3", - "aggExpr" : "(sum(data.b) + CAST(2 AS BIGINT)) AS `(sum(b) + 2)`" + "aggExpr" : "(sum(data.b) + CAST(2 AS BIGINT)) AS `(sum(b) + 2)`", + "index" : "3" }, "queryContext" : [ { "objectType" : "", @@ -446,8 +446,8 @@ org.apache.spark.sql.AnalysisException "errorClass" : "GROUP_BY_POS_REFERS_AGG_EXPR", "sqlState" : "42000", "messageParameters" : { - "index" : "3", - "aggExpr" : "count(1) AS `count(1)`" + "aggExpr" : "count(1) AS `count(1)`", + "index" : "3" }, "queryContext" : [ { "objectType" : "", @@ -492,8 +492,8 @@ org.apache.spark.sql.AnalysisException "errorClass" : "GROUP_BY_POS_REFERS_AGG_EXPR", "sqlState" : "42000", "messageParameters" : { - "index" : "3", - "aggExpr" : "count(1) AS `count(1)`" + "aggExpr" : "count(1) AS `count(1)`", + "index" : "3" }, "queryContext" : [ { "objectType" : "", diff --git a/sql/core/src/test/resources/sql-tests/results/having.sql.out b/sql/core/src/test/resources/sql-tests/results/having.sql.out index 098f7bb3e257d..00b59827fde9b 100644 --- a/sql/core/src/test/resources/sql-tests/results/having.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/having.sql.out @@ -39,9 +39,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(v = array(1))\"", "left" : "\"INT\"", - "right" : "\"ARRAY\"" + "right" : "\"ARRAY\"", + "sqlExpr" : "\"(v = array(1))\"" }, "queryContext" : [ { "objectType" : "", diff --git a/sql/core/src/test/resources/sql-tests/results/interval.sql.out b/sql/core/src/test/resources/sql-tests/results/interval.sql.out index e75ceb3ff7d3b..39fcd3aba41f2 100644 --- a/sql/core/src/test/resources/sql-tests/results/interval.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/interval.sql.out @@ -187,9 +187,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(2 / INTERVAL '02' SECOND)\"", "left" : "\"DOUBLE\"", - "right" : "\"INTERVAL SECOND\"" + "right" : "\"INTERVAL SECOND\"", + "sqlExpr" : "\"(2 / INTERVAL '02' SECOND)\"" }, "queryContext" : [ { "objectType" : "", @@ -211,9 +211,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(2 / INTERVAL '2' YEAR)\"", "left" : "\"DOUBLE\"", - "right" : "\"INTERVAL YEAR\"" + "right" : "\"INTERVAL YEAR\"", + "sqlExpr" : "\"(2 / INTERVAL '2' YEAR)\"" }, "queryContext" : [ { "objectType" : "", @@ -321,9 +321,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(2 / INTERVAL '2' YEAR)\"", "left" : "\"INT\"", - "right" : "\"INTERVAL YEAR\"" + "right" : "\"INTERVAL YEAR\"", + "sqlExpr" : "\"(2 / INTERVAL '2' YEAR)\"" }, "queryContext" : [ { "objectType" : "", @@ -345,9 +345,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(2 / INTERVAL '02' HOUR)\"", "left" : "\"INT\"", - "right" : "\"INTERVAL HOUR\"" + "right" : "\"INTERVAL HOUR\"", + "sqlExpr" : "\"(2 / INTERVAL '02' HOUR)\"" }, "queryContext" : [ { "objectType" : "", @@ -369,9 +369,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(NULL / INTERVAL '2' YEAR)\"", "left" : "\"VOID\"", - "right" : "\"INTERVAL YEAR\"" + "right" : "\"INTERVAL YEAR\"", + "sqlExpr" : "\"(NULL / INTERVAL '2' YEAR)\"" }, "queryContext" : [ { "objectType" : "", @@ -393,9 +393,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(NULL / INTERVAL '02' HOUR)\"", "left" : "\"VOID\"", - "right" : "\"INTERVAL HOUR\"" + "right" : "\"INTERVAL HOUR\"", + "sqlExpr" : "\"(NULL / INTERVAL '02' HOUR)\"" }, "queryContext" : [ { "objectType" : "", @@ -1554,9 +1554,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(INTERVAL '2' YEAR + 3-3 year to month)\"", "left" : "\"INTERVAL YEAR\"", - "right" : "\"DOUBLE\"" + "right" : "\"DOUBLE\"", + "sqlExpr" : "\"(INTERVAL '2' YEAR + 3-3 year to month)\"" }, "queryContext" : [ { "objectType" : "", @@ -1594,9 +1594,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(INTERVAL '2' YEAR + 3-3)\"", "left" : "\"INTERVAL YEAR\"", - "right" : "\"DOUBLE\"" + "right" : "\"DOUBLE\"", + "sqlExpr" : "\"(INTERVAL '2' YEAR + 3-3)\"" }, "queryContext" : [ { "objectType" : "", @@ -1618,9 +1618,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(INTERVAL '2' YEAR - 4)\"", "left" : "\"INTERVAL YEAR\"", - "right" : "\"DOUBLE\"" + "right" : "\"DOUBLE\"", + "sqlExpr" : "\"(INTERVAL '2' YEAR - 4)\"" }, "queryContext" : [ { "objectType" : "", @@ -1666,9 +1666,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(INTERVAL '2' YEAR + str)\"", "left" : "\"INTERVAL YEAR\"", - "right" : "\"DOUBLE\"" + "right" : "\"DOUBLE\"", + "sqlExpr" : "\"(INTERVAL '2' YEAR + str)\"" }, "queryContext" : [ { "objectType" : "", @@ -1690,9 +1690,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(INTERVAL '2' YEAR - str)\"", "left" : "\"INTERVAL YEAR\"", - "right" : "\"DOUBLE\"" + "right" : "\"DOUBLE\"", + "sqlExpr" : "\"(INTERVAL '2' YEAR - str)\"" }, "queryContext" : [ { "objectType" : "", @@ -1757,9 +1757,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(INTERVAL '3' DAY - INTERVAL '2-2' YEAR TO MONTH)\"", "left" : "\"INTERVAL DAY\"", - "right" : "\"INTERVAL YEAR TO MONTH\"" + "right" : "\"INTERVAL YEAR TO MONTH\"", + "sqlExpr" : "\"(INTERVAL '3' DAY - INTERVAL '2-2' YEAR TO MONTH)\"" }, "queryContext" : [ { "objectType" : "", @@ -1790,9 +1790,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(1 + INTERVAL '2' MONTH)\"", "left" : "\"INT\"", - "right" : "\"INTERVAL MONTH\"" + "right" : "\"INTERVAL MONTH\"", + "sqlExpr" : "\"(1 + INTERVAL '2' MONTH)\"" }, "queryContext" : [ { "objectType" : "", @@ -1823,9 +1823,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(INTERVAL '2' MONTH - 1)\"", "left" : "\"INTERVAL MONTH\"", - "right" : "\"INT\"" + "right" : "\"INT\"", + "sqlExpr" : "\"(INTERVAL '2' MONTH - 1)\"" }, "queryContext" : [ { "objectType" : "", @@ -1974,8 +1974,8 @@ org.apache.spark.SparkArithmeticException "errorClass" : "INTERVAL_ARITHMETIC_OVERFLOW", "sqlState" : "22003", "messageParameters" : { - "message" : "integer overflow", - "alternative" : "" + "alternative" : "", + "message" : "integer overflow" } } @@ -1990,8 +1990,8 @@ org.apache.spark.SparkArithmeticException "errorClass" : "INTERVAL_ARITHMETIC_OVERFLOW", "sqlState" : "22003", "messageParameters" : { - "message" : "integer overflow", - "alternative" : " Use 'try_subtract' to tolerate overflow and return NULL instead." + "alternative" : " Use 'try_subtract' to tolerate overflow and return NULL instead.", + "message" : "integer overflow" } } @@ -2006,8 +2006,8 @@ org.apache.spark.SparkArithmeticException "errorClass" : "INTERVAL_ARITHMETIC_OVERFLOW", "sqlState" : "22003", "messageParameters" : { - "message" : "integer overflow", - "alternative" : " Use 'try_add' to tolerate overflow and return NULL instead." + "alternative" : " Use 'try_add' to tolerate overflow and return NULL instead.", + "message" : "integer overflow" } } @@ -2242,8 +2242,8 @@ org.apache.spark.SparkArithmeticException "errorClass" : "INTERVAL_ARITHMETIC_OVERFLOW", "sqlState" : "22003", "messageParameters" : { - "message" : "Interval value overflows after being divided by -1", - "alternative" : " Use 'try_divide' to tolerate overflow and return NULL instead." + "alternative" : " Use 'try_divide' to tolerate overflow and return NULL instead.", + "message" : "Interval value overflows after being divided by -1" }, "queryContext" : [ { "objectType" : "", @@ -2265,8 +2265,8 @@ org.apache.spark.SparkArithmeticException "errorClass" : "INTERVAL_ARITHMETIC_OVERFLOW", "sqlState" : "22003", "messageParameters" : { - "message" : "Interval value overflows after being divided by -1", - "alternative" : " Use 'try_divide' to tolerate overflow and return NULL instead." + "alternative" : " Use 'try_divide' to tolerate overflow and return NULL instead.", + "message" : "Interval value overflows after being divided by -1" }, "queryContext" : [ { "objectType" : "", @@ -2322,8 +2322,8 @@ org.apache.spark.SparkArithmeticException "errorClass" : "INTERVAL_ARITHMETIC_OVERFLOW", "sqlState" : "22003", "messageParameters" : { - "message" : "Interval value overflows after being divided by -1", - "alternative" : " Use 'try_divide' to tolerate overflow and return NULL instead." + "alternative" : " Use 'try_divide' to tolerate overflow and return NULL instead.", + "message" : "Interval value overflows after being divided by -1" }, "queryContext" : [ { "objectType" : "", @@ -2345,8 +2345,8 @@ org.apache.spark.SparkArithmeticException "errorClass" : "INTERVAL_ARITHMETIC_OVERFLOW", "sqlState" : "22003", "messageParameters" : { - "message" : "Interval value overflows after being divided by -1", - "alternative" : " Use 'try_divide' to tolerate overflow and return NULL instead." + "alternative" : " Use 'try_divide' to tolerate overflow and return NULL instead.", + "message" : "Interval value overflows after being divided by -1" }, "queryContext" : [ { "objectType" : "", @@ -2694,9 +2694,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(INTERVAL '1' MONTH > INTERVAL '20' DAY)\"", "left" : "\"INTERVAL MONTH\"", - "right" : "\"INTERVAL DAY\"" + "right" : "\"INTERVAL DAY\"", + "sqlExpr" : "\"(INTERVAL '1' MONTH > INTERVAL '20' DAY)\"" }, "queryContext" : [ { "objectType" : "", @@ -2718,9 +2718,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(INTERVAL '1' DAY < 1)\"", "left" : "\"INTERVAL DAY\"", - "right" : "\"STRING\"" + "right" : "\"STRING\"", + "sqlExpr" : "\"(INTERVAL '1' DAY < 1)\"" }, "queryContext" : [ { "objectType" : "", @@ -2742,9 +2742,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(INTERVAL '1' DAY = 1)\"", "left" : "\"INTERVAL DAY\"", - "right" : "\"STRING\"" + "right" : "\"STRING\"", + "sqlExpr" : "\"(INTERVAL '1' DAY = 1)\"" }, "queryContext" : [ { "objectType" : "", @@ -2766,9 +2766,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(INTERVAL '1' DAY > 1)\"", "left" : "\"INTERVAL DAY\"", - "right" : "\"STRING\"" + "right" : "\"STRING\"", + "sqlExpr" : "\"(INTERVAL '1' DAY > 1)\"" }, "queryContext" : [ { "objectType" : "", @@ -2790,9 +2790,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(1 < INTERVAL '1' DAY)\"", "left" : "\"STRING\"", - "right" : "\"INTERVAL DAY\"" + "right" : "\"INTERVAL DAY\"", + "sqlExpr" : "\"(1 < INTERVAL '1' DAY)\"" }, "queryContext" : [ { "objectType" : "", @@ -2814,9 +2814,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(1 = INTERVAL '1' DAY)\"", "left" : "\"STRING\"", - "right" : "\"INTERVAL DAY\"" + "right" : "\"INTERVAL DAY\"", + "sqlExpr" : "\"(1 = INTERVAL '1' DAY)\"" }, "queryContext" : [ { "objectType" : "", @@ -2838,9 +2838,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(1 > INTERVAL '1' DAY)\"", "left" : "\"STRING\"", - "right" : "\"INTERVAL DAY\"" + "right" : "\"INTERVAL DAY\"", + "sqlExpr" : "\"(1 > INTERVAL '1' DAY)\"" }, "queryContext" : [ { "objectType" : "", @@ -2862,9 +2862,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(INTERVAL '1' YEAR < 1)\"", "left" : "\"INTERVAL YEAR\"", - "right" : "\"STRING\"" + "right" : "\"STRING\"", + "sqlExpr" : "\"(INTERVAL '1' YEAR < 1)\"" }, "queryContext" : [ { "objectType" : "", @@ -2886,9 +2886,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(INTERVAL '1' YEAR = 1)\"", "left" : "\"INTERVAL YEAR\"", - "right" : "\"STRING\"" + "right" : "\"STRING\"", + "sqlExpr" : "\"(INTERVAL '1' YEAR = 1)\"" }, "queryContext" : [ { "objectType" : "", @@ -2910,9 +2910,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(INTERVAL '1' YEAR > 1)\"", "left" : "\"INTERVAL YEAR\"", - "right" : "\"STRING\"" + "right" : "\"STRING\"", + "sqlExpr" : "\"(INTERVAL '1' YEAR > 1)\"" }, "queryContext" : [ { "objectType" : "", @@ -2934,9 +2934,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(1 < INTERVAL '1' YEAR)\"", "left" : "\"STRING\"", - "right" : "\"INTERVAL YEAR\"" + "right" : "\"INTERVAL YEAR\"", + "sqlExpr" : "\"(1 < INTERVAL '1' YEAR)\"" }, "queryContext" : [ { "objectType" : "", @@ -2958,9 +2958,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(1 = INTERVAL '1' YEAR)\"", "left" : "\"STRING\"", - "right" : "\"INTERVAL YEAR\"" + "right" : "\"INTERVAL YEAR\"", + "sqlExpr" : "\"(1 = INTERVAL '1' YEAR)\"" }, "queryContext" : [ { "objectType" : "", @@ -2982,9 +2982,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(1 > INTERVAL '1' YEAR)\"", "left" : "\"STRING\"", - "right" : "\"INTERVAL YEAR\"" + "right" : "\"INTERVAL YEAR\"", + "sqlExpr" : "\"(1 > INTERVAL '1' YEAR)\"" }, "queryContext" : [ { "objectType" : "", @@ -3104,9 +3104,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(INTERVAL '1' MONTH div INTERVAL '-1' DAY)\"", "left" : "\"INTERVAL MONTH\"", - "right" : "\"INTERVAL DAY\"" + "right" : "\"INTERVAL DAY\"", + "sqlExpr" : "\"(INTERVAL '1' MONTH div INTERVAL '-1' DAY)\"" }, "queryContext" : [ { "objectType" : "", diff --git a/sql/core/src/test/resources/sql-tests/results/json-functions.sql.out b/sql/core/src/test/resources/sql-tests/results/json-functions.sql.out index ca079e4add048..97b219589a4f2 100644 --- a/sql/core/src/test/resources/sql-tests/results/json-functions.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/json-functions.sql.out @@ -340,8 +340,8 @@ org.apache.spark.SparkUpgradeException "errorClass" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION", "errorSubClass" : "PARSE_DATETIME_BY_NEW_PARSER", "messageParameters" : { - "datetime" : "'02-29'", - "config" : "\"spark.sql.legacy.timeParserPolicy\"" + "config" : "\"spark.sql.legacy.timeParserPolicy\"", + "datetime" : "'02-29'" } } @@ -359,8 +359,8 @@ org.apache.spark.SparkUpgradeException "errorClass" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION", "errorSubClass" : "PARSE_DATETIME_BY_NEW_PARSER", "messageParameters" : { - "datetime" : "'02-29'", - "config" : "\"spark.sql.legacy.timeParserPolicy\"" + "config" : "\"spark.sql.legacy.timeParserPolicy\"", + "datetime" : "'02-29'" } } diff --git a/sql/core/src/test/resources/sql-tests/results/pivot.sql.out b/sql/core/src/test/resources/sql-tests/results/pivot.sql.out index d5f260d3acea2..b25362aba1efd 100644 --- a/sql/core/src/test/resources/sql-tests/results/pivot.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/pivot.sql.out @@ -329,9 +329,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "PIVOT_VALUE_DATA_TYPE_MISMATCH", "sqlState" : "42000", "messageParameters" : { + "pivotType" : "struct", "value" : "dotNET", - "valueType" : "string", - "pivotType" : "struct" + "valueType" : "string" } } diff --git a/sql/core/src/test/resources/sql-tests/results/postgreSQL/boolean.sql.out b/sql/core/src/test/resources/sql-tests/results/postgreSQL/boolean.sql.out index 2ef04e1669232..371ae63424f48 100644 --- a/sql/core/src/test/resources/sql-tests/results/postgreSQL/boolean.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/postgreSQL/boolean.sql.out @@ -57,10 +57,10 @@ org.apache.spark.SparkRuntimeException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "42000", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'test'", "sourceType" : "\"STRING\"", - "targetType" : "\"BOOLEAN\"", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "targetType" : "\"BOOLEAN\"" }, "queryContext" : [ { "objectType" : "", @@ -90,10 +90,10 @@ org.apache.spark.SparkRuntimeException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "42000", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'foo'", "sourceType" : "\"STRING\"", - "targetType" : "\"BOOLEAN\"", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "targetType" : "\"BOOLEAN\"" }, "queryContext" : [ { "objectType" : "", @@ -131,10 +131,10 @@ org.apache.spark.SparkRuntimeException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "42000", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'yeah'", "sourceType" : "\"STRING\"", - "targetType" : "\"BOOLEAN\"", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "targetType" : "\"BOOLEAN\"" }, "queryContext" : [ { "objectType" : "", @@ -172,10 +172,10 @@ org.apache.spark.SparkRuntimeException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "42000", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'nay'", "sourceType" : "\"STRING\"", - "targetType" : "\"BOOLEAN\"", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "targetType" : "\"BOOLEAN\"" }, "queryContext" : [ { "objectType" : "", @@ -197,10 +197,10 @@ org.apache.spark.SparkRuntimeException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "42000", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'on'", "sourceType" : "\"STRING\"", - "targetType" : "\"BOOLEAN\"", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "targetType" : "\"BOOLEAN\"" }, "queryContext" : [ { "objectType" : "", @@ -222,10 +222,10 @@ org.apache.spark.SparkRuntimeException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "42000", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'off'", "sourceType" : "\"STRING\"", - "targetType" : "\"BOOLEAN\"", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "targetType" : "\"BOOLEAN\"" }, "queryContext" : [ { "objectType" : "", @@ -247,10 +247,10 @@ org.apache.spark.SparkRuntimeException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "42000", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'of'", "sourceType" : "\"STRING\"", - "targetType" : "\"BOOLEAN\"", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "targetType" : "\"BOOLEAN\"" }, "queryContext" : [ { "objectType" : "", @@ -272,10 +272,10 @@ org.apache.spark.SparkRuntimeException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "42000", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'o'", "sourceType" : "\"STRING\"", - "targetType" : "\"BOOLEAN\"", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "targetType" : "\"BOOLEAN\"" }, "queryContext" : [ { "objectType" : "", @@ -297,10 +297,10 @@ org.apache.spark.SparkRuntimeException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "42000", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'on_'", "sourceType" : "\"STRING\"", - "targetType" : "\"BOOLEAN\"", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "targetType" : "\"BOOLEAN\"" }, "queryContext" : [ { "objectType" : "", @@ -322,10 +322,10 @@ org.apache.spark.SparkRuntimeException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "42000", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'off_'", "sourceType" : "\"STRING\"", - "targetType" : "\"BOOLEAN\"", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "targetType" : "\"BOOLEAN\"" }, "queryContext" : [ { "objectType" : "", @@ -355,10 +355,10 @@ org.apache.spark.SparkRuntimeException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "42000", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'11'", "sourceType" : "\"STRING\"", - "targetType" : "\"BOOLEAN\"", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "targetType" : "\"BOOLEAN\"" }, "queryContext" : [ { "objectType" : "", @@ -388,10 +388,10 @@ org.apache.spark.SparkRuntimeException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "42000", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'000'", "sourceType" : "\"STRING\"", - "targetType" : "\"BOOLEAN\"", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "targetType" : "\"BOOLEAN\"" }, "queryContext" : [ { "objectType" : "", @@ -413,10 +413,10 @@ org.apache.spark.SparkRuntimeException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "42000", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "''", "sourceType" : "\"STRING\"", - "targetType" : "\"BOOLEAN\"", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "targetType" : "\"BOOLEAN\"" }, "queryContext" : [ { "objectType" : "", @@ -535,10 +535,10 @@ org.apache.spark.SparkRuntimeException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "42000", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "' tru e '", "sourceType" : "\"STRING\"", - "targetType" : "\"BOOLEAN\"", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "targetType" : "\"BOOLEAN\"" }, "queryContext" : [ { "objectType" : "", @@ -560,10 +560,10 @@ org.apache.spark.SparkRuntimeException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "42000", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "''", "sourceType" : "\"STRING\"", - "targetType" : "\"BOOLEAN\"", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "targetType" : "\"BOOLEAN\"" }, "queryContext" : [ { "objectType" : "", diff --git a/sql/core/src/test/resources/sql-tests/results/postgreSQL/float4.sql.out b/sql/core/src/test/resources/sql-tests/results/postgreSQL/float4.sql.out index 673252707709c..a2fb4276f6e37 100644 --- a/sql/core/src/test/resources/sql-tests/results/postgreSQL/float4.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/postgreSQL/float4.sql.out @@ -97,10 +97,10 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "42000", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'N A N'", "sourceType" : "\"STRING\"", - "targetType" : "\"FLOAT\"", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "targetType" : "\"FLOAT\"" }, "queryContext" : [ { "objectType" : "", @@ -122,10 +122,10 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "42000", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'NaN x'", "sourceType" : "\"STRING\"", - "targetType" : "\"FLOAT\"", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "targetType" : "\"FLOAT\"" }, "queryContext" : [ { "objectType" : "", @@ -147,10 +147,10 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "42000", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "' INFINITY x'", "sourceType" : "\"STRING\"", - "targetType" : "\"FLOAT\"", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "targetType" : "\"FLOAT\"" }, "queryContext" : [ { "objectType" : "", @@ -196,10 +196,10 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "42000", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'nan'", "sourceType" : "\"STRING\"", - "targetType" : "\"DECIMAL(10,0)\"", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "targetType" : "\"DECIMAL(10,0)\"" }, "queryContext" : [ { "objectType" : "", @@ -393,10 +393,10 @@ org.apache.spark.SparkArithmeticException "errorClass" : "CAST_OVERFLOW", "sqlState" : "22005", "messageParameters" : { - "value" : "2.14748365E9", + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "sourceType" : "\"FLOAT\"", "targetType" : "\"INT\"", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "value" : "2.14748365E9" } } @@ -419,10 +419,10 @@ org.apache.spark.SparkArithmeticException "errorClass" : "CAST_OVERFLOW", "sqlState" : "22005", "messageParameters" : { - "value" : "-2.1474839E9", + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "sourceType" : "\"FLOAT\"", "targetType" : "\"INT\"", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "value" : "-2.1474839E9" } } @@ -461,10 +461,10 @@ org.apache.spark.SparkArithmeticException "errorClass" : "CAST_OVERFLOW", "sqlState" : "22005", "messageParameters" : { - "value" : "-9.22338E18", + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "sourceType" : "\"FLOAT\"", "targetType" : "\"BIGINT\"", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "value" : "-9.22338E18" } } diff --git a/sql/core/src/test/resources/sql-tests/results/postgreSQL/float8.sql.out b/sql/core/src/test/resources/sql-tests/results/postgreSQL/float8.sql.out index a8dbeb13b2d89..a9f7bdc925c4a 100644 --- a/sql/core/src/test/resources/sql-tests/results/postgreSQL/float8.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/postgreSQL/float8.sql.out @@ -129,10 +129,10 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "42000", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'N A N'", "sourceType" : "\"STRING\"", - "targetType" : "\"DOUBLE\"", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "targetType" : "\"DOUBLE\"" }, "queryContext" : [ { "objectType" : "", @@ -154,10 +154,10 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "42000", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'NaN x'", "sourceType" : "\"STRING\"", - "targetType" : "\"DOUBLE\"", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "targetType" : "\"DOUBLE\"" }, "queryContext" : [ { "objectType" : "", @@ -179,10 +179,10 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "42000", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "' INFINITY x'", "sourceType" : "\"STRING\"", - "targetType" : "\"DOUBLE\"", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "targetType" : "\"DOUBLE\"" }, "queryContext" : [ { "objectType" : "", @@ -228,10 +228,10 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "42000", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'nan'", "sourceType" : "\"STRING\"", - "targetType" : "\"DECIMAL(10,0)\"", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "targetType" : "\"DECIMAL(10,0)\"" }, "queryContext" : [ { "objectType" : "", @@ -898,10 +898,10 @@ org.apache.spark.SparkArithmeticException "errorClass" : "CAST_OVERFLOW", "sqlState" : "22005", "messageParameters" : { - "value" : "-9.22337203685478E18D", + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "sourceType" : "\"DOUBLE\"", "targetType" : "\"BIGINT\"", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "value" : "-9.22337203685478E18D" } } diff --git a/sql/core/src/test/resources/sql-tests/results/postgreSQL/int4.sql.out b/sql/core/src/test/resources/sql-tests/results/postgreSQL/int4.sql.out index 7b2382a4f5710..16c18c86f2919 100755 --- a/sql/core/src/test/resources/sql-tests/results/postgreSQL/int4.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/postgreSQL/int4.sql.out @@ -201,9 +201,9 @@ org.apache.spark.SparkArithmeticException "errorClass" : "ARITHMETIC_OVERFLOW", "sqlState" : "22003", "messageParameters" : { - "message" : "integer overflow", "alternative" : " Use 'try_multiply' to tolerate overflow and return NULL instead.", - "config" : "\"spark.sql.ansi.enabled\"" + "config" : "\"spark.sql.ansi.enabled\"", + "message" : "integer overflow" }, "queryContext" : [ { "objectType" : "", @@ -236,9 +236,9 @@ org.apache.spark.SparkArithmeticException "errorClass" : "ARITHMETIC_OVERFLOW", "sqlState" : "22003", "messageParameters" : { - "message" : "integer overflow", "alternative" : " Use 'try_multiply' to tolerate overflow and return NULL instead.", - "config" : "\"spark.sql.ansi.enabled\"" + "config" : "\"spark.sql.ansi.enabled\"", + "message" : "integer overflow" }, "queryContext" : [ { "objectType" : "", @@ -271,9 +271,9 @@ org.apache.spark.SparkArithmeticException "errorClass" : "ARITHMETIC_OVERFLOW", "sqlState" : "22003", "messageParameters" : { - "message" : "integer overflow", "alternative" : " Use 'try_add' to tolerate overflow and return NULL instead.", - "config" : "\"spark.sql.ansi.enabled\"" + "config" : "\"spark.sql.ansi.enabled\"", + "message" : "integer overflow" }, "queryContext" : [ { "objectType" : "", @@ -307,9 +307,9 @@ org.apache.spark.SparkArithmeticException "errorClass" : "ARITHMETIC_OVERFLOW", "sqlState" : "22003", "messageParameters" : { - "message" : "integer overflow", "alternative" : " Use 'try_add' to tolerate overflow and return NULL instead.", - "config" : "\"spark.sql.ansi.enabled\"" + "config" : "\"spark.sql.ansi.enabled\"", + "message" : "integer overflow" }, "queryContext" : [ { "objectType" : "", @@ -343,9 +343,9 @@ org.apache.spark.SparkArithmeticException "errorClass" : "ARITHMETIC_OVERFLOW", "sqlState" : "22003", "messageParameters" : { - "message" : "integer overflow", "alternative" : " Use 'try_subtract' to tolerate overflow and return NULL instead.", - "config" : "\"spark.sql.ansi.enabled\"" + "config" : "\"spark.sql.ansi.enabled\"", + "message" : "integer overflow" }, "queryContext" : [ { "objectType" : "", @@ -379,9 +379,9 @@ org.apache.spark.SparkArithmeticException "errorClass" : "ARITHMETIC_OVERFLOW", "sqlState" : "22003", "messageParameters" : { - "message" : "integer overflow", "alternative" : " Use 'try_subtract' to tolerate overflow and return NULL instead.", - "config" : "\"spark.sql.ansi.enabled\"" + "config" : "\"spark.sql.ansi.enabled\"", + "message" : "integer overflow" }, "queryContext" : [ { "objectType" : "", diff --git a/sql/core/src/test/resources/sql-tests/results/postgreSQL/int8.sql.out b/sql/core/src/test/resources/sql-tests/results/postgreSQL/int8.sql.out index 39a8a2441afee..d1974a1c0d8ee 100755 --- a/sql/core/src/test/resources/sql-tests/results/postgreSQL/int8.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/postgreSQL/int8.sql.out @@ -393,9 +393,9 @@ org.apache.spark.SparkArithmeticException "errorClass" : "ARITHMETIC_OVERFLOW", "sqlState" : "22003", "messageParameters" : { - "message" : "long overflow", "alternative" : " Use 'try_multiply' to tolerate overflow and return NULL instead.", - "config" : "\"spark.sql.ansi.enabled\"" + "config" : "\"spark.sql.ansi.enabled\"", + "message" : "long overflow" }, "queryContext" : [ { "objectType" : "", @@ -737,10 +737,10 @@ org.apache.spark.SparkArithmeticException "errorClass" : "CAST_OVERFLOW", "sqlState" : "22005", "messageParameters" : { - "value" : "4567890123456789L", + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "sourceType" : "\"BIGINT\"", "targetType" : "\"INT\"", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "value" : "4567890123456789L" } } @@ -763,10 +763,10 @@ org.apache.spark.SparkArithmeticException "errorClass" : "CAST_OVERFLOW", "sqlState" : "22005", "messageParameters" : { - "value" : "4567890123456789L", + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "sourceType" : "\"BIGINT\"", "targetType" : "\"SMALLINT\"", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "value" : "4567890123456789L" } } @@ -809,10 +809,10 @@ org.apache.spark.SparkArithmeticException "errorClass" : "CAST_OVERFLOW", "sqlState" : "22005", "messageParameters" : { - "value" : "9.223372036854776E20D", + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "sourceType" : "\"DOUBLE\"", "targetType" : "\"BIGINT\"", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "value" : "9.223372036854776E20D" } } @@ -890,10 +890,10 @@ org.apache.spark.SparkArithmeticException "errorClass" : "CAST_OVERFLOW", "sqlState" : "22005", "messageParameters" : { - "value" : "-9223372036854775808L", + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "sourceType" : "\"BIGINT\"", "targetType" : "\"INT\"", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "value" : "-9223372036854775808L" } } @@ -908,9 +908,9 @@ org.apache.spark.SparkArithmeticException "errorClass" : "ARITHMETIC_OVERFLOW", "sqlState" : "22003", "messageParameters" : { - "message" : "long overflow", "alternative" : " Use 'try_multiply' to tolerate overflow and return NULL instead.", - "config" : "\"spark.sql.ansi.enabled\"" + "config" : "\"spark.sql.ansi.enabled\"", + "message" : "long overflow" }, "queryContext" : [ { "objectType" : "", @@ -948,9 +948,9 @@ org.apache.spark.SparkArithmeticException "errorClass" : "ARITHMETIC_OVERFLOW", "sqlState" : "22003", "messageParameters" : { - "message" : "long overflow", "alternative" : " Use 'try_multiply' to tolerate overflow and return NULL instead.", - "config" : "\"spark.sql.ansi.enabled\"" + "config" : "\"spark.sql.ansi.enabled\"", + "message" : "long overflow" }, "queryContext" : [ { "objectType" : "", @@ -988,9 +988,9 @@ org.apache.spark.SparkArithmeticException "errorClass" : "ARITHMETIC_OVERFLOW", "sqlState" : "22003", "messageParameters" : { - "message" : "long overflow", "alternative" : " Use 'try_multiply' to tolerate overflow and return NULL instead.", - "config" : "\"spark.sql.ansi.enabled\"" + "config" : "\"spark.sql.ansi.enabled\"", + "message" : "long overflow" }, "queryContext" : [ { "objectType" : "", diff --git a/sql/core/src/test/resources/sql-tests/results/postgreSQL/text.sql.out b/sql/core/src/test/resources/sql-tests/results/postgreSQL/text.sql.out index 1067fb6ee18d8..4867410806d3b 100755 --- a/sql/core/src/test/resources/sql-tests/results/postgreSQL/text.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/postgreSQL/text.sql.out @@ -66,10 +66,10 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "42000", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'four: 2'", "sourceType" : "\"STRING\"", - "targetType" : "\"BIGINT\"", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "targetType" : "\"BIGINT\"" }, "queryContext" : [ { "objectType" : "", @@ -91,10 +91,10 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "42000", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'four: 2'", "sourceType" : "\"STRING\"", - "targetType" : "\"BIGINT\"", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "targetType" : "\"BIGINT\"" }, "queryContext" : [ { "objectType" : "", @@ -312,9 +312,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "INVALID_PARAMETER_VALUE", "sqlState" : "22023", "messageParameters" : { - "parameter" : "strfmt", + "expected" : "expects %1$, %2$ and so on, but got %0$.", "functionName" : "`format_string`", - "expected" : "expects %1$, %2$ and so on, but got %0$." + "parameter" : "strfmt" }, "queryContext" : [ { "objectType" : "", diff --git a/sql/core/src/test/resources/sql-tests/results/postgreSQL/window_part2.sql.out b/sql/core/src/test/resources/sql-tests/results/postgreSQL/window_part2.sql.out index 7289de2c0c57b..f70d694080387 100644 --- a/sql/core/src/test/resources/sql-tests/results/postgreSQL/window_part2.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/postgreSQL/window_part2.sql.out @@ -226,9 +226,9 @@ org.apache.spark.SparkArithmeticException "errorClass" : "ARITHMETIC_OVERFLOW", "sqlState" : "22003", "messageParameters" : { - "message" : "long overflow", "alternative" : " Use 'try_add' to tolerate overflow and return NULL instead.", - "config" : "\"spark.sql.ansi.enabled\"" + "config" : "\"spark.sql.ansi.enabled\"", + "message" : "long overflow" }, "queryContext" : [ { "objectType" : "", @@ -249,9 +249,9 @@ org.apache.spark.SparkArithmeticException "errorClass" : "ARITHMETIC_OVERFLOW", "sqlState" : "22003", "messageParameters" : { - "message" : "long overflow", "alternative" : " Use 'try_add' to tolerate overflow and return NULL instead.", - "config" : "\"spark.sql.ansi.enabled\"" + "config" : "\"spark.sql.ansi.enabled\"", + "message" : "long overflow" }, "queryContext" : [ { "objectType" : "", @@ -489,10 +489,10 @@ org.apache.spark.SparkNumberFormatException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "42000", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'NaN'", "sourceType" : "\"STRING\"", - "targetType" : "\"INT\"", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "targetType" : "\"INT\"" }, "queryContext" : [ { "objectType" : "", diff --git a/sql/core/src/test/resources/sql-tests/results/regexp-functions.sql.out b/sql/core/src/test/resources/sql-tests/results/regexp-functions.sql.out index 65e1e31ae7cf3..9ef1802738a1f 100644 --- a/sql/core/src/test/resources/sql-tests/results/regexp-functions.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/regexp-functions.sql.out @@ -135,9 +135,9 @@ org.apache.spark.SparkRuntimeException "errorClass" : "INVALID_PARAMETER_VALUE", "sqlState" : "22023", "messageParameters" : { - "parameter" : "regexp", + "expected" : "(?l)", "functionName" : "`regexp_extract`", - "expected" : "(?l)" + "parameter" : "regexp" } } @@ -278,9 +278,9 @@ org.apache.spark.SparkRuntimeException "errorClass" : "INVALID_PARAMETER_VALUE", "sqlState" : "22023", "messageParameters" : { - "parameter" : "regexp", + "expected" : "], [", "functionName" : "`regexp_extract_all`", - "expected" : "], [" + "parameter" : "regexp" } } @@ -585,8 +585,8 @@ org.apache.spark.SparkRuntimeException "errorClass" : "INVALID_PARAMETER_VALUE", "sqlState" : "22023", "messageParameters" : { - "parameter" : "regexp", + "expected" : ") ?", "functionName" : "`regexp_instr`", - "expected" : ") ?" + "parameter" : "regexp" } } diff --git a/sql/core/src/test/resources/sql-tests/results/string-functions.sql.out b/sql/core/src/test/resources/sql-tests/results/string-functions.sql.out index a8ad802dd9899..8502c60ce6e27 100644 --- a/sql/core/src/test/resources/sql-tests/results/string-functions.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/string-functions.sql.out @@ -1119,10 +1119,10 @@ org.apache.spark.SparkIllegalArgumentException { "errorClass" : "CONVERSION_INVALID_INPUT", "messageParameters" : { - "str" : "' ab cdef= = '", "fmt" : "'BASE64'", - "targetType" : "\"BINARY\"", - "suggestion" : "`try_to_binary`" + "str" : "' ab cdef= = '", + "suggestion" : "`try_to_binary`", + "targetType" : "\"BINARY\"" } } @@ -1146,10 +1146,10 @@ org.apache.spark.SparkIllegalArgumentException { "errorClass" : "CONVERSION_INVALID_INPUT", "messageParameters" : { - "str" : "'a'", "fmt" : "'BASE64'", - "targetType" : "\"BINARY\"", - "suggestion" : "`try_to_binary`" + "str" : "'a'", + "suggestion" : "`try_to_binary`", + "targetType" : "\"BINARY\"" } } @@ -1163,10 +1163,10 @@ org.apache.spark.SparkIllegalArgumentException { "errorClass" : "CONVERSION_INVALID_INPUT", "messageParameters" : { - "str" : "'a?'", "fmt" : "'BASE64'", - "targetType" : "\"BINARY\"", - "suggestion" : "`try_to_binary`" + "str" : "'a?'", + "suggestion" : "`try_to_binary`", + "targetType" : "\"BINARY\"" } } @@ -1180,10 +1180,10 @@ org.apache.spark.SparkIllegalArgumentException { "errorClass" : "CONVERSION_INVALID_INPUT", "messageParameters" : { - "str" : "'abcde'", "fmt" : "'BASE64'", - "targetType" : "\"BINARY\"", - "suggestion" : "`try_to_binary`" + "str" : "'abcde'", + "suggestion" : "`try_to_binary`", + "targetType" : "\"BINARY\"" } } @@ -1197,10 +1197,10 @@ org.apache.spark.SparkIllegalArgumentException { "errorClass" : "CONVERSION_INVALID_INPUT", "messageParameters" : { - "str" : "'abcd='", "fmt" : "'BASE64'", - "targetType" : "\"BINARY\"", - "suggestion" : "`try_to_binary`" + "str" : "'abcd='", + "suggestion" : "`try_to_binary`", + "targetType" : "\"BINARY\"" } } @@ -1214,10 +1214,10 @@ org.apache.spark.SparkIllegalArgumentException { "errorClass" : "CONVERSION_INVALID_INPUT", "messageParameters" : { - "str" : "'a==='", "fmt" : "'BASE64'", - "targetType" : "\"BINARY\"", - "suggestion" : "`try_to_binary`" + "str" : "'a==='", + "suggestion" : "`try_to_binary`", + "targetType" : "\"BINARY\"" } } @@ -1231,10 +1231,10 @@ org.apache.spark.SparkIllegalArgumentException { "errorClass" : "CONVERSION_INVALID_INPUT", "messageParameters" : { - "str" : "'ab==f'", "fmt" : "'BASE64'", - "targetType" : "\"BINARY\"", - "suggestion" : "`try_to_binary`" + "str" : "'ab==f'", + "suggestion" : "`try_to_binary`", + "targetType" : "\"BINARY\"" } } @@ -1321,10 +1321,10 @@ org.apache.spark.SparkIllegalArgumentException { "errorClass" : "CONVERSION_INVALID_INPUT", "messageParameters" : { - "str" : "'GG'", "fmt" : "'HEX'", - "targetType" : "\"BINARY\"", - "suggestion" : "`try_to_binary`" + "str" : "'GG'", + "suggestion" : "`try_to_binary`", + "targetType" : "\"BINARY\"" } } @@ -1338,10 +1338,10 @@ org.apache.spark.SparkIllegalArgumentException { "errorClass" : "CONVERSION_INVALID_INPUT", "messageParameters" : { - "str" : "'01 AF'", "fmt" : "'HEX'", - "targetType" : "\"BINARY\"", - "suggestion" : "`try_to_binary`" + "str" : "'01 AF'", + "suggestion" : "`try_to_binary`", + "targetType" : "\"BINARY\"" } } @@ -1363,10 +1363,10 @@ org.apache.spark.SparkIllegalArgumentException { "errorClass" : "CONVERSION_INVALID_INPUT", "messageParameters" : { - "str" : "' ab cdef= = '", "fmt" : "'BASE64'", - "targetType" : "\"BINARY\"", - "suggestion" : "`try_to_binary`" + "str" : "' ab cdef= = '", + "suggestion" : "`try_to_binary`", + "targetType" : "\"BINARY\"" } } @@ -1380,10 +1380,10 @@ org.apache.spark.SparkIllegalArgumentException { "errorClass" : "CONVERSION_INVALID_INPUT", "messageParameters" : { - "str" : "' ab cdef= = '", "fmt" : "'HEX'", - "targetType" : "\"BINARY\"", - "suggestion" : "`try_to_binary`" + "str" : "' ab cdef= = '", + "suggestion" : "`try_to_binary`", + "targetType" : "\"BINARY\"" } } diff --git a/sql/core/src/test/resources/sql-tests/results/timestamp.sql.out b/sql/core/src/test/resources/sql-tests/results/timestamp.sql.out index 6cf116a49e0a3..d9042f0bbb2cd 100644 --- a/sql/core/src/test/resources/sql-tests/results/timestamp.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/timestamp.sql.out @@ -669,9 +669,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(TIMESTAMP '2011-11-11 11:11:11' + 1)\"", "left" : "\"TIMESTAMP\"", - "right" : "\"DOUBLE\"" + "right" : "\"DOUBLE\"", + "sqlExpr" : "\"(TIMESTAMP '2011-11-11 11:11:11' + 1)\"" }, "queryContext" : [ { "objectType" : "", @@ -693,9 +693,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(1 + TIMESTAMP '2011-11-11 11:11:11')\"", "left" : "\"DOUBLE\"", - "right" : "\"TIMESTAMP\"" + "right" : "\"TIMESTAMP\"", + "sqlExpr" : "\"(1 + TIMESTAMP '2011-11-11 11:11:11')\"" }, "queryContext" : [ { "objectType" : "", @@ -717,9 +717,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(TIMESTAMP '2011-11-11 11:11:11' + NULL)\"", "left" : "\"TIMESTAMP\"", - "right" : "\"VOID\"" + "right" : "\"VOID\"", + "sqlExpr" : "\"(TIMESTAMP '2011-11-11 11:11:11' + NULL)\"" }, "queryContext" : [ { "objectType" : "", @@ -741,9 +741,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(NULL + TIMESTAMP '2011-11-11 11:11:11')\"", "left" : "\"VOID\"", - "right" : "\"TIMESTAMP\"" + "right" : "\"TIMESTAMP\"", + "sqlExpr" : "\"(NULL + TIMESTAMP '2011-11-11 11:11:11')\"" }, "queryContext" : [ { "objectType" : "", @@ -788,8 +788,8 @@ org.apache.spark.SparkUpgradeException "errorClass" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION", "errorSubClass" : "DATETIME_PATTERN_RECOGNITION", "messageParameters" : { - "pattern" : "'yyyy-MM-dd GGGGG'", - "config" : "\"spark.sql.legacy.timeParserPolicy\"" + "config" : "\"spark.sql.legacy.timeParserPolicy\"", + "pattern" : "'yyyy-MM-dd GGGGG'" } } @@ -804,8 +804,8 @@ org.apache.spark.SparkUpgradeException "errorClass" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION", "errorSubClass" : "DATETIME_PATTERN_RECOGNITION", "messageParameters" : { - "pattern" : "'dd MM yyyy EEEEEE'", - "config" : "\"spark.sql.legacy.timeParserPolicy\"" + "config" : "\"spark.sql.legacy.timeParserPolicy\"", + "pattern" : "'dd MM yyyy EEEEEE'" } } @@ -820,8 +820,8 @@ org.apache.spark.SparkUpgradeException "errorClass" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION", "errorSubClass" : "DATETIME_PATTERN_RECOGNITION", "messageParameters" : { - "pattern" : "'dd MM yyyy EEEEE'", - "config" : "\"spark.sql.legacy.timeParserPolicy\"" + "config" : "\"spark.sql.legacy.timeParserPolicy\"", + "pattern" : "'dd MM yyyy EEEEE'" } } @@ -836,8 +836,8 @@ org.apache.spark.SparkUpgradeException "errorClass" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION", "errorSubClass" : "DATETIME_PATTERN_RECOGNITION", "messageParameters" : { - "pattern" : "'dd MM yyyy EEEEE'", - "config" : "\"spark.sql.legacy.timeParserPolicy\"" + "config" : "\"spark.sql.legacy.timeParserPolicy\"", + "pattern" : "'dd MM yyyy EEEEE'" } } @@ -852,8 +852,8 @@ org.apache.spark.SparkUpgradeException "errorClass" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION", "errorSubClass" : "DATETIME_PATTERN_RECOGNITION", "messageParameters" : { - "pattern" : "'dd/MMMMM/yyyy'", - "config" : "\"spark.sql.legacy.timeParserPolicy\"" + "config" : "\"spark.sql.legacy.timeParserPolicy\"", + "pattern" : "'dd/MMMMM/yyyy'" } } @@ -868,8 +868,8 @@ org.apache.spark.SparkUpgradeException "errorClass" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION", "errorSubClass" : "DATETIME_PATTERN_RECOGNITION", "messageParameters" : { - "pattern" : "'dd/MMMMM/yyyy'", - "config" : "\"spark.sql.legacy.timeParserPolicy\"" + "config" : "\"spark.sql.legacy.timeParserPolicy\"", + "pattern" : "'dd/MMMMM/yyyy'" } } diff --git a/sql/core/src/test/resources/sql-tests/results/timestampNTZ/timestamp-ansi.sql.out b/sql/core/src/test/resources/sql-tests/results/timestampNTZ/timestamp-ansi.sql.out index b494b62bdba6b..269b6d5ff0c23 100644 --- a/sql/core/src/test/resources/sql-tests/results/timestampNTZ/timestamp-ansi.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/timestampNTZ/timestamp-ansi.sql.out @@ -339,10 +339,10 @@ org.apache.spark.SparkDateTimeException "errorClass" : "CAST_INVALID_INPUT", "sqlState" : "42000", "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", "expression" : "'1'", "sourceType" : "\"STRING\"", - "targetType" : "\"TIMESTAMP_NTZ\"", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "targetType" : "\"TIMESTAMP_NTZ\"" }, "queryContext" : [ { "objectType" : "", @@ -362,8 +362,8 @@ org.apache.spark.SparkDateTimeException "errorClass" : "CANNOT_PARSE_TIMESTAMP", "sqlState" : "42000", "messageParameters" : { - "message" : "Text '2019-10-06 10:11:12.' could not be parsed at index 20", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "ansiConfig" : "\"spark.sql.ansi.enabled\"", + "message" : "Text '2019-10-06 10:11:12.' could not be parsed at index 20" } } @@ -434,8 +434,8 @@ org.apache.spark.SparkDateTimeException "errorClass" : "CANNOT_PARSE_TIMESTAMP", "sqlState" : "42000", "messageParameters" : { - "message" : "Text '2019-10-06 10:11:12.1234567PST' could not be parsed, unparsed text found at index 26", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "ansiConfig" : "\"spark.sql.ansi.enabled\"", + "message" : "Text '2019-10-06 10:11:12.1234567PST' could not be parsed, unparsed text found at index 26" } } @@ -458,8 +458,8 @@ org.apache.spark.SparkDateTimeException "errorClass" : "CANNOT_PARSE_TIMESTAMP", "sqlState" : "42000", "messageParameters" : { - "message" : "Text '223456 2019-10-06 10:11:12.123456PST' could not be parsed at index 27", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "ansiConfig" : "\"spark.sql.ansi.enabled\"", + "message" : "Text '223456 2019-10-06 10:11:12.123456PST' could not be parsed at index 27" } } @@ -530,8 +530,8 @@ org.apache.spark.SparkDateTimeException "errorClass" : "CANNOT_PARSE_TIMESTAMP", "sqlState" : "42000", "messageParameters" : { - "message" : "Text '12.1232019-10-06S10:11' could not be parsed at index 7", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "ansiConfig" : "\"spark.sql.ansi.enabled\"", + "message" : "Text '12.1232019-10-06S10:11' could not be parsed at index 7" } } @@ -546,8 +546,8 @@ org.apache.spark.SparkDateTimeException "errorClass" : "CANNOT_PARSE_TIMESTAMP", "sqlState" : "42000", "messageParameters" : { - "message" : "Text '12.1232019-10-06S10:11' could not be parsed at index 9", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "ansiConfig" : "\"spark.sql.ansi.enabled\"", + "message" : "Text '12.1232019-10-06S10:11' could not be parsed at index 9" } } @@ -626,8 +626,8 @@ org.apache.spark.SparkDateTimeException "errorClass" : "CANNOT_PARSE_TIMESTAMP", "sqlState" : "42000", "messageParameters" : { - "message" : "Invalid date 'February 29' as '1970' is not a leap year", - "ansiConfig" : "\"spark.sql.ansi.enabled\"" + "ansiConfig" : "\"spark.sql.ansi.enabled\"", + "message" : "Invalid date 'February 29' as '1970' is not a leap year" } } @@ -738,9 +738,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_WRONG_TYPE", "messageParameters" : { - "sqlExpr" : "\"(TIMESTAMP_NTZ '2011-11-11 11:11:11' + 1)\"", + "actualDataType" : "\"TIMESTAMP_NTZ\"", "inputType" : "(\"NUMERIC\" or \"INTERVAL DAY TO SECOND\" or \"INTERVAL YEAR TO MONTH\" or \"INTERVAL\")", - "actualDataType" : "\"TIMESTAMP_NTZ\"" + "sqlExpr" : "\"(TIMESTAMP_NTZ '2011-11-11 11:11:11' + 1)\"" }, "queryContext" : [ { "objectType" : "", @@ -762,9 +762,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_WRONG_TYPE", "messageParameters" : { - "sqlExpr" : "\"(1 + TIMESTAMP_NTZ '2011-11-11 11:11:11')\"", + "actualDataType" : "\"TIMESTAMP_NTZ\"", "inputType" : "(\"NUMERIC\" or \"INTERVAL DAY TO SECOND\" or \"INTERVAL YEAR TO MONTH\" or \"INTERVAL\")", - "actualDataType" : "\"TIMESTAMP_NTZ\"" + "sqlExpr" : "\"(1 + TIMESTAMP_NTZ '2011-11-11 11:11:11')\"" }, "queryContext" : [ { "objectType" : "", @@ -786,9 +786,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(TIMESTAMP_NTZ '2011-11-11 11:11:11' + NULL)\"", "left" : "\"TIMESTAMP_NTZ\"", - "right" : "\"VOID\"" + "right" : "\"VOID\"", + "sqlExpr" : "\"(TIMESTAMP_NTZ '2011-11-11 11:11:11' + NULL)\"" }, "queryContext" : [ { "objectType" : "", @@ -810,9 +810,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(NULL + TIMESTAMP_NTZ '2011-11-11 11:11:11')\"", "left" : "\"VOID\"", - "right" : "\"TIMESTAMP_NTZ\"" + "right" : "\"TIMESTAMP_NTZ\"", + "sqlExpr" : "\"(NULL + TIMESTAMP_NTZ '2011-11-11 11:11:11')\"" }, "queryContext" : [ { "objectType" : "", @@ -884,8 +884,8 @@ org.apache.spark.SparkUpgradeException "errorClass" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION", "errorSubClass" : "DATETIME_PATTERN_RECOGNITION", "messageParameters" : { - "pattern" : "'dd MM yyyy EEEEE'", - "config" : "\"spark.sql.legacy.timeParserPolicy\"" + "config" : "\"spark.sql.legacy.timeParserPolicy\"", + "pattern" : "'dd MM yyyy EEEEE'" } } diff --git a/sql/core/src/test/resources/sql-tests/results/timestampNTZ/timestamp.sql.out b/sql/core/src/test/resources/sql-tests/results/timestampNTZ/timestamp.sql.out index e99fa0d882e1f..b1ff61b4c9dd8 100644 --- a/sql/core/src/test/resources/sql-tests/results/timestampNTZ/timestamp.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/timestampNTZ/timestamp.sql.out @@ -669,9 +669,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(TIMESTAMP_NTZ '2011-11-11 11:11:11' + 1)\"", "left" : "\"TIMESTAMP_NTZ\"", - "right" : "\"DOUBLE\"" + "right" : "\"DOUBLE\"", + "sqlExpr" : "\"(TIMESTAMP_NTZ '2011-11-11 11:11:11' + 1)\"" }, "queryContext" : [ { "objectType" : "", @@ -693,9 +693,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(1 + TIMESTAMP_NTZ '2011-11-11 11:11:11')\"", "left" : "\"DOUBLE\"", - "right" : "\"TIMESTAMP_NTZ\"" + "right" : "\"TIMESTAMP_NTZ\"", + "sqlExpr" : "\"(1 + TIMESTAMP_NTZ '2011-11-11 11:11:11')\"" }, "queryContext" : [ { "objectType" : "", @@ -717,9 +717,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(TIMESTAMP_NTZ '2011-11-11 11:11:11' + NULL)\"", "left" : "\"TIMESTAMP_NTZ\"", - "right" : "\"VOID\"" + "right" : "\"VOID\"", + "sqlExpr" : "\"(TIMESTAMP_NTZ '2011-11-11 11:11:11' + NULL)\"" }, "queryContext" : [ { "objectType" : "", @@ -741,9 +741,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(NULL + TIMESTAMP_NTZ '2011-11-11 11:11:11')\"", "left" : "\"VOID\"", - "right" : "\"TIMESTAMP_NTZ\"" + "right" : "\"TIMESTAMP_NTZ\"", + "sqlExpr" : "\"(NULL + TIMESTAMP_NTZ '2011-11-11 11:11:11')\"" }, "queryContext" : [ { "objectType" : "", @@ -815,8 +815,8 @@ org.apache.spark.SparkUpgradeException "errorClass" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION", "errorSubClass" : "DATETIME_PATTERN_RECOGNITION", "messageParameters" : { - "pattern" : "'dd MM yyyy EEEEE'", - "config" : "\"spark.sql.legacy.timeParserPolicy\"" + "config" : "\"spark.sql.legacy.timeParserPolicy\"", + "pattern" : "'dd MM yyyy EEEEE'" } } diff --git a/sql/core/src/test/resources/sql-tests/results/try_datetime_functions.sql.out b/sql/core/src/test/resources/sql-tests/results/try_datetime_functions.sql.out index ab2172f56de86..c49637869e1fc 100644 --- a/sql/core/src/test/resources/sql-tests/results/try_datetime_functions.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/try_datetime_functions.sql.out @@ -49,7 +49,7 @@ org.apache.spark.SparkUpgradeException "errorClass" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION", "errorSubClass" : "DATETIME_PATTERN_RECOGNITION", "messageParameters" : { - "pattern" : "'dd MM yyyy EEEEEE'", - "config" : "\"spark.sql.legacy.timeParserPolicy\"" + "config" : "\"spark.sql.legacy.timeParserPolicy\"", + "pattern" : "'dd MM yyyy EEEEEE'" } } diff --git a/sql/core/src/test/resources/sql-tests/results/typeCoercion/native/booleanEquality.sql.out b/sql/core/src/test/resources/sql-tests/results/typeCoercion/native/booleanEquality.sql.out index d93351dae77fe..2912754249912 100644 --- a/sql/core/src/test/resources/sql-tests/results/typeCoercion/native/booleanEquality.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/typeCoercion/native/booleanEquality.sql.out @@ -81,9 +81,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(true = CAST(1 AS BINARY))\"", "left" : "\"BOOLEAN\"", - "right" : "\"BINARY\"" + "right" : "\"BINARY\"", + "sqlExpr" : "\"(true = CAST(1 AS BINARY))\"" }, "queryContext" : [ { "objectType" : "", @@ -113,9 +113,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(true = CAST(2017-12-11 09:30:00.0 AS TIMESTAMP))\"", "left" : "\"BOOLEAN\"", - "right" : "\"TIMESTAMP\"" + "right" : "\"TIMESTAMP\"", + "sqlExpr" : "\"(true = CAST(2017-12-11 09:30:00.0 AS TIMESTAMP))\"" }, "queryContext" : [ { "objectType" : "", @@ -137,9 +137,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(true = CAST(2017-12-11 09:30:00 AS DATE))\"", "left" : "\"BOOLEAN\"", - "right" : "\"DATE\"" + "right" : "\"DATE\"", + "sqlExpr" : "\"(true = CAST(2017-12-11 09:30:00 AS DATE))\"" }, "queryContext" : [ { "objectType" : "", @@ -225,9 +225,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(true <=> CAST(1 AS BINARY))\"", "left" : "\"BOOLEAN\"", - "right" : "\"BINARY\"" + "right" : "\"BINARY\"", + "sqlExpr" : "\"(true <=> CAST(1 AS BINARY))\"" }, "queryContext" : [ { "objectType" : "", @@ -257,9 +257,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(true <=> CAST(2017-12-11 09:30:00.0 AS TIMESTAMP))\"", "left" : "\"BOOLEAN\"", - "right" : "\"TIMESTAMP\"" + "right" : "\"TIMESTAMP\"", + "sqlExpr" : "\"(true <=> CAST(2017-12-11 09:30:00.0 AS TIMESTAMP))\"" }, "queryContext" : [ { "objectType" : "", @@ -281,9 +281,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(true <=> CAST(2017-12-11 09:30:00 AS DATE))\"", "left" : "\"BOOLEAN\"", - "right" : "\"DATE\"" + "right" : "\"DATE\"", + "sqlExpr" : "\"(true <=> CAST(2017-12-11 09:30:00 AS DATE))\"" }, "queryContext" : [ { "objectType" : "", @@ -369,9 +369,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS BINARY) = true)\"", "left" : "\"BINARY\"", - "right" : "\"BOOLEAN\"" + "right" : "\"BOOLEAN\"", + "sqlExpr" : "\"(CAST(1 AS BINARY) = true)\"" }, "queryContext" : [ { "objectType" : "", @@ -401,9 +401,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(2017-12-11 09:30:00.0 AS TIMESTAMP) = true)\"", "left" : "\"TIMESTAMP\"", - "right" : "\"BOOLEAN\"" + "right" : "\"BOOLEAN\"", + "sqlExpr" : "\"(CAST(2017-12-11 09:30:00.0 AS TIMESTAMP) = true)\"" }, "queryContext" : [ { "objectType" : "", @@ -425,9 +425,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(2017-12-11 09:30:00 AS DATE) = true)\"", "left" : "\"DATE\"", - "right" : "\"BOOLEAN\"" + "right" : "\"BOOLEAN\"", + "sqlExpr" : "\"(CAST(2017-12-11 09:30:00 AS DATE) = true)\"" }, "queryContext" : [ { "objectType" : "", @@ -513,9 +513,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS BINARY) <=> true)\"", "left" : "\"BINARY\"", - "right" : "\"BOOLEAN\"" + "right" : "\"BOOLEAN\"", + "sqlExpr" : "\"(CAST(1 AS BINARY) <=> true)\"" }, "queryContext" : [ { "objectType" : "", @@ -545,9 +545,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(2017-12-11 09:30:00.0 AS TIMESTAMP) <=> true)\"", "left" : "\"TIMESTAMP\"", - "right" : "\"BOOLEAN\"" + "right" : "\"BOOLEAN\"", + "sqlExpr" : "\"(CAST(2017-12-11 09:30:00.0 AS TIMESTAMP) <=> true)\"" }, "queryContext" : [ { "objectType" : "", @@ -569,9 +569,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(2017-12-11 09:30:00 AS DATE) <=> true)\"", "left" : "\"DATE\"", - "right" : "\"BOOLEAN\"" + "right" : "\"BOOLEAN\"", + "sqlExpr" : "\"(CAST(2017-12-11 09:30:00 AS DATE) <=> true)\"" }, "queryContext" : [ { "objectType" : "", @@ -657,9 +657,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(false = CAST(0 AS BINARY))\"", "left" : "\"BOOLEAN\"", - "right" : "\"BINARY\"" + "right" : "\"BINARY\"", + "sqlExpr" : "\"(false = CAST(0 AS BINARY))\"" }, "queryContext" : [ { "objectType" : "", @@ -689,9 +689,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(false = CAST(2017-12-11 09:30:00.0 AS TIMESTAMP))\"", "left" : "\"BOOLEAN\"", - "right" : "\"TIMESTAMP\"" + "right" : "\"TIMESTAMP\"", + "sqlExpr" : "\"(false = CAST(2017-12-11 09:30:00.0 AS TIMESTAMP))\"" }, "queryContext" : [ { "objectType" : "", @@ -713,9 +713,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(false = CAST(2017-12-11 09:30:00 AS DATE))\"", "left" : "\"BOOLEAN\"", - "right" : "\"DATE\"" + "right" : "\"DATE\"", + "sqlExpr" : "\"(false = CAST(2017-12-11 09:30:00 AS DATE))\"" }, "queryContext" : [ { "objectType" : "", @@ -801,9 +801,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(false <=> CAST(0 AS BINARY))\"", "left" : "\"BOOLEAN\"", - "right" : "\"BINARY\"" + "right" : "\"BINARY\"", + "sqlExpr" : "\"(false <=> CAST(0 AS BINARY))\"" }, "queryContext" : [ { "objectType" : "", @@ -833,9 +833,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(false <=> CAST(2017-12-11 09:30:00.0 AS TIMESTAMP))\"", "left" : "\"BOOLEAN\"", - "right" : "\"TIMESTAMP\"" + "right" : "\"TIMESTAMP\"", + "sqlExpr" : "\"(false <=> CAST(2017-12-11 09:30:00.0 AS TIMESTAMP))\"" }, "queryContext" : [ { "objectType" : "", @@ -857,9 +857,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(false <=> CAST(2017-12-11 09:30:00 AS DATE))\"", "left" : "\"BOOLEAN\"", - "right" : "\"DATE\"" + "right" : "\"DATE\"", + "sqlExpr" : "\"(false <=> CAST(2017-12-11 09:30:00 AS DATE))\"" }, "queryContext" : [ { "objectType" : "", @@ -945,9 +945,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(0 AS BINARY) = false)\"", "left" : "\"BINARY\"", - "right" : "\"BOOLEAN\"" + "right" : "\"BOOLEAN\"", + "sqlExpr" : "\"(CAST(0 AS BINARY) = false)\"" }, "queryContext" : [ { "objectType" : "", @@ -977,9 +977,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(2017-12-11 09:30:00.0 AS TIMESTAMP) = false)\"", "left" : "\"TIMESTAMP\"", - "right" : "\"BOOLEAN\"" + "right" : "\"BOOLEAN\"", + "sqlExpr" : "\"(CAST(2017-12-11 09:30:00.0 AS TIMESTAMP) = false)\"" }, "queryContext" : [ { "objectType" : "", @@ -1001,9 +1001,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(2017-12-11 09:30:00 AS DATE) = false)\"", "left" : "\"DATE\"", - "right" : "\"BOOLEAN\"" + "right" : "\"BOOLEAN\"", + "sqlExpr" : "\"(CAST(2017-12-11 09:30:00 AS DATE) = false)\"" }, "queryContext" : [ { "objectType" : "", @@ -1089,9 +1089,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(0 AS BINARY) <=> false)\"", "left" : "\"BINARY\"", - "right" : "\"BOOLEAN\"" + "right" : "\"BOOLEAN\"", + "sqlExpr" : "\"(CAST(0 AS BINARY) <=> false)\"" }, "queryContext" : [ { "objectType" : "", @@ -1121,9 +1121,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(2017-12-11 09:30:00.0 AS TIMESTAMP) <=> false)\"", "left" : "\"TIMESTAMP\"", - "right" : "\"BOOLEAN\"" + "right" : "\"BOOLEAN\"", + "sqlExpr" : "\"(CAST(2017-12-11 09:30:00.0 AS TIMESTAMP) <=> false)\"" }, "queryContext" : [ { "objectType" : "", @@ -1145,9 +1145,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(2017-12-11 09:30:00 AS DATE) <=> false)\"", "left" : "\"DATE\"", - "right" : "\"BOOLEAN\"" + "right" : "\"BOOLEAN\"", + "sqlExpr" : "\"(CAST(2017-12-11 09:30:00 AS DATE) <=> false)\"" }, "queryContext" : [ { "objectType" : "", diff --git a/sql/core/src/test/resources/sql-tests/results/typeCoercion/native/decimalPrecision.sql.out b/sql/core/src/test/resources/sql-tests/results/typeCoercion/native/decimalPrecision.sql.out index 387a85099528b..4b9973c53cae5 100644 --- a/sql/core/src/test/resources/sql-tests/results/typeCoercion/native/decimalPrecision.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/typeCoercion/native/decimalPrecision.sql.out @@ -241,9 +241,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS BINARY) + CAST(1 AS DECIMAL(3,0)))\"", "left" : "\"BINARY\"", - "right" : "\"DECIMAL(3,0)\"" + "right" : "\"DECIMAL(3,0)\"", + "sqlExpr" : "\"(CAST(1 AS BINARY) + CAST(1 AS DECIMAL(3,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -265,9 +265,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS BINARY) + CAST(1 AS DECIMAL(5,0)))\"", "left" : "\"BINARY\"", - "right" : "\"DECIMAL(5,0)\"" + "right" : "\"DECIMAL(5,0)\"", + "sqlExpr" : "\"(CAST(1 AS BINARY) + CAST(1 AS DECIMAL(5,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -289,9 +289,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS BINARY) + CAST(1 AS DECIMAL(10,0)))\"", "left" : "\"BINARY\"", - "right" : "\"DECIMAL(10,0)\"" + "right" : "\"DECIMAL(10,0)\"", + "sqlExpr" : "\"(CAST(1 AS BINARY) + CAST(1 AS DECIMAL(10,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -313,9 +313,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS BINARY) + CAST(1 AS DECIMAL(20,0)))\"", "left" : "\"BINARY\"", - "right" : "\"DECIMAL(20,0)\"" + "right" : "\"DECIMAL(20,0)\"", + "sqlExpr" : "\"(CAST(1 AS BINARY) + CAST(1 AS DECIMAL(20,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -337,9 +337,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(2017-12-11 09:30:00.0 AS TIMESTAMP) + CAST(1 AS DECIMAL(3,0)))\"", "left" : "\"TIMESTAMP\"", - "right" : "\"DECIMAL(3,0)\"" + "right" : "\"DECIMAL(3,0)\"", + "sqlExpr" : "\"(CAST(2017-12-11 09:30:00.0 AS TIMESTAMP) + CAST(1 AS DECIMAL(3,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -361,9 +361,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(2017-12-11 09:30:00.0 AS TIMESTAMP) + CAST(1 AS DECIMAL(5,0)))\"", "left" : "\"TIMESTAMP\"", - "right" : "\"DECIMAL(5,0)\"" + "right" : "\"DECIMAL(5,0)\"", + "sqlExpr" : "\"(CAST(2017-12-11 09:30:00.0 AS TIMESTAMP) + CAST(1 AS DECIMAL(5,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -385,9 +385,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(2017-12-11 09:30:00.0 AS TIMESTAMP) + CAST(1 AS DECIMAL(10,0)))\"", "left" : "\"TIMESTAMP\"", - "right" : "\"DECIMAL(10,0)\"" + "right" : "\"DECIMAL(10,0)\"", + "sqlExpr" : "\"(CAST(2017-12-11 09:30:00.0 AS TIMESTAMP) + CAST(1 AS DECIMAL(10,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -409,9 +409,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(2017-12-11 09:30:00.0 AS TIMESTAMP) + CAST(1 AS DECIMAL(20,0)))\"", "left" : "\"TIMESTAMP\"", - "right" : "\"DECIMAL(20,0)\"" + "right" : "\"DECIMAL(20,0)\"", + "sqlExpr" : "\"(CAST(2017-12-11 09:30:00.0 AS TIMESTAMP) + CAST(1 AS DECIMAL(20,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -725,9 +725,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(3,0)) + CAST(1 AS BINARY))\"", "left" : "\"DECIMAL(3,0)\"", - "right" : "\"BINARY\"" + "right" : "\"BINARY\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(3,0)) + CAST(1 AS BINARY))\"" }, "queryContext" : [ { "objectType" : "", @@ -749,9 +749,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(5,0)) + CAST(1 AS BINARY))\"", "left" : "\"DECIMAL(5,0)\"", - "right" : "\"BINARY\"" + "right" : "\"BINARY\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(5,0)) + CAST(1 AS BINARY))\"" }, "queryContext" : [ { "objectType" : "", @@ -773,9 +773,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(10,0)) + CAST(1 AS BINARY))\"", "left" : "\"DECIMAL(10,0)\"", - "right" : "\"BINARY\"" + "right" : "\"BINARY\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(10,0)) + CAST(1 AS BINARY))\"" }, "queryContext" : [ { "objectType" : "", @@ -797,9 +797,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(20,0)) + CAST(1 AS BINARY))\"", "left" : "\"DECIMAL(20,0)\"", - "right" : "\"BINARY\"" + "right" : "\"BINARY\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(20,0)) + CAST(1 AS BINARY))\"" }, "queryContext" : [ { "objectType" : "", @@ -821,9 +821,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(3,0)) + CAST(1 AS BOOLEAN))\"", "left" : "\"DECIMAL(3,0)\"", - "right" : "\"BOOLEAN\"" + "right" : "\"BOOLEAN\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(3,0)) + CAST(1 AS BOOLEAN))\"" }, "queryContext" : [ { "objectType" : "", @@ -845,9 +845,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(5,0)) + CAST(1 AS BOOLEAN))\"", "left" : "\"DECIMAL(5,0)\"", - "right" : "\"BOOLEAN\"" + "right" : "\"BOOLEAN\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(5,0)) + CAST(1 AS BOOLEAN))\"" }, "queryContext" : [ { "objectType" : "", @@ -869,9 +869,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(10,0)) + CAST(1 AS BOOLEAN))\"", "left" : "\"DECIMAL(10,0)\"", - "right" : "\"BOOLEAN\"" + "right" : "\"BOOLEAN\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(10,0)) + CAST(1 AS BOOLEAN))\"" }, "queryContext" : [ { "objectType" : "", @@ -893,9 +893,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(20,0)) + CAST(1 AS BOOLEAN))\"", "left" : "\"DECIMAL(20,0)\"", - "right" : "\"BOOLEAN\"" + "right" : "\"BOOLEAN\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(20,0)) + CAST(1 AS BOOLEAN))\"" }, "queryContext" : [ { "objectType" : "", @@ -917,9 +917,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(3,0)) + CAST(2017-12-11 09:30:00.0 AS TIMESTAMP))\"", "left" : "\"DECIMAL(3,0)\"", - "right" : "\"TIMESTAMP\"" + "right" : "\"TIMESTAMP\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(3,0)) + CAST(2017-12-11 09:30:00.0 AS TIMESTAMP))\"" }, "queryContext" : [ { "objectType" : "", @@ -941,9 +941,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(5,0)) + CAST(2017-12-11 09:30:00.0 AS TIMESTAMP))\"", "left" : "\"DECIMAL(5,0)\"", - "right" : "\"TIMESTAMP\"" + "right" : "\"TIMESTAMP\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(5,0)) + CAST(2017-12-11 09:30:00.0 AS TIMESTAMP))\"" }, "queryContext" : [ { "objectType" : "", @@ -965,9 +965,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(10,0)) + CAST(2017-12-11 09:30:00.0 AS TIMESTAMP))\"", "left" : "\"DECIMAL(10,0)\"", - "right" : "\"TIMESTAMP\"" + "right" : "\"TIMESTAMP\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(10,0)) + CAST(2017-12-11 09:30:00.0 AS TIMESTAMP))\"" }, "queryContext" : [ { "objectType" : "", @@ -989,9 +989,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(20,0)) + CAST(2017-12-11 09:30:00.0 AS TIMESTAMP))\"", "left" : "\"DECIMAL(20,0)\"", - "right" : "\"TIMESTAMP\"" + "right" : "\"TIMESTAMP\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(20,0)) + CAST(2017-12-11 09:30:00.0 AS TIMESTAMP))\"" }, "queryContext" : [ { "objectType" : "", @@ -1273,9 +1273,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS BINARY) - CAST(1 AS DECIMAL(3,0)))\"", "left" : "\"BINARY\"", - "right" : "\"DECIMAL(3,0)\"" + "right" : "\"DECIMAL(3,0)\"", + "sqlExpr" : "\"(CAST(1 AS BINARY) - CAST(1 AS DECIMAL(3,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -1297,9 +1297,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS BINARY) - CAST(1 AS DECIMAL(5,0)))\"", "left" : "\"BINARY\"", - "right" : "\"DECIMAL(5,0)\"" + "right" : "\"DECIMAL(5,0)\"", + "sqlExpr" : "\"(CAST(1 AS BINARY) - CAST(1 AS DECIMAL(5,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -1321,9 +1321,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS BINARY) - CAST(1 AS DECIMAL(10,0)))\"", "left" : "\"BINARY\"", - "right" : "\"DECIMAL(10,0)\"" + "right" : "\"DECIMAL(10,0)\"", + "sqlExpr" : "\"(CAST(1 AS BINARY) - CAST(1 AS DECIMAL(10,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -1345,9 +1345,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS BINARY) - CAST(1 AS DECIMAL(20,0)))\"", "left" : "\"BINARY\"", - "right" : "\"DECIMAL(20,0)\"" + "right" : "\"DECIMAL(20,0)\"", + "sqlExpr" : "\"(CAST(1 AS BINARY) - CAST(1 AS DECIMAL(20,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -1697,9 +1697,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(3,0)) - CAST(1 AS BINARY))\"", "left" : "\"DECIMAL(3,0)\"", - "right" : "\"BINARY\"" + "right" : "\"BINARY\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(3,0)) - CAST(1 AS BINARY))\"" }, "queryContext" : [ { "objectType" : "", @@ -1721,9 +1721,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(5,0)) - CAST(1 AS BINARY))\"", "left" : "\"DECIMAL(5,0)\"", - "right" : "\"BINARY\"" + "right" : "\"BINARY\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(5,0)) - CAST(1 AS BINARY))\"" }, "queryContext" : [ { "objectType" : "", @@ -1745,9 +1745,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(10,0)) - CAST(1 AS BINARY))\"", "left" : "\"DECIMAL(10,0)\"", - "right" : "\"BINARY\"" + "right" : "\"BINARY\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(10,0)) - CAST(1 AS BINARY))\"" }, "queryContext" : [ { "objectType" : "", @@ -1769,9 +1769,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(20,0)) - CAST(1 AS BINARY))\"", "left" : "\"DECIMAL(20,0)\"", - "right" : "\"BINARY\"" + "right" : "\"BINARY\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(20,0)) - CAST(1 AS BINARY))\"" }, "queryContext" : [ { "objectType" : "", @@ -1793,9 +1793,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(3,0)) - CAST(1 AS BOOLEAN))\"", "left" : "\"DECIMAL(3,0)\"", - "right" : "\"BOOLEAN\"" + "right" : "\"BOOLEAN\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(3,0)) - CAST(1 AS BOOLEAN))\"" }, "queryContext" : [ { "objectType" : "", @@ -1817,9 +1817,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(5,0)) - CAST(1 AS BOOLEAN))\"", "left" : "\"DECIMAL(5,0)\"", - "right" : "\"BOOLEAN\"" + "right" : "\"BOOLEAN\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(5,0)) - CAST(1 AS BOOLEAN))\"" }, "queryContext" : [ { "objectType" : "", @@ -1841,9 +1841,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(10,0)) - CAST(1 AS BOOLEAN))\"", "left" : "\"DECIMAL(10,0)\"", - "right" : "\"BOOLEAN\"" + "right" : "\"BOOLEAN\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(10,0)) - CAST(1 AS BOOLEAN))\"" }, "queryContext" : [ { "objectType" : "", @@ -1865,9 +1865,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(20,0)) - CAST(1 AS BOOLEAN))\"", "left" : "\"DECIMAL(20,0)\"", - "right" : "\"BOOLEAN\"" + "right" : "\"BOOLEAN\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(20,0)) - CAST(1 AS BOOLEAN))\"" }, "queryContext" : [ { "objectType" : "", @@ -2185,9 +2185,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS BINARY) * CAST(1 AS DECIMAL(3,0)))\"", "left" : "\"BINARY\"", - "right" : "\"DECIMAL(3,0)\"" + "right" : "\"DECIMAL(3,0)\"", + "sqlExpr" : "\"(CAST(1 AS BINARY) * CAST(1 AS DECIMAL(3,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -2209,9 +2209,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS BINARY) * CAST(1 AS DECIMAL(5,0)))\"", "left" : "\"BINARY\"", - "right" : "\"DECIMAL(5,0)\"" + "right" : "\"DECIMAL(5,0)\"", + "sqlExpr" : "\"(CAST(1 AS BINARY) * CAST(1 AS DECIMAL(5,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -2233,9 +2233,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS BINARY) * CAST(1 AS DECIMAL(10,0)))\"", "left" : "\"BINARY\"", - "right" : "\"DECIMAL(10,0)\"" + "right" : "\"DECIMAL(10,0)\"", + "sqlExpr" : "\"(CAST(1 AS BINARY) * CAST(1 AS DECIMAL(10,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -2257,9 +2257,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS BINARY) * CAST(1 AS DECIMAL(20,0)))\"", "left" : "\"BINARY\"", - "right" : "\"DECIMAL(20,0)\"" + "right" : "\"DECIMAL(20,0)\"", + "sqlExpr" : "\"(CAST(1 AS BINARY) * CAST(1 AS DECIMAL(20,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -2281,9 +2281,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(2017*12*11 09:30:00.0 AS TIMESTAMP) * CAST(1 AS DECIMAL(3,0)))\"", "left" : "\"TIMESTAMP\"", - "right" : "\"DECIMAL(3,0)\"" + "right" : "\"DECIMAL(3,0)\"", + "sqlExpr" : "\"(CAST(2017*12*11 09:30:00.0 AS TIMESTAMP) * CAST(1 AS DECIMAL(3,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -2305,9 +2305,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(2017*12*11 09:30:00.0 AS TIMESTAMP) * CAST(1 AS DECIMAL(5,0)))\"", "left" : "\"TIMESTAMP\"", - "right" : "\"DECIMAL(5,0)\"" + "right" : "\"DECIMAL(5,0)\"", + "sqlExpr" : "\"(CAST(2017*12*11 09:30:00.0 AS TIMESTAMP) * CAST(1 AS DECIMAL(5,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -2329,9 +2329,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(2017*12*11 09:30:00.0 AS TIMESTAMP) * CAST(1 AS DECIMAL(10,0)))\"", "left" : "\"TIMESTAMP\"", - "right" : "\"DECIMAL(10,0)\"" + "right" : "\"DECIMAL(10,0)\"", + "sqlExpr" : "\"(CAST(2017*12*11 09:30:00.0 AS TIMESTAMP) * CAST(1 AS DECIMAL(10,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -2353,9 +2353,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(2017*12*11 09:30:00.0 AS TIMESTAMP) * CAST(1 AS DECIMAL(20,0)))\"", "left" : "\"TIMESTAMP\"", - "right" : "\"DECIMAL(20,0)\"" + "right" : "\"DECIMAL(20,0)\"", + "sqlExpr" : "\"(CAST(2017*12*11 09:30:00.0 AS TIMESTAMP) * CAST(1 AS DECIMAL(20,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -2377,9 +2377,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(2017*12*11 09:30:00 AS DATE) * CAST(1 AS DECIMAL(3,0)))\"", "left" : "\"DATE\"", - "right" : "\"DECIMAL(3,0)\"" + "right" : "\"DECIMAL(3,0)\"", + "sqlExpr" : "\"(CAST(2017*12*11 09:30:00 AS DATE) * CAST(1 AS DECIMAL(3,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -2401,9 +2401,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(2017*12*11 09:30:00 AS DATE) * CAST(1 AS DECIMAL(5,0)))\"", "left" : "\"DATE\"", - "right" : "\"DECIMAL(5,0)\"" + "right" : "\"DECIMAL(5,0)\"", + "sqlExpr" : "\"(CAST(2017*12*11 09:30:00 AS DATE) * CAST(1 AS DECIMAL(5,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -2425,9 +2425,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(2017*12*11 09:30:00 AS DATE) * CAST(1 AS DECIMAL(10,0)))\"", "left" : "\"DATE\"", - "right" : "\"DECIMAL(10,0)\"" + "right" : "\"DECIMAL(10,0)\"", + "sqlExpr" : "\"(CAST(2017*12*11 09:30:00 AS DATE) * CAST(1 AS DECIMAL(10,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -2449,9 +2449,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(2017*12*11 09:30:00 AS DATE) * CAST(1 AS DECIMAL(20,0)))\"", "left" : "\"DATE\"", - "right" : "\"DECIMAL(20,0)\"" + "right" : "\"DECIMAL(20,0)\"", + "sqlExpr" : "\"(CAST(2017*12*11 09:30:00 AS DATE) * CAST(1 AS DECIMAL(20,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -2729,9 +2729,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(3,0)) * CAST(1 AS BINARY))\"", "left" : "\"DECIMAL(3,0)\"", - "right" : "\"BINARY\"" + "right" : "\"BINARY\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(3,0)) * CAST(1 AS BINARY))\"" }, "queryContext" : [ { "objectType" : "", @@ -2753,9 +2753,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(5,0)) * CAST(1 AS BINARY))\"", "left" : "\"DECIMAL(5,0)\"", - "right" : "\"BINARY\"" + "right" : "\"BINARY\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(5,0)) * CAST(1 AS BINARY))\"" }, "queryContext" : [ { "objectType" : "", @@ -2777,9 +2777,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(10,0)) * CAST(1 AS BINARY))\"", "left" : "\"DECIMAL(10,0)\"", - "right" : "\"BINARY\"" + "right" : "\"BINARY\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(10,0)) * CAST(1 AS BINARY))\"" }, "queryContext" : [ { "objectType" : "", @@ -2801,9 +2801,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(20,0)) * CAST(1 AS BINARY))\"", "left" : "\"DECIMAL(20,0)\"", - "right" : "\"BINARY\"" + "right" : "\"BINARY\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(20,0)) * CAST(1 AS BINARY))\"" }, "queryContext" : [ { "objectType" : "", @@ -2825,9 +2825,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(3,0)) * CAST(1 AS BOOLEAN))\"", "left" : "\"DECIMAL(3,0)\"", - "right" : "\"BOOLEAN\"" + "right" : "\"BOOLEAN\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(3,0)) * CAST(1 AS BOOLEAN))\"" }, "queryContext" : [ { "objectType" : "", @@ -2849,9 +2849,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(5,0)) * CAST(1 AS BOOLEAN))\"", "left" : "\"DECIMAL(5,0)\"", - "right" : "\"BOOLEAN\"" + "right" : "\"BOOLEAN\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(5,0)) * CAST(1 AS BOOLEAN))\"" }, "queryContext" : [ { "objectType" : "", @@ -2873,9 +2873,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(10,0)) * CAST(1 AS BOOLEAN))\"", "left" : "\"DECIMAL(10,0)\"", - "right" : "\"BOOLEAN\"" + "right" : "\"BOOLEAN\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(10,0)) * CAST(1 AS BOOLEAN))\"" }, "queryContext" : [ { "objectType" : "", @@ -2897,9 +2897,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(20,0)) * CAST(1 AS BOOLEAN))\"", "left" : "\"DECIMAL(20,0)\"", - "right" : "\"BOOLEAN\"" + "right" : "\"BOOLEAN\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(20,0)) * CAST(1 AS BOOLEAN))\"" }, "queryContext" : [ { "objectType" : "", @@ -2921,9 +2921,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(3,0)) * CAST(2017*12*11 09:30:00.0 AS TIMESTAMP))\"", "left" : "\"DECIMAL(3,0)\"", - "right" : "\"TIMESTAMP\"" + "right" : "\"TIMESTAMP\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(3,0)) * CAST(2017*12*11 09:30:00.0 AS TIMESTAMP))\"" }, "queryContext" : [ { "objectType" : "", @@ -2945,9 +2945,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(5,0)) * CAST(2017*12*11 09:30:00.0 AS TIMESTAMP))\"", "left" : "\"DECIMAL(5,0)\"", - "right" : "\"TIMESTAMP\"" + "right" : "\"TIMESTAMP\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(5,0)) * CAST(2017*12*11 09:30:00.0 AS TIMESTAMP))\"" }, "queryContext" : [ { "objectType" : "", @@ -2969,9 +2969,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(10,0)) * CAST(2017*12*11 09:30:00.0 AS TIMESTAMP))\"", "left" : "\"DECIMAL(10,0)\"", - "right" : "\"TIMESTAMP\"" + "right" : "\"TIMESTAMP\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(10,0)) * CAST(2017*12*11 09:30:00.0 AS TIMESTAMP))\"" }, "queryContext" : [ { "objectType" : "", @@ -2993,9 +2993,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(20,0)) * CAST(2017*12*11 09:30:00.0 AS TIMESTAMP))\"", "left" : "\"DECIMAL(20,0)\"", - "right" : "\"TIMESTAMP\"" + "right" : "\"TIMESTAMP\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(20,0)) * CAST(2017*12*11 09:30:00.0 AS TIMESTAMP))\"" }, "queryContext" : [ { "objectType" : "", @@ -3017,9 +3017,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(3,0)) * CAST(2017*12*11 09:30:00 AS DATE))\"", "left" : "\"DECIMAL(3,0)\"", - "right" : "\"DATE\"" + "right" : "\"DATE\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(3,0)) * CAST(2017*12*11 09:30:00 AS DATE))\"" }, "queryContext" : [ { "objectType" : "", @@ -3041,9 +3041,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(5,0)) * CAST(2017*12*11 09:30:00 AS DATE))\"", "left" : "\"DECIMAL(5,0)\"", - "right" : "\"DATE\"" + "right" : "\"DATE\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(5,0)) * CAST(2017*12*11 09:30:00 AS DATE))\"" }, "queryContext" : [ { "objectType" : "", @@ -3065,9 +3065,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(10,0)) * CAST(2017*12*11 09:30:00 AS DATE))\"", "left" : "\"DECIMAL(10,0)\"", - "right" : "\"DATE\"" + "right" : "\"DATE\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(10,0)) * CAST(2017*12*11 09:30:00 AS DATE))\"" }, "queryContext" : [ { "objectType" : "", @@ -3089,9 +3089,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(20,0)) * CAST(2017*12*11 09:30:00 AS DATE))\"", "left" : "\"DECIMAL(20,0)\"", - "right" : "\"DATE\"" + "right" : "\"DATE\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(20,0)) * CAST(2017*12*11 09:30:00 AS DATE))\"" }, "queryContext" : [ { "objectType" : "", @@ -3337,9 +3337,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS BINARY) / CAST(1 AS DECIMAL(3,0)))\"", "left" : "\"BINARY\"", - "right" : "\"DECIMAL(3,0)\"" + "right" : "\"DECIMAL(3,0)\"", + "sqlExpr" : "\"(CAST(1 AS BINARY) / CAST(1 AS DECIMAL(3,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -3361,9 +3361,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS BINARY) / CAST(1 AS DECIMAL(5,0)))\"", "left" : "\"BINARY\"", - "right" : "\"DECIMAL(5,0)\"" + "right" : "\"DECIMAL(5,0)\"", + "sqlExpr" : "\"(CAST(1 AS BINARY) / CAST(1 AS DECIMAL(5,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -3385,9 +3385,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS BINARY) / CAST(1 AS DECIMAL(10,0)))\"", "left" : "\"BINARY\"", - "right" : "\"DECIMAL(10,0)\"" + "right" : "\"DECIMAL(10,0)\"", + "sqlExpr" : "\"(CAST(1 AS BINARY) / CAST(1 AS DECIMAL(10,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -3409,9 +3409,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS BINARY) / CAST(1 AS DECIMAL(20,0)))\"", "left" : "\"BINARY\"", - "right" : "\"DECIMAL(20,0)\"" + "right" : "\"DECIMAL(20,0)\"", + "sqlExpr" : "\"(CAST(1 AS BINARY) / CAST(1 AS DECIMAL(20,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -3433,9 +3433,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(2017/12/11 09:30:00.0 AS TIMESTAMP) / CAST(1 AS DECIMAL(3,0)))\"", "left" : "\"TIMESTAMP\"", - "right" : "\"DECIMAL(3,0)\"" + "right" : "\"DECIMAL(3,0)\"", + "sqlExpr" : "\"(CAST(2017/12/11 09:30:00.0 AS TIMESTAMP) / CAST(1 AS DECIMAL(3,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -3457,9 +3457,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(2017/12/11 09:30:00.0 AS TIMESTAMP) / CAST(1 AS DECIMAL(5,0)))\"", "left" : "\"TIMESTAMP\"", - "right" : "\"DECIMAL(5,0)\"" + "right" : "\"DECIMAL(5,0)\"", + "sqlExpr" : "\"(CAST(2017/12/11 09:30:00.0 AS TIMESTAMP) / CAST(1 AS DECIMAL(5,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -3481,9 +3481,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(2017/12/11 09:30:00.0 AS TIMESTAMP) / CAST(1 AS DECIMAL(10,0)))\"", "left" : "\"TIMESTAMP\"", - "right" : "\"DECIMAL(10,0)\"" + "right" : "\"DECIMAL(10,0)\"", + "sqlExpr" : "\"(CAST(2017/12/11 09:30:00.0 AS TIMESTAMP) / CAST(1 AS DECIMAL(10,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -3505,9 +3505,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(2017/12/11 09:30:00.0 AS TIMESTAMP) / CAST(1 AS DECIMAL(20,0)))\"", "left" : "\"TIMESTAMP\"", - "right" : "\"DECIMAL(20,0)\"" + "right" : "\"DECIMAL(20,0)\"", + "sqlExpr" : "\"(CAST(2017/12/11 09:30:00.0 AS TIMESTAMP) / CAST(1 AS DECIMAL(20,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -3529,9 +3529,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(2017/12/11 09:30:00 AS DATE) / CAST(1 AS DECIMAL(3,0)))\"", "left" : "\"DATE\"", - "right" : "\"DECIMAL(3,0)\"" + "right" : "\"DECIMAL(3,0)\"", + "sqlExpr" : "\"(CAST(2017/12/11 09:30:00 AS DATE) / CAST(1 AS DECIMAL(3,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -3553,9 +3553,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(2017/12/11 09:30:00 AS DATE) / CAST(1 AS DECIMAL(5,0)))\"", "left" : "\"DATE\"", - "right" : "\"DECIMAL(5,0)\"" + "right" : "\"DECIMAL(5,0)\"", + "sqlExpr" : "\"(CAST(2017/12/11 09:30:00 AS DATE) / CAST(1 AS DECIMAL(5,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -3577,9 +3577,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(2017/12/11 09:30:00 AS DATE) / CAST(1 AS DECIMAL(10,0)))\"", "left" : "\"DATE\"", - "right" : "\"DECIMAL(10,0)\"" + "right" : "\"DECIMAL(10,0)\"", + "sqlExpr" : "\"(CAST(2017/12/11 09:30:00 AS DATE) / CAST(1 AS DECIMAL(10,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -3601,9 +3601,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(2017/12/11 09:30:00 AS DATE) / CAST(1 AS DECIMAL(20,0)))\"", "left" : "\"DATE\"", - "right" : "\"DECIMAL(20,0)\"" + "right" : "\"DECIMAL(20,0)\"", + "sqlExpr" : "\"(CAST(2017/12/11 09:30:00 AS DATE) / CAST(1 AS DECIMAL(20,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -3881,9 +3881,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(3,0)) / CAST(1 AS BINARY))\"", "left" : "\"DECIMAL(3,0)\"", - "right" : "\"BINARY\"" + "right" : "\"BINARY\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(3,0)) / CAST(1 AS BINARY))\"" }, "queryContext" : [ { "objectType" : "", @@ -3905,9 +3905,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(5,0)) / CAST(1 AS BINARY))\"", "left" : "\"DECIMAL(5,0)\"", - "right" : "\"BINARY\"" + "right" : "\"BINARY\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(5,0)) / CAST(1 AS BINARY))\"" }, "queryContext" : [ { "objectType" : "", @@ -3929,9 +3929,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(10,0)) / CAST(1 AS BINARY))\"", "left" : "\"DECIMAL(10,0)\"", - "right" : "\"BINARY\"" + "right" : "\"BINARY\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(10,0)) / CAST(1 AS BINARY))\"" }, "queryContext" : [ { "objectType" : "", @@ -3953,9 +3953,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(20,0)) / CAST(1 AS BINARY))\"", "left" : "\"DECIMAL(20,0)\"", - "right" : "\"BINARY\"" + "right" : "\"BINARY\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(20,0)) / CAST(1 AS BINARY))\"" }, "queryContext" : [ { "objectType" : "", @@ -3977,9 +3977,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(3,0)) / CAST(1 AS BOOLEAN))\"", "left" : "\"DECIMAL(3,0)\"", - "right" : "\"BOOLEAN\"" + "right" : "\"BOOLEAN\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(3,0)) / CAST(1 AS BOOLEAN))\"" }, "queryContext" : [ { "objectType" : "", @@ -4001,9 +4001,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(5,0)) / CAST(1 AS BOOLEAN))\"", "left" : "\"DECIMAL(5,0)\"", - "right" : "\"BOOLEAN\"" + "right" : "\"BOOLEAN\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(5,0)) / CAST(1 AS BOOLEAN))\"" }, "queryContext" : [ { "objectType" : "", @@ -4025,9 +4025,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(10,0)) / CAST(1 AS BOOLEAN))\"", "left" : "\"DECIMAL(10,0)\"", - "right" : "\"BOOLEAN\"" + "right" : "\"BOOLEAN\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(10,0)) / CAST(1 AS BOOLEAN))\"" }, "queryContext" : [ { "objectType" : "", @@ -4049,9 +4049,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(20,0)) / CAST(1 AS BOOLEAN))\"", "left" : "\"DECIMAL(20,0)\"", - "right" : "\"BOOLEAN\"" + "right" : "\"BOOLEAN\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(20,0)) / CAST(1 AS BOOLEAN))\"" }, "queryContext" : [ { "objectType" : "", @@ -4073,9 +4073,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(3,0)) / CAST(2017/12/11 09:30:00.0 AS TIMESTAMP))\"", "left" : "\"DECIMAL(3,0)\"", - "right" : "\"TIMESTAMP\"" + "right" : "\"TIMESTAMP\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(3,0)) / CAST(2017/12/11 09:30:00.0 AS TIMESTAMP))\"" }, "queryContext" : [ { "objectType" : "", @@ -4097,9 +4097,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(5,0)) / CAST(2017/12/11 09:30:00.0 AS TIMESTAMP))\"", "left" : "\"DECIMAL(5,0)\"", - "right" : "\"TIMESTAMP\"" + "right" : "\"TIMESTAMP\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(5,0)) / CAST(2017/12/11 09:30:00.0 AS TIMESTAMP))\"" }, "queryContext" : [ { "objectType" : "", @@ -4121,9 +4121,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(10,0)) / CAST(2017/12/11 09:30:00.0 AS TIMESTAMP))\"", "left" : "\"DECIMAL(10,0)\"", - "right" : "\"TIMESTAMP\"" + "right" : "\"TIMESTAMP\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(10,0)) / CAST(2017/12/11 09:30:00.0 AS TIMESTAMP))\"" }, "queryContext" : [ { "objectType" : "", @@ -4145,9 +4145,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(20,0)) / CAST(2017/12/11 09:30:00.0 AS TIMESTAMP))\"", "left" : "\"DECIMAL(20,0)\"", - "right" : "\"TIMESTAMP\"" + "right" : "\"TIMESTAMP\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(20,0)) / CAST(2017/12/11 09:30:00.0 AS TIMESTAMP))\"" }, "queryContext" : [ { "objectType" : "", @@ -4169,9 +4169,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(3,0)) / CAST(2017/12/11 09:30:00 AS DATE))\"", "left" : "\"DECIMAL(3,0)\"", - "right" : "\"DATE\"" + "right" : "\"DATE\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(3,0)) / CAST(2017/12/11 09:30:00 AS DATE))\"" }, "queryContext" : [ { "objectType" : "", @@ -4193,9 +4193,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(5,0)) / CAST(2017/12/11 09:30:00 AS DATE))\"", "left" : "\"DECIMAL(5,0)\"", - "right" : "\"DATE\"" + "right" : "\"DATE\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(5,0)) / CAST(2017/12/11 09:30:00 AS DATE))\"" }, "queryContext" : [ { "objectType" : "", @@ -4217,9 +4217,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(10,0)) / CAST(2017/12/11 09:30:00 AS DATE))\"", "left" : "\"DECIMAL(10,0)\"", - "right" : "\"DATE\"" + "right" : "\"DATE\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(10,0)) / CAST(2017/12/11 09:30:00 AS DATE))\"" }, "queryContext" : [ { "objectType" : "", @@ -4241,9 +4241,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(20,0)) / CAST(2017/12/11 09:30:00 AS DATE))\"", "left" : "\"DECIMAL(20,0)\"", - "right" : "\"DATE\"" + "right" : "\"DATE\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(20,0)) / CAST(2017/12/11 09:30:00 AS DATE))\"" }, "queryContext" : [ { "objectType" : "", @@ -4489,9 +4489,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS BINARY) % CAST(1 AS DECIMAL(3,0)))\"", "left" : "\"BINARY\"", - "right" : "\"DECIMAL(3,0)\"" + "right" : "\"DECIMAL(3,0)\"", + "sqlExpr" : "\"(CAST(1 AS BINARY) % CAST(1 AS DECIMAL(3,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -4513,9 +4513,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS BINARY) % CAST(1 AS DECIMAL(5,0)))\"", "left" : "\"BINARY\"", - "right" : "\"DECIMAL(5,0)\"" + "right" : "\"DECIMAL(5,0)\"", + "sqlExpr" : "\"(CAST(1 AS BINARY) % CAST(1 AS DECIMAL(5,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -4537,9 +4537,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS BINARY) % CAST(1 AS DECIMAL(10,0)))\"", "left" : "\"BINARY\"", - "right" : "\"DECIMAL(10,0)\"" + "right" : "\"DECIMAL(10,0)\"", + "sqlExpr" : "\"(CAST(1 AS BINARY) % CAST(1 AS DECIMAL(10,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -4561,9 +4561,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS BINARY) % CAST(1 AS DECIMAL(20,0)))\"", "left" : "\"BINARY\"", - "right" : "\"DECIMAL(20,0)\"" + "right" : "\"DECIMAL(20,0)\"", + "sqlExpr" : "\"(CAST(1 AS BINARY) % CAST(1 AS DECIMAL(20,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -4585,9 +4585,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(2017-12-11 09:30:00.0 AS TIMESTAMP) % CAST(1 AS DECIMAL(3,0)))\"", "left" : "\"TIMESTAMP\"", - "right" : "\"DECIMAL(3,0)\"" + "right" : "\"DECIMAL(3,0)\"", + "sqlExpr" : "\"(CAST(2017-12-11 09:30:00.0 AS TIMESTAMP) % CAST(1 AS DECIMAL(3,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -4609,9 +4609,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(2017-12-11 09:30:00.0 AS TIMESTAMP) % CAST(1 AS DECIMAL(5,0)))\"", "left" : "\"TIMESTAMP\"", - "right" : "\"DECIMAL(5,0)\"" + "right" : "\"DECIMAL(5,0)\"", + "sqlExpr" : "\"(CAST(2017-12-11 09:30:00.0 AS TIMESTAMP) % CAST(1 AS DECIMAL(5,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -4633,9 +4633,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(2017-12-11 09:30:00.0 AS TIMESTAMP) % CAST(1 AS DECIMAL(10,0)))\"", "left" : "\"TIMESTAMP\"", - "right" : "\"DECIMAL(10,0)\"" + "right" : "\"DECIMAL(10,0)\"", + "sqlExpr" : "\"(CAST(2017-12-11 09:30:00.0 AS TIMESTAMP) % CAST(1 AS DECIMAL(10,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -4657,9 +4657,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(2017-12-11 09:30:00.0 AS TIMESTAMP) % CAST(1 AS DECIMAL(20,0)))\"", "left" : "\"TIMESTAMP\"", - "right" : "\"DECIMAL(20,0)\"" + "right" : "\"DECIMAL(20,0)\"", + "sqlExpr" : "\"(CAST(2017-12-11 09:30:00.0 AS TIMESTAMP) % CAST(1 AS DECIMAL(20,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -4681,9 +4681,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(2017-12-11 09:30:00 AS DATE) % CAST(1 AS DECIMAL(3,0)))\"", "left" : "\"DATE\"", - "right" : "\"DECIMAL(3,0)\"" + "right" : "\"DECIMAL(3,0)\"", + "sqlExpr" : "\"(CAST(2017-12-11 09:30:00 AS DATE) % CAST(1 AS DECIMAL(3,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -4705,9 +4705,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(2017-12-11 09:30:00 AS DATE) % CAST(1 AS DECIMAL(5,0)))\"", "left" : "\"DATE\"", - "right" : "\"DECIMAL(5,0)\"" + "right" : "\"DECIMAL(5,0)\"", + "sqlExpr" : "\"(CAST(2017-12-11 09:30:00 AS DATE) % CAST(1 AS DECIMAL(5,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -4729,9 +4729,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(2017-12-11 09:30:00 AS DATE) % CAST(1 AS DECIMAL(10,0)))\"", "left" : "\"DATE\"", - "right" : "\"DECIMAL(10,0)\"" + "right" : "\"DECIMAL(10,0)\"", + "sqlExpr" : "\"(CAST(2017-12-11 09:30:00 AS DATE) % CAST(1 AS DECIMAL(10,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -4753,9 +4753,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(2017-12-11 09:30:00 AS DATE) % CAST(1 AS DECIMAL(20,0)))\"", "left" : "\"DATE\"", - "right" : "\"DECIMAL(20,0)\"" + "right" : "\"DECIMAL(20,0)\"", + "sqlExpr" : "\"(CAST(2017-12-11 09:30:00 AS DATE) % CAST(1 AS DECIMAL(20,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -5033,9 +5033,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(3,0)) % CAST(1 AS BINARY))\"", "left" : "\"DECIMAL(3,0)\"", - "right" : "\"BINARY\"" + "right" : "\"BINARY\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(3,0)) % CAST(1 AS BINARY))\"" }, "queryContext" : [ { "objectType" : "", @@ -5057,9 +5057,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(5,0)) % CAST(1 AS BINARY))\"", "left" : "\"DECIMAL(5,0)\"", - "right" : "\"BINARY\"" + "right" : "\"BINARY\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(5,0)) % CAST(1 AS BINARY))\"" }, "queryContext" : [ { "objectType" : "", @@ -5081,9 +5081,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(10,0)) % CAST(1 AS BINARY))\"", "left" : "\"DECIMAL(10,0)\"", - "right" : "\"BINARY\"" + "right" : "\"BINARY\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(10,0)) % CAST(1 AS BINARY))\"" }, "queryContext" : [ { "objectType" : "", @@ -5105,9 +5105,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(20,0)) % CAST(1 AS BINARY))\"", "left" : "\"DECIMAL(20,0)\"", - "right" : "\"BINARY\"" + "right" : "\"BINARY\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(20,0)) % CAST(1 AS BINARY))\"" }, "queryContext" : [ { "objectType" : "", @@ -5129,9 +5129,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(3,0)) % CAST(1 AS BOOLEAN))\"", "left" : "\"DECIMAL(3,0)\"", - "right" : "\"BOOLEAN\"" + "right" : "\"BOOLEAN\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(3,0)) % CAST(1 AS BOOLEAN))\"" }, "queryContext" : [ { "objectType" : "", @@ -5153,9 +5153,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(5,0)) % CAST(1 AS BOOLEAN))\"", "left" : "\"DECIMAL(5,0)\"", - "right" : "\"BOOLEAN\"" + "right" : "\"BOOLEAN\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(5,0)) % CAST(1 AS BOOLEAN))\"" }, "queryContext" : [ { "objectType" : "", @@ -5177,9 +5177,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(10,0)) % CAST(1 AS BOOLEAN))\"", "left" : "\"DECIMAL(10,0)\"", - "right" : "\"BOOLEAN\"" + "right" : "\"BOOLEAN\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(10,0)) % CAST(1 AS BOOLEAN))\"" }, "queryContext" : [ { "objectType" : "", @@ -5201,9 +5201,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(20,0)) % CAST(1 AS BOOLEAN))\"", "left" : "\"DECIMAL(20,0)\"", - "right" : "\"BOOLEAN\"" + "right" : "\"BOOLEAN\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(20,0)) % CAST(1 AS BOOLEAN))\"" }, "queryContext" : [ { "objectType" : "", @@ -5225,9 +5225,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(3,0)) % CAST(2017-12-11 09:30:00.0 AS TIMESTAMP))\"", "left" : "\"DECIMAL(3,0)\"", - "right" : "\"TIMESTAMP\"" + "right" : "\"TIMESTAMP\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(3,0)) % CAST(2017-12-11 09:30:00.0 AS TIMESTAMP))\"" }, "queryContext" : [ { "objectType" : "", @@ -5249,9 +5249,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(5,0)) % CAST(2017-12-11 09:30:00.0 AS TIMESTAMP))\"", "left" : "\"DECIMAL(5,0)\"", - "right" : "\"TIMESTAMP\"" + "right" : "\"TIMESTAMP\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(5,0)) % CAST(2017-12-11 09:30:00.0 AS TIMESTAMP))\"" }, "queryContext" : [ { "objectType" : "", @@ -5273,9 +5273,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(10,0)) % CAST(2017-12-11 09:30:00.0 AS TIMESTAMP))\"", "left" : "\"DECIMAL(10,0)\"", - "right" : "\"TIMESTAMP\"" + "right" : "\"TIMESTAMP\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(10,0)) % CAST(2017-12-11 09:30:00.0 AS TIMESTAMP))\"" }, "queryContext" : [ { "objectType" : "", @@ -5297,9 +5297,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(20,0)) % CAST(2017-12-11 09:30:00.0 AS TIMESTAMP))\"", "left" : "\"DECIMAL(20,0)\"", - "right" : "\"TIMESTAMP\"" + "right" : "\"TIMESTAMP\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(20,0)) % CAST(2017-12-11 09:30:00.0 AS TIMESTAMP))\"" }, "queryContext" : [ { "objectType" : "", @@ -5321,9 +5321,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(3,0)) % CAST(2017-12-11 09:30:00 AS DATE))\"", "left" : "\"DECIMAL(3,0)\"", - "right" : "\"DATE\"" + "right" : "\"DATE\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(3,0)) % CAST(2017-12-11 09:30:00 AS DATE))\"" }, "queryContext" : [ { "objectType" : "", @@ -5345,9 +5345,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(5,0)) % CAST(2017-12-11 09:30:00 AS DATE))\"", "left" : "\"DECIMAL(5,0)\"", - "right" : "\"DATE\"" + "right" : "\"DATE\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(5,0)) % CAST(2017-12-11 09:30:00 AS DATE))\"" }, "queryContext" : [ { "objectType" : "", @@ -5369,9 +5369,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(10,0)) % CAST(2017-12-11 09:30:00 AS DATE))\"", "left" : "\"DECIMAL(10,0)\"", - "right" : "\"DATE\"" + "right" : "\"DATE\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(10,0)) % CAST(2017-12-11 09:30:00 AS DATE))\"" }, "queryContext" : [ { "objectType" : "", @@ -5393,9 +5393,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(20,0)) % CAST(2017-12-11 09:30:00 AS DATE))\"", "left" : "\"DECIMAL(20,0)\"", - "right" : "\"DATE\"" + "right" : "\"DATE\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(20,0)) % CAST(2017-12-11 09:30:00 AS DATE))\"" }, "queryContext" : [ { "objectType" : "", @@ -5641,9 +5641,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"pmod(CAST(1 AS BINARY), CAST(1 AS DECIMAL(3,0)))\"", "left" : "\"BINARY\"", - "right" : "\"DECIMAL(3,0)\"" + "right" : "\"DECIMAL(3,0)\"", + "sqlExpr" : "\"pmod(CAST(1 AS BINARY), CAST(1 AS DECIMAL(3,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -5665,9 +5665,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"pmod(CAST(1 AS BINARY), CAST(1 AS DECIMAL(5,0)))\"", "left" : "\"BINARY\"", - "right" : "\"DECIMAL(5,0)\"" + "right" : "\"DECIMAL(5,0)\"", + "sqlExpr" : "\"pmod(CAST(1 AS BINARY), CAST(1 AS DECIMAL(5,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -5689,9 +5689,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"pmod(CAST(1 AS BINARY), CAST(1 AS DECIMAL(10,0)))\"", "left" : "\"BINARY\"", - "right" : "\"DECIMAL(10,0)\"" + "right" : "\"DECIMAL(10,0)\"", + "sqlExpr" : "\"pmod(CAST(1 AS BINARY), CAST(1 AS DECIMAL(10,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -5713,9 +5713,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"pmod(CAST(1 AS BINARY), CAST(1 AS DECIMAL(20,0)))\"", "left" : "\"BINARY\"", - "right" : "\"DECIMAL(20,0)\"" + "right" : "\"DECIMAL(20,0)\"", + "sqlExpr" : "\"pmod(CAST(1 AS BINARY), CAST(1 AS DECIMAL(20,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -5737,9 +5737,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"pmod(CAST(2017-12-11 09:30:00.0 AS TIMESTAMP), CAST(1 AS DECIMAL(3,0)))\"", "left" : "\"TIMESTAMP\"", - "right" : "\"DECIMAL(3,0)\"" + "right" : "\"DECIMAL(3,0)\"", + "sqlExpr" : "\"pmod(CAST(2017-12-11 09:30:00.0 AS TIMESTAMP), CAST(1 AS DECIMAL(3,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -5761,9 +5761,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"pmod(CAST(2017-12-11 09:30:00.0 AS TIMESTAMP), CAST(1 AS DECIMAL(5,0)))\"", "left" : "\"TIMESTAMP\"", - "right" : "\"DECIMAL(5,0)\"" + "right" : "\"DECIMAL(5,0)\"", + "sqlExpr" : "\"pmod(CAST(2017-12-11 09:30:00.0 AS TIMESTAMP), CAST(1 AS DECIMAL(5,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -5785,9 +5785,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"pmod(CAST(2017-12-11 09:30:00.0 AS TIMESTAMP), CAST(1 AS DECIMAL(10,0)))\"", "left" : "\"TIMESTAMP\"", - "right" : "\"DECIMAL(10,0)\"" + "right" : "\"DECIMAL(10,0)\"", + "sqlExpr" : "\"pmod(CAST(2017-12-11 09:30:00.0 AS TIMESTAMP), CAST(1 AS DECIMAL(10,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -5809,9 +5809,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"pmod(CAST(2017-12-11 09:30:00.0 AS TIMESTAMP), CAST(1 AS DECIMAL(20,0)))\"", "left" : "\"TIMESTAMP\"", - "right" : "\"DECIMAL(20,0)\"" + "right" : "\"DECIMAL(20,0)\"", + "sqlExpr" : "\"pmod(CAST(2017-12-11 09:30:00.0 AS TIMESTAMP), CAST(1 AS DECIMAL(20,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -5833,9 +5833,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"pmod(CAST(2017-12-11 09:30:00 AS DATE), CAST(1 AS DECIMAL(3,0)))\"", "left" : "\"DATE\"", - "right" : "\"DECIMAL(3,0)\"" + "right" : "\"DECIMAL(3,0)\"", + "sqlExpr" : "\"pmod(CAST(2017-12-11 09:30:00 AS DATE), CAST(1 AS DECIMAL(3,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -5857,9 +5857,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"pmod(CAST(2017-12-11 09:30:00 AS DATE), CAST(1 AS DECIMAL(5,0)))\"", "left" : "\"DATE\"", - "right" : "\"DECIMAL(5,0)\"" + "right" : "\"DECIMAL(5,0)\"", + "sqlExpr" : "\"pmod(CAST(2017-12-11 09:30:00 AS DATE), CAST(1 AS DECIMAL(5,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -5881,9 +5881,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"pmod(CAST(2017-12-11 09:30:00 AS DATE), CAST(1 AS DECIMAL(10,0)))\"", "left" : "\"DATE\"", - "right" : "\"DECIMAL(10,0)\"" + "right" : "\"DECIMAL(10,0)\"", + "sqlExpr" : "\"pmod(CAST(2017-12-11 09:30:00 AS DATE), CAST(1 AS DECIMAL(10,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -5905,9 +5905,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"pmod(CAST(2017-12-11 09:30:00 AS DATE), CAST(1 AS DECIMAL(20,0)))\"", "left" : "\"DATE\"", - "right" : "\"DECIMAL(20,0)\"" + "right" : "\"DECIMAL(20,0)\"", + "sqlExpr" : "\"pmod(CAST(2017-12-11 09:30:00 AS DATE), CAST(1 AS DECIMAL(20,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -6185,9 +6185,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"pmod(CAST(1 AS DECIMAL(3,0)), CAST(1 AS BINARY))\"", "left" : "\"DECIMAL(3,0)\"", - "right" : "\"BINARY\"" + "right" : "\"BINARY\"", + "sqlExpr" : "\"pmod(CAST(1 AS DECIMAL(3,0)), CAST(1 AS BINARY))\"" }, "queryContext" : [ { "objectType" : "", @@ -6209,9 +6209,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"pmod(CAST(1 AS DECIMAL(5,0)), CAST(1 AS BINARY))\"", "left" : "\"DECIMAL(5,0)\"", - "right" : "\"BINARY\"" + "right" : "\"BINARY\"", + "sqlExpr" : "\"pmod(CAST(1 AS DECIMAL(5,0)), CAST(1 AS BINARY))\"" }, "queryContext" : [ { "objectType" : "", @@ -6233,9 +6233,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"pmod(CAST(1 AS DECIMAL(10,0)), CAST(1 AS BINARY))\"", "left" : "\"DECIMAL(10,0)\"", - "right" : "\"BINARY\"" + "right" : "\"BINARY\"", + "sqlExpr" : "\"pmod(CAST(1 AS DECIMAL(10,0)), CAST(1 AS BINARY))\"" }, "queryContext" : [ { "objectType" : "", @@ -6257,9 +6257,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"pmod(CAST(1 AS DECIMAL(20,0)), CAST(1 AS BINARY))\"", "left" : "\"DECIMAL(20,0)\"", - "right" : "\"BINARY\"" + "right" : "\"BINARY\"", + "sqlExpr" : "\"pmod(CAST(1 AS DECIMAL(20,0)), CAST(1 AS BINARY))\"" }, "queryContext" : [ { "objectType" : "", @@ -6281,9 +6281,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"pmod(CAST(1 AS DECIMAL(3,0)), CAST(1 AS BOOLEAN))\"", "left" : "\"DECIMAL(3,0)\"", - "right" : "\"BOOLEAN\"" + "right" : "\"BOOLEAN\"", + "sqlExpr" : "\"pmod(CAST(1 AS DECIMAL(3,0)), CAST(1 AS BOOLEAN))\"" }, "queryContext" : [ { "objectType" : "", @@ -6305,9 +6305,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"pmod(CAST(1 AS DECIMAL(5,0)), CAST(1 AS BOOLEAN))\"", "left" : "\"DECIMAL(5,0)\"", - "right" : "\"BOOLEAN\"" + "right" : "\"BOOLEAN\"", + "sqlExpr" : "\"pmod(CAST(1 AS DECIMAL(5,0)), CAST(1 AS BOOLEAN))\"" }, "queryContext" : [ { "objectType" : "", @@ -6329,9 +6329,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"pmod(CAST(1 AS DECIMAL(10,0)), CAST(1 AS BOOLEAN))\"", "left" : "\"DECIMAL(10,0)\"", - "right" : "\"BOOLEAN\"" + "right" : "\"BOOLEAN\"", + "sqlExpr" : "\"pmod(CAST(1 AS DECIMAL(10,0)), CAST(1 AS BOOLEAN))\"" }, "queryContext" : [ { "objectType" : "", @@ -6353,9 +6353,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"pmod(CAST(1 AS DECIMAL(20,0)), CAST(1 AS BOOLEAN))\"", "left" : "\"DECIMAL(20,0)\"", - "right" : "\"BOOLEAN\"" + "right" : "\"BOOLEAN\"", + "sqlExpr" : "\"pmod(CAST(1 AS DECIMAL(20,0)), CAST(1 AS BOOLEAN))\"" }, "queryContext" : [ { "objectType" : "", @@ -6377,9 +6377,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"pmod(CAST(1 AS DECIMAL(3,0)), CAST(2017-12-11 09:30:00.0 AS TIMESTAMP))\"", "left" : "\"DECIMAL(3,0)\"", - "right" : "\"TIMESTAMP\"" + "right" : "\"TIMESTAMP\"", + "sqlExpr" : "\"pmod(CAST(1 AS DECIMAL(3,0)), CAST(2017-12-11 09:30:00.0 AS TIMESTAMP))\"" }, "queryContext" : [ { "objectType" : "", @@ -6401,9 +6401,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"pmod(CAST(1 AS DECIMAL(5,0)), CAST(2017-12-11 09:30:00.0 AS TIMESTAMP))\"", "left" : "\"DECIMAL(5,0)\"", - "right" : "\"TIMESTAMP\"" + "right" : "\"TIMESTAMP\"", + "sqlExpr" : "\"pmod(CAST(1 AS DECIMAL(5,0)), CAST(2017-12-11 09:30:00.0 AS TIMESTAMP))\"" }, "queryContext" : [ { "objectType" : "", @@ -6425,9 +6425,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"pmod(CAST(1 AS DECIMAL(10,0)), CAST(2017-12-11 09:30:00.0 AS TIMESTAMP))\"", "left" : "\"DECIMAL(10,0)\"", - "right" : "\"TIMESTAMP\"" + "right" : "\"TIMESTAMP\"", + "sqlExpr" : "\"pmod(CAST(1 AS DECIMAL(10,0)), CAST(2017-12-11 09:30:00.0 AS TIMESTAMP))\"" }, "queryContext" : [ { "objectType" : "", @@ -6449,9 +6449,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"pmod(CAST(1 AS DECIMAL(20,0)), CAST(2017-12-11 09:30:00.0 AS TIMESTAMP))\"", "left" : "\"DECIMAL(20,0)\"", - "right" : "\"TIMESTAMP\"" + "right" : "\"TIMESTAMP\"", + "sqlExpr" : "\"pmod(CAST(1 AS DECIMAL(20,0)), CAST(2017-12-11 09:30:00.0 AS TIMESTAMP))\"" }, "queryContext" : [ { "objectType" : "", @@ -6473,9 +6473,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"pmod(CAST(1 AS DECIMAL(3,0)), CAST(2017-12-11 09:30:00 AS DATE))\"", "left" : "\"DECIMAL(3,0)\"", - "right" : "\"DATE\"" + "right" : "\"DATE\"", + "sqlExpr" : "\"pmod(CAST(1 AS DECIMAL(3,0)), CAST(2017-12-11 09:30:00 AS DATE))\"" }, "queryContext" : [ { "objectType" : "", @@ -6497,9 +6497,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"pmod(CAST(1 AS DECIMAL(5,0)), CAST(2017-12-11 09:30:00 AS DATE))\"", "left" : "\"DECIMAL(5,0)\"", - "right" : "\"DATE\"" + "right" : "\"DATE\"", + "sqlExpr" : "\"pmod(CAST(1 AS DECIMAL(5,0)), CAST(2017-12-11 09:30:00 AS DATE))\"" }, "queryContext" : [ { "objectType" : "", @@ -6521,9 +6521,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"pmod(CAST(1 AS DECIMAL(10,0)), CAST(2017-12-11 09:30:00 AS DATE))\"", "left" : "\"DECIMAL(10,0)\"", - "right" : "\"DATE\"" + "right" : "\"DATE\"", + "sqlExpr" : "\"pmod(CAST(1 AS DECIMAL(10,0)), CAST(2017-12-11 09:30:00 AS DATE))\"" }, "queryContext" : [ { "objectType" : "", @@ -6545,9 +6545,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"pmod(CAST(1 AS DECIMAL(20,0)), CAST(2017-12-11 09:30:00 AS DATE))\"", "left" : "\"DECIMAL(20,0)\"", - "right" : "\"DATE\"" + "right" : "\"DATE\"", + "sqlExpr" : "\"pmod(CAST(1 AS DECIMAL(20,0)), CAST(2017-12-11 09:30:00 AS DATE))\"" }, "queryContext" : [ { "objectType" : "", @@ -6793,9 +6793,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS BINARY) = CAST(1 AS DECIMAL(3,0)))\"", "left" : "\"BINARY\"", - "right" : "\"DECIMAL(3,0)\"" + "right" : "\"DECIMAL(3,0)\"", + "sqlExpr" : "\"(CAST(1 AS BINARY) = CAST(1 AS DECIMAL(3,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -6817,9 +6817,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS BINARY) = CAST(1 AS DECIMAL(5,0)))\"", "left" : "\"BINARY\"", - "right" : "\"DECIMAL(5,0)\"" + "right" : "\"DECIMAL(5,0)\"", + "sqlExpr" : "\"(CAST(1 AS BINARY) = CAST(1 AS DECIMAL(5,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -6841,9 +6841,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS BINARY) = CAST(1 AS DECIMAL(10,0)))\"", "left" : "\"BINARY\"", - "right" : "\"DECIMAL(10,0)\"" + "right" : "\"DECIMAL(10,0)\"", + "sqlExpr" : "\"(CAST(1 AS BINARY) = CAST(1 AS DECIMAL(10,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -6865,9 +6865,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS BINARY) = CAST(1 AS DECIMAL(20,0)))\"", "left" : "\"BINARY\"", - "right" : "\"DECIMAL(20,0)\"" + "right" : "\"DECIMAL(20,0)\"", + "sqlExpr" : "\"(CAST(1 AS BINARY) = CAST(1 AS DECIMAL(20,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -6889,9 +6889,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(2017-12-11 09:30:00.0 AS TIMESTAMP) = CAST(1 AS DECIMAL(3,0)))\"", "left" : "\"TIMESTAMP\"", - "right" : "\"DECIMAL(3,0)\"" + "right" : "\"DECIMAL(3,0)\"", + "sqlExpr" : "\"(CAST(2017-12-11 09:30:00.0 AS TIMESTAMP) = CAST(1 AS DECIMAL(3,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -6913,9 +6913,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(2017-12-11 09:30:00.0 AS TIMESTAMP) = CAST(1 AS DECIMAL(5,0)))\"", "left" : "\"TIMESTAMP\"", - "right" : "\"DECIMAL(5,0)\"" + "right" : "\"DECIMAL(5,0)\"", + "sqlExpr" : "\"(CAST(2017-12-11 09:30:00.0 AS TIMESTAMP) = CAST(1 AS DECIMAL(5,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -6937,9 +6937,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(2017-12-11 09:30:00.0 AS TIMESTAMP) = CAST(1 AS DECIMAL(10,0)))\"", "left" : "\"TIMESTAMP\"", - "right" : "\"DECIMAL(10,0)\"" + "right" : "\"DECIMAL(10,0)\"", + "sqlExpr" : "\"(CAST(2017-12-11 09:30:00.0 AS TIMESTAMP) = CAST(1 AS DECIMAL(10,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -6961,9 +6961,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(2017-12-11 09:30:00.0 AS TIMESTAMP) = CAST(1 AS DECIMAL(20,0)))\"", "left" : "\"TIMESTAMP\"", - "right" : "\"DECIMAL(20,0)\"" + "right" : "\"DECIMAL(20,0)\"", + "sqlExpr" : "\"(CAST(2017-12-11 09:30:00.0 AS TIMESTAMP) = CAST(1 AS DECIMAL(20,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -6985,9 +6985,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(2017-12-11 09:30:00 AS DATE) = CAST(1 AS DECIMAL(3,0)))\"", "left" : "\"DATE\"", - "right" : "\"DECIMAL(3,0)\"" + "right" : "\"DECIMAL(3,0)\"", + "sqlExpr" : "\"(CAST(2017-12-11 09:30:00 AS DATE) = CAST(1 AS DECIMAL(3,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -7009,9 +7009,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(2017-12-11 09:30:00 AS DATE) = CAST(1 AS DECIMAL(5,0)))\"", "left" : "\"DATE\"", - "right" : "\"DECIMAL(5,0)\"" + "right" : "\"DECIMAL(5,0)\"", + "sqlExpr" : "\"(CAST(2017-12-11 09:30:00 AS DATE) = CAST(1 AS DECIMAL(5,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -7033,9 +7033,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(2017-12-11 09:30:00 AS DATE) = CAST(1 AS DECIMAL(10,0)))\"", "left" : "\"DATE\"", - "right" : "\"DECIMAL(10,0)\"" + "right" : "\"DECIMAL(10,0)\"", + "sqlExpr" : "\"(CAST(2017-12-11 09:30:00 AS DATE) = CAST(1 AS DECIMAL(10,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -7057,9 +7057,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(2017-12-11 09:30:00 AS DATE) = CAST(1 AS DECIMAL(20,0)))\"", "left" : "\"DATE\"", - "right" : "\"DECIMAL(20,0)\"" + "right" : "\"DECIMAL(20,0)\"", + "sqlExpr" : "\"(CAST(2017-12-11 09:30:00 AS DATE) = CAST(1 AS DECIMAL(20,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -7337,9 +7337,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(3,0)) = CAST(1 AS BINARY))\"", "left" : "\"DECIMAL(3,0)\"", - "right" : "\"BINARY\"" + "right" : "\"BINARY\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(3,0)) = CAST(1 AS BINARY))\"" }, "queryContext" : [ { "objectType" : "", @@ -7361,9 +7361,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(5,0)) = CAST(1 AS BINARY))\"", "left" : "\"DECIMAL(5,0)\"", - "right" : "\"BINARY\"" + "right" : "\"BINARY\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(5,0)) = CAST(1 AS BINARY))\"" }, "queryContext" : [ { "objectType" : "", @@ -7385,9 +7385,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(10,0)) = CAST(1 AS BINARY))\"", "left" : "\"DECIMAL(10,0)\"", - "right" : "\"BINARY\"" + "right" : "\"BINARY\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(10,0)) = CAST(1 AS BINARY))\"" }, "queryContext" : [ { "objectType" : "", @@ -7409,9 +7409,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(20,0)) = CAST(1 AS BINARY))\"", "left" : "\"DECIMAL(20,0)\"", - "right" : "\"BINARY\"" + "right" : "\"BINARY\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(20,0)) = CAST(1 AS BINARY))\"" }, "queryContext" : [ { "objectType" : "", @@ -7465,9 +7465,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(3,0)) = CAST(2017-12-11 09:30:00.0 AS TIMESTAMP))\"", "left" : "\"DECIMAL(3,0)\"", - "right" : "\"TIMESTAMP\"" + "right" : "\"TIMESTAMP\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(3,0)) = CAST(2017-12-11 09:30:00.0 AS TIMESTAMP))\"" }, "queryContext" : [ { "objectType" : "", @@ -7489,9 +7489,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(5,0)) = CAST(2017-12-11 09:30:00.0 AS TIMESTAMP))\"", "left" : "\"DECIMAL(5,0)\"", - "right" : "\"TIMESTAMP\"" + "right" : "\"TIMESTAMP\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(5,0)) = CAST(2017-12-11 09:30:00.0 AS TIMESTAMP))\"" }, "queryContext" : [ { "objectType" : "", @@ -7513,9 +7513,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(10,0)) = CAST(2017-12-11 09:30:00.0 AS TIMESTAMP))\"", "left" : "\"DECIMAL(10,0)\"", - "right" : "\"TIMESTAMP\"" + "right" : "\"TIMESTAMP\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(10,0)) = CAST(2017-12-11 09:30:00.0 AS TIMESTAMP))\"" }, "queryContext" : [ { "objectType" : "", @@ -7537,9 +7537,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(20,0)) = CAST(2017-12-11 09:30:00.0 AS TIMESTAMP))\"", "left" : "\"DECIMAL(20,0)\"", - "right" : "\"TIMESTAMP\"" + "right" : "\"TIMESTAMP\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(20,0)) = CAST(2017-12-11 09:30:00.0 AS TIMESTAMP))\"" }, "queryContext" : [ { "objectType" : "", @@ -7561,9 +7561,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(3,0)) = CAST(2017-12-11 09:30:00 AS DATE))\"", "left" : "\"DECIMAL(3,0)\"", - "right" : "\"DATE\"" + "right" : "\"DATE\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(3,0)) = CAST(2017-12-11 09:30:00 AS DATE))\"" }, "queryContext" : [ { "objectType" : "", @@ -7585,9 +7585,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(5,0)) = CAST(2017-12-11 09:30:00 AS DATE))\"", "left" : "\"DECIMAL(5,0)\"", - "right" : "\"DATE\"" + "right" : "\"DATE\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(5,0)) = CAST(2017-12-11 09:30:00 AS DATE))\"" }, "queryContext" : [ { "objectType" : "", @@ -7609,9 +7609,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(10,0)) = CAST(2017-12-11 09:30:00 AS DATE))\"", "left" : "\"DECIMAL(10,0)\"", - "right" : "\"DATE\"" + "right" : "\"DATE\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(10,0)) = CAST(2017-12-11 09:30:00 AS DATE))\"" }, "queryContext" : [ { "objectType" : "", @@ -7633,9 +7633,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(20,0)) = CAST(2017-12-11 09:30:00 AS DATE))\"", "left" : "\"DECIMAL(20,0)\"", - "right" : "\"DATE\"" + "right" : "\"DATE\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(20,0)) = CAST(2017-12-11 09:30:00 AS DATE))\"" }, "queryContext" : [ { "objectType" : "", @@ -7881,9 +7881,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS BINARY) <=> CAST(1 AS DECIMAL(3,0)))\"", "left" : "\"BINARY\"", - "right" : "\"DECIMAL(3,0)\"" + "right" : "\"DECIMAL(3,0)\"", + "sqlExpr" : "\"(CAST(1 AS BINARY) <=> CAST(1 AS DECIMAL(3,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -7905,9 +7905,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS BINARY) <=> CAST(1 AS DECIMAL(5,0)))\"", "left" : "\"BINARY\"", - "right" : "\"DECIMAL(5,0)\"" + "right" : "\"DECIMAL(5,0)\"", + "sqlExpr" : "\"(CAST(1 AS BINARY) <=> CAST(1 AS DECIMAL(5,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -7929,9 +7929,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS BINARY) <=> CAST(1 AS DECIMAL(10,0)))\"", "left" : "\"BINARY\"", - "right" : "\"DECIMAL(10,0)\"" + "right" : "\"DECIMAL(10,0)\"", + "sqlExpr" : "\"(CAST(1 AS BINARY) <=> CAST(1 AS DECIMAL(10,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -7953,9 +7953,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS BINARY) <=> CAST(1 AS DECIMAL(20,0)))\"", "left" : "\"BINARY\"", - "right" : "\"DECIMAL(20,0)\"" + "right" : "\"DECIMAL(20,0)\"", + "sqlExpr" : "\"(CAST(1 AS BINARY) <=> CAST(1 AS DECIMAL(20,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -7977,9 +7977,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(2017-12-11 09:30:00.0 AS TIMESTAMP) <=> CAST(1 AS DECIMAL(3,0)))\"", "left" : "\"TIMESTAMP\"", - "right" : "\"DECIMAL(3,0)\"" + "right" : "\"DECIMAL(3,0)\"", + "sqlExpr" : "\"(CAST(2017-12-11 09:30:00.0 AS TIMESTAMP) <=> CAST(1 AS DECIMAL(3,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -8001,9 +8001,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(2017-12-11 09:30:00.0 AS TIMESTAMP) <=> CAST(1 AS DECIMAL(5,0)))\"", "left" : "\"TIMESTAMP\"", - "right" : "\"DECIMAL(5,0)\"" + "right" : "\"DECIMAL(5,0)\"", + "sqlExpr" : "\"(CAST(2017-12-11 09:30:00.0 AS TIMESTAMP) <=> CAST(1 AS DECIMAL(5,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -8025,9 +8025,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(2017-12-11 09:30:00.0 AS TIMESTAMP) <=> CAST(1 AS DECIMAL(10,0)))\"", "left" : "\"TIMESTAMP\"", - "right" : "\"DECIMAL(10,0)\"" + "right" : "\"DECIMAL(10,0)\"", + "sqlExpr" : "\"(CAST(2017-12-11 09:30:00.0 AS TIMESTAMP) <=> CAST(1 AS DECIMAL(10,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -8049,9 +8049,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(2017-12-11 09:30:00.0 AS TIMESTAMP) <=> CAST(1 AS DECIMAL(20,0)))\"", "left" : "\"TIMESTAMP\"", - "right" : "\"DECIMAL(20,0)\"" + "right" : "\"DECIMAL(20,0)\"", + "sqlExpr" : "\"(CAST(2017-12-11 09:30:00.0 AS TIMESTAMP) <=> CAST(1 AS DECIMAL(20,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -8073,9 +8073,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(2017-12-11 09:30:00 AS DATE) <=> CAST(1 AS DECIMAL(3,0)))\"", "left" : "\"DATE\"", - "right" : "\"DECIMAL(3,0)\"" + "right" : "\"DECIMAL(3,0)\"", + "sqlExpr" : "\"(CAST(2017-12-11 09:30:00 AS DATE) <=> CAST(1 AS DECIMAL(3,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -8097,9 +8097,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(2017-12-11 09:30:00 AS DATE) <=> CAST(1 AS DECIMAL(5,0)))\"", "left" : "\"DATE\"", - "right" : "\"DECIMAL(5,0)\"" + "right" : "\"DECIMAL(5,0)\"", + "sqlExpr" : "\"(CAST(2017-12-11 09:30:00 AS DATE) <=> CAST(1 AS DECIMAL(5,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -8121,9 +8121,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(2017-12-11 09:30:00 AS DATE) <=> CAST(1 AS DECIMAL(10,0)))\"", "left" : "\"DATE\"", - "right" : "\"DECIMAL(10,0)\"" + "right" : "\"DECIMAL(10,0)\"", + "sqlExpr" : "\"(CAST(2017-12-11 09:30:00 AS DATE) <=> CAST(1 AS DECIMAL(10,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -8145,9 +8145,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(2017-12-11 09:30:00 AS DATE) <=> CAST(1 AS DECIMAL(20,0)))\"", "left" : "\"DATE\"", - "right" : "\"DECIMAL(20,0)\"" + "right" : "\"DECIMAL(20,0)\"", + "sqlExpr" : "\"(CAST(2017-12-11 09:30:00 AS DATE) <=> CAST(1 AS DECIMAL(20,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -8425,9 +8425,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(3,0)) <=> CAST(1 AS BINARY))\"", "left" : "\"DECIMAL(3,0)\"", - "right" : "\"BINARY\"" + "right" : "\"BINARY\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(3,0)) <=> CAST(1 AS BINARY))\"" }, "queryContext" : [ { "objectType" : "", @@ -8449,9 +8449,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(5,0)) <=> CAST(1 AS BINARY))\"", "left" : "\"DECIMAL(5,0)\"", - "right" : "\"BINARY\"" + "right" : "\"BINARY\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(5,0)) <=> CAST(1 AS BINARY))\"" }, "queryContext" : [ { "objectType" : "", @@ -8473,9 +8473,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(10,0)) <=> CAST(1 AS BINARY))\"", "left" : "\"DECIMAL(10,0)\"", - "right" : "\"BINARY\"" + "right" : "\"BINARY\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(10,0)) <=> CAST(1 AS BINARY))\"" }, "queryContext" : [ { "objectType" : "", @@ -8497,9 +8497,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(20,0)) <=> CAST(1 AS BINARY))\"", "left" : "\"DECIMAL(20,0)\"", - "right" : "\"BINARY\"" + "right" : "\"BINARY\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(20,0)) <=> CAST(1 AS BINARY))\"" }, "queryContext" : [ { "objectType" : "", @@ -8553,9 +8553,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(3,0)) <=> CAST(2017-12-11 09:30:00.0 AS TIMESTAMP))\"", "left" : "\"DECIMAL(3,0)\"", - "right" : "\"TIMESTAMP\"" + "right" : "\"TIMESTAMP\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(3,0)) <=> CAST(2017-12-11 09:30:00.0 AS TIMESTAMP))\"" }, "queryContext" : [ { "objectType" : "", @@ -8577,9 +8577,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(5,0)) <=> CAST(2017-12-11 09:30:00.0 AS TIMESTAMP))\"", "left" : "\"DECIMAL(5,0)\"", - "right" : "\"TIMESTAMP\"" + "right" : "\"TIMESTAMP\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(5,0)) <=> CAST(2017-12-11 09:30:00.0 AS TIMESTAMP))\"" }, "queryContext" : [ { "objectType" : "", @@ -8601,9 +8601,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(10,0)) <=> CAST(2017-12-11 09:30:00.0 AS TIMESTAMP))\"", "left" : "\"DECIMAL(10,0)\"", - "right" : "\"TIMESTAMP\"" + "right" : "\"TIMESTAMP\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(10,0)) <=> CAST(2017-12-11 09:30:00.0 AS TIMESTAMP))\"" }, "queryContext" : [ { "objectType" : "", @@ -8625,9 +8625,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(20,0)) <=> CAST(2017-12-11 09:30:00.0 AS TIMESTAMP))\"", "left" : "\"DECIMAL(20,0)\"", - "right" : "\"TIMESTAMP\"" + "right" : "\"TIMESTAMP\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(20,0)) <=> CAST(2017-12-11 09:30:00.0 AS TIMESTAMP))\"" }, "queryContext" : [ { "objectType" : "", @@ -8649,9 +8649,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(3,0)) <=> CAST(2017-12-11 09:30:00 AS DATE))\"", "left" : "\"DECIMAL(3,0)\"", - "right" : "\"DATE\"" + "right" : "\"DATE\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(3,0)) <=> CAST(2017-12-11 09:30:00 AS DATE))\"" }, "queryContext" : [ { "objectType" : "", @@ -8673,9 +8673,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(5,0)) <=> CAST(2017-12-11 09:30:00 AS DATE))\"", "left" : "\"DECIMAL(5,0)\"", - "right" : "\"DATE\"" + "right" : "\"DATE\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(5,0)) <=> CAST(2017-12-11 09:30:00 AS DATE))\"" }, "queryContext" : [ { "objectType" : "", @@ -8697,9 +8697,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(10,0)) <=> CAST(2017-12-11 09:30:00 AS DATE))\"", "left" : "\"DECIMAL(10,0)\"", - "right" : "\"DATE\"" + "right" : "\"DATE\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(10,0)) <=> CAST(2017-12-11 09:30:00 AS DATE))\"" }, "queryContext" : [ { "objectType" : "", @@ -8721,9 +8721,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(20,0)) <=> CAST(2017-12-11 09:30:00 AS DATE))\"", "left" : "\"DECIMAL(20,0)\"", - "right" : "\"DATE\"" + "right" : "\"DATE\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(20,0)) <=> CAST(2017-12-11 09:30:00 AS DATE))\"" }, "queryContext" : [ { "objectType" : "", @@ -8969,9 +8969,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS BINARY) < CAST(1 AS DECIMAL(3,0)))\"", "left" : "\"BINARY\"", - "right" : "\"DECIMAL(3,0)\"" + "right" : "\"DECIMAL(3,0)\"", + "sqlExpr" : "\"(CAST(1 AS BINARY) < CAST(1 AS DECIMAL(3,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -8993,9 +8993,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS BINARY) < CAST(1 AS DECIMAL(5,0)))\"", "left" : "\"BINARY\"", - "right" : "\"DECIMAL(5,0)\"" + "right" : "\"DECIMAL(5,0)\"", + "sqlExpr" : "\"(CAST(1 AS BINARY) < CAST(1 AS DECIMAL(5,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -9017,9 +9017,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS BINARY) < CAST(1 AS DECIMAL(10,0)))\"", "left" : "\"BINARY\"", - "right" : "\"DECIMAL(10,0)\"" + "right" : "\"DECIMAL(10,0)\"", + "sqlExpr" : "\"(CAST(1 AS BINARY) < CAST(1 AS DECIMAL(10,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -9041,9 +9041,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS BINARY) < CAST(1 AS DECIMAL(20,0)))\"", "left" : "\"BINARY\"", - "right" : "\"DECIMAL(20,0)\"" + "right" : "\"DECIMAL(20,0)\"", + "sqlExpr" : "\"(CAST(1 AS BINARY) < CAST(1 AS DECIMAL(20,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -9065,9 +9065,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(2017-12-11 09:30:00.0 AS TIMESTAMP) < CAST(1 AS DECIMAL(3,0)))\"", "left" : "\"TIMESTAMP\"", - "right" : "\"DECIMAL(3,0)\"" + "right" : "\"DECIMAL(3,0)\"", + "sqlExpr" : "\"(CAST(2017-12-11 09:30:00.0 AS TIMESTAMP) < CAST(1 AS DECIMAL(3,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -9089,9 +9089,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(2017-12-11 09:30:00.0 AS TIMESTAMP) < CAST(1 AS DECIMAL(5,0)))\"", "left" : "\"TIMESTAMP\"", - "right" : "\"DECIMAL(5,0)\"" + "right" : "\"DECIMAL(5,0)\"", + "sqlExpr" : "\"(CAST(2017-12-11 09:30:00.0 AS TIMESTAMP) < CAST(1 AS DECIMAL(5,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -9113,9 +9113,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(2017-12-11 09:30:00.0 AS TIMESTAMP) < CAST(1 AS DECIMAL(10,0)))\"", "left" : "\"TIMESTAMP\"", - "right" : "\"DECIMAL(10,0)\"" + "right" : "\"DECIMAL(10,0)\"", + "sqlExpr" : "\"(CAST(2017-12-11 09:30:00.0 AS TIMESTAMP) < CAST(1 AS DECIMAL(10,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -9137,9 +9137,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(2017-12-11 09:30:00.0 AS TIMESTAMP) < CAST(1 AS DECIMAL(20,0)))\"", "left" : "\"TIMESTAMP\"", - "right" : "\"DECIMAL(20,0)\"" + "right" : "\"DECIMAL(20,0)\"", + "sqlExpr" : "\"(CAST(2017-12-11 09:30:00.0 AS TIMESTAMP) < CAST(1 AS DECIMAL(20,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -9161,9 +9161,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(2017-12-11 09:30:00 AS DATE) < CAST(1 AS DECIMAL(3,0)))\"", "left" : "\"DATE\"", - "right" : "\"DECIMAL(3,0)\"" + "right" : "\"DECIMAL(3,0)\"", + "sqlExpr" : "\"(CAST(2017-12-11 09:30:00 AS DATE) < CAST(1 AS DECIMAL(3,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -9185,9 +9185,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(2017-12-11 09:30:00 AS DATE) < CAST(1 AS DECIMAL(5,0)))\"", "left" : "\"DATE\"", - "right" : "\"DECIMAL(5,0)\"" + "right" : "\"DECIMAL(5,0)\"", + "sqlExpr" : "\"(CAST(2017-12-11 09:30:00 AS DATE) < CAST(1 AS DECIMAL(5,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -9209,9 +9209,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(2017-12-11 09:30:00 AS DATE) < CAST(1 AS DECIMAL(10,0)))\"", "left" : "\"DATE\"", - "right" : "\"DECIMAL(10,0)\"" + "right" : "\"DECIMAL(10,0)\"", + "sqlExpr" : "\"(CAST(2017-12-11 09:30:00 AS DATE) < CAST(1 AS DECIMAL(10,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -9233,9 +9233,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(2017-12-11 09:30:00 AS DATE) < CAST(1 AS DECIMAL(20,0)))\"", "left" : "\"DATE\"", - "right" : "\"DECIMAL(20,0)\"" + "right" : "\"DECIMAL(20,0)\"", + "sqlExpr" : "\"(CAST(2017-12-11 09:30:00 AS DATE) < CAST(1 AS DECIMAL(20,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -9513,9 +9513,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(3,0)) < CAST(1 AS BINARY))\"", "left" : "\"DECIMAL(3,0)\"", - "right" : "\"BINARY\"" + "right" : "\"BINARY\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(3,0)) < CAST(1 AS BINARY))\"" }, "queryContext" : [ { "objectType" : "", @@ -9537,9 +9537,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(5,0)) < CAST(1 AS BINARY))\"", "left" : "\"DECIMAL(5,0)\"", - "right" : "\"BINARY\"" + "right" : "\"BINARY\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(5,0)) < CAST(1 AS BINARY))\"" }, "queryContext" : [ { "objectType" : "", @@ -9561,9 +9561,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(10,0)) < CAST(1 AS BINARY))\"", "left" : "\"DECIMAL(10,0)\"", - "right" : "\"BINARY\"" + "right" : "\"BINARY\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(10,0)) < CAST(1 AS BINARY))\"" }, "queryContext" : [ { "objectType" : "", @@ -9585,9 +9585,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(20,0)) < CAST(1 AS BINARY))\"", "left" : "\"DECIMAL(20,0)\"", - "right" : "\"BINARY\"" + "right" : "\"BINARY\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(20,0)) < CAST(1 AS BINARY))\"" }, "queryContext" : [ { "objectType" : "", @@ -9609,9 +9609,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(3,0)) < CAST(1 AS BOOLEAN))\"", "left" : "\"DECIMAL(3,0)\"", - "right" : "\"BOOLEAN\"" + "right" : "\"BOOLEAN\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(3,0)) < CAST(1 AS BOOLEAN))\"" }, "queryContext" : [ { "objectType" : "", @@ -9633,9 +9633,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(5,0)) < CAST(1 AS BOOLEAN))\"", "left" : "\"DECIMAL(5,0)\"", - "right" : "\"BOOLEAN\"" + "right" : "\"BOOLEAN\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(5,0)) < CAST(1 AS BOOLEAN))\"" }, "queryContext" : [ { "objectType" : "", @@ -9657,9 +9657,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(10,0)) < CAST(1 AS BOOLEAN))\"", "left" : "\"DECIMAL(10,0)\"", - "right" : "\"BOOLEAN\"" + "right" : "\"BOOLEAN\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(10,0)) < CAST(1 AS BOOLEAN))\"" }, "queryContext" : [ { "objectType" : "", @@ -9681,9 +9681,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(20,0)) < CAST(1 AS BOOLEAN))\"", "left" : "\"DECIMAL(20,0)\"", - "right" : "\"BOOLEAN\"" + "right" : "\"BOOLEAN\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(20,0)) < CAST(1 AS BOOLEAN))\"" }, "queryContext" : [ { "objectType" : "", @@ -9705,9 +9705,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(3,0)) < CAST(2017-12-11 09:30:00.0 AS TIMESTAMP))\"", "left" : "\"DECIMAL(3,0)\"", - "right" : "\"TIMESTAMP\"" + "right" : "\"TIMESTAMP\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(3,0)) < CAST(2017-12-11 09:30:00.0 AS TIMESTAMP))\"" }, "queryContext" : [ { "objectType" : "", @@ -9729,9 +9729,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(5,0)) < CAST(2017-12-11 09:30:00.0 AS TIMESTAMP))\"", "left" : "\"DECIMAL(5,0)\"", - "right" : "\"TIMESTAMP\"" + "right" : "\"TIMESTAMP\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(5,0)) < CAST(2017-12-11 09:30:00.0 AS TIMESTAMP))\"" }, "queryContext" : [ { "objectType" : "", @@ -9753,9 +9753,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(10,0)) < CAST(2017-12-11 09:30:00.0 AS TIMESTAMP))\"", "left" : "\"DECIMAL(10,0)\"", - "right" : "\"TIMESTAMP\"" + "right" : "\"TIMESTAMP\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(10,0)) < CAST(2017-12-11 09:30:00.0 AS TIMESTAMP))\"" }, "queryContext" : [ { "objectType" : "", @@ -9777,9 +9777,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(20,0)) < CAST(2017-12-11 09:30:00.0 AS TIMESTAMP))\"", "left" : "\"DECIMAL(20,0)\"", - "right" : "\"TIMESTAMP\"" + "right" : "\"TIMESTAMP\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(20,0)) < CAST(2017-12-11 09:30:00.0 AS TIMESTAMP))\"" }, "queryContext" : [ { "objectType" : "", @@ -9801,9 +9801,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(3,0)) < CAST(2017-12-11 09:30:00 AS DATE))\"", "left" : "\"DECIMAL(3,0)\"", - "right" : "\"DATE\"" + "right" : "\"DATE\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(3,0)) < CAST(2017-12-11 09:30:00 AS DATE))\"" }, "queryContext" : [ { "objectType" : "", @@ -9825,9 +9825,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(5,0)) < CAST(2017-12-11 09:30:00 AS DATE))\"", "left" : "\"DECIMAL(5,0)\"", - "right" : "\"DATE\"" + "right" : "\"DATE\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(5,0)) < CAST(2017-12-11 09:30:00 AS DATE))\"" }, "queryContext" : [ { "objectType" : "", @@ -9849,9 +9849,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(10,0)) < CAST(2017-12-11 09:30:00 AS DATE))\"", "left" : "\"DECIMAL(10,0)\"", - "right" : "\"DATE\"" + "right" : "\"DATE\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(10,0)) < CAST(2017-12-11 09:30:00 AS DATE))\"" }, "queryContext" : [ { "objectType" : "", @@ -9873,9 +9873,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(20,0)) < CAST(2017-12-11 09:30:00 AS DATE))\"", "left" : "\"DECIMAL(20,0)\"", - "right" : "\"DATE\"" + "right" : "\"DATE\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(20,0)) < CAST(2017-12-11 09:30:00 AS DATE))\"" }, "queryContext" : [ { "objectType" : "", @@ -10121,9 +10121,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS BINARY) <= CAST(1 AS DECIMAL(3,0)))\"", "left" : "\"BINARY\"", - "right" : "\"DECIMAL(3,0)\"" + "right" : "\"DECIMAL(3,0)\"", + "sqlExpr" : "\"(CAST(1 AS BINARY) <= CAST(1 AS DECIMAL(3,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -10145,9 +10145,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS BINARY) <= CAST(1 AS DECIMAL(5,0)))\"", "left" : "\"BINARY\"", - "right" : "\"DECIMAL(5,0)\"" + "right" : "\"DECIMAL(5,0)\"", + "sqlExpr" : "\"(CAST(1 AS BINARY) <= CAST(1 AS DECIMAL(5,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -10169,9 +10169,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS BINARY) <= CAST(1 AS DECIMAL(10,0)))\"", "left" : "\"BINARY\"", - "right" : "\"DECIMAL(10,0)\"" + "right" : "\"DECIMAL(10,0)\"", + "sqlExpr" : "\"(CAST(1 AS BINARY) <= CAST(1 AS DECIMAL(10,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -10193,9 +10193,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS BINARY) <= CAST(1 AS DECIMAL(20,0)))\"", "left" : "\"BINARY\"", - "right" : "\"DECIMAL(20,0)\"" + "right" : "\"DECIMAL(20,0)\"", + "sqlExpr" : "\"(CAST(1 AS BINARY) <= CAST(1 AS DECIMAL(20,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -10217,9 +10217,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(2017-12-11 09:30:00.0 AS TIMESTAMP) <= CAST(1 AS DECIMAL(3,0)))\"", "left" : "\"TIMESTAMP\"", - "right" : "\"DECIMAL(3,0)\"" + "right" : "\"DECIMAL(3,0)\"", + "sqlExpr" : "\"(CAST(2017-12-11 09:30:00.0 AS TIMESTAMP) <= CAST(1 AS DECIMAL(3,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -10241,9 +10241,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(2017-12-11 09:30:00.0 AS TIMESTAMP) <= CAST(1 AS DECIMAL(5,0)))\"", "left" : "\"TIMESTAMP\"", - "right" : "\"DECIMAL(5,0)\"" + "right" : "\"DECIMAL(5,0)\"", + "sqlExpr" : "\"(CAST(2017-12-11 09:30:00.0 AS TIMESTAMP) <= CAST(1 AS DECIMAL(5,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -10265,9 +10265,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(2017-12-11 09:30:00.0 AS TIMESTAMP) <= CAST(1 AS DECIMAL(10,0)))\"", "left" : "\"TIMESTAMP\"", - "right" : "\"DECIMAL(10,0)\"" + "right" : "\"DECIMAL(10,0)\"", + "sqlExpr" : "\"(CAST(2017-12-11 09:30:00.0 AS TIMESTAMP) <= CAST(1 AS DECIMAL(10,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -10289,9 +10289,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(2017-12-11 09:30:00.0 AS TIMESTAMP) <= CAST(1 AS DECIMAL(20,0)))\"", "left" : "\"TIMESTAMP\"", - "right" : "\"DECIMAL(20,0)\"" + "right" : "\"DECIMAL(20,0)\"", + "sqlExpr" : "\"(CAST(2017-12-11 09:30:00.0 AS TIMESTAMP) <= CAST(1 AS DECIMAL(20,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -10313,9 +10313,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(2017-12-11 09:30:00 AS DATE) <= CAST(1 AS DECIMAL(3,0)))\"", "left" : "\"DATE\"", - "right" : "\"DECIMAL(3,0)\"" + "right" : "\"DECIMAL(3,0)\"", + "sqlExpr" : "\"(CAST(2017-12-11 09:30:00 AS DATE) <= CAST(1 AS DECIMAL(3,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -10337,9 +10337,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(2017-12-11 09:30:00 AS DATE) <= CAST(1 AS DECIMAL(5,0)))\"", "left" : "\"DATE\"", - "right" : "\"DECIMAL(5,0)\"" + "right" : "\"DECIMAL(5,0)\"", + "sqlExpr" : "\"(CAST(2017-12-11 09:30:00 AS DATE) <= CAST(1 AS DECIMAL(5,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -10361,9 +10361,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(2017-12-11 09:30:00 AS DATE) <= CAST(1 AS DECIMAL(10,0)))\"", "left" : "\"DATE\"", - "right" : "\"DECIMAL(10,0)\"" + "right" : "\"DECIMAL(10,0)\"", + "sqlExpr" : "\"(CAST(2017-12-11 09:30:00 AS DATE) <= CAST(1 AS DECIMAL(10,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -10385,9 +10385,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(2017-12-11 09:30:00 AS DATE) <= CAST(1 AS DECIMAL(20,0)))\"", "left" : "\"DATE\"", - "right" : "\"DECIMAL(20,0)\"" + "right" : "\"DECIMAL(20,0)\"", + "sqlExpr" : "\"(CAST(2017-12-11 09:30:00 AS DATE) <= CAST(1 AS DECIMAL(20,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -10665,9 +10665,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(3,0)) <= CAST(1 AS BINARY))\"", "left" : "\"DECIMAL(3,0)\"", - "right" : "\"BINARY\"" + "right" : "\"BINARY\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(3,0)) <= CAST(1 AS BINARY))\"" }, "queryContext" : [ { "objectType" : "", @@ -10689,9 +10689,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(5,0)) <= CAST(1 AS BINARY))\"", "left" : "\"DECIMAL(5,0)\"", - "right" : "\"BINARY\"" + "right" : "\"BINARY\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(5,0)) <= CAST(1 AS BINARY))\"" }, "queryContext" : [ { "objectType" : "", @@ -10713,9 +10713,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(10,0)) <= CAST(1 AS BINARY))\"", "left" : "\"DECIMAL(10,0)\"", - "right" : "\"BINARY\"" + "right" : "\"BINARY\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(10,0)) <= CAST(1 AS BINARY))\"" }, "queryContext" : [ { "objectType" : "", @@ -10737,9 +10737,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(20,0)) <= CAST(1 AS BINARY))\"", "left" : "\"DECIMAL(20,0)\"", - "right" : "\"BINARY\"" + "right" : "\"BINARY\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(20,0)) <= CAST(1 AS BINARY))\"" }, "queryContext" : [ { "objectType" : "", @@ -10761,9 +10761,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(3,0)) <= CAST(1 AS BOOLEAN))\"", "left" : "\"DECIMAL(3,0)\"", - "right" : "\"BOOLEAN\"" + "right" : "\"BOOLEAN\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(3,0)) <= CAST(1 AS BOOLEAN))\"" }, "queryContext" : [ { "objectType" : "", @@ -10785,9 +10785,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(5,0)) <= CAST(1 AS BOOLEAN))\"", "left" : "\"DECIMAL(5,0)\"", - "right" : "\"BOOLEAN\"" + "right" : "\"BOOLEAN\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(5,0)) <= CAST(1 AS BOOLEAN))\"" }, "queryContext" : [ { "objectType" : "", @@ -10809,9 +10809,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(10,0)) <= CAST(1 AS BOOLEAN))\"", "left" : "\"DECIMAL(10,0)\"", - "right" : "\"BOOLEAN\"" + "right" : "\"BOOLEAN\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(10,0)) <= CAST(1 AS BOOLEAN))\"" }, "queryContext" : [ { "objectType" : "", @@ -10833,9 +10833,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(20,0)) <= CAST(1 AS BOOLEAN))\"", "left" : "\"DECIMAL(20,0)\"", - "right" : "\"BOOLEAN\"" + "right" : "\"BOOLEAN\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(20,0)) <= CAST(1 AS BOOLEAN))\"" }, "queryContext" : [ { "objectType" : "", @@ -10857,9 +10857,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(3,0)) <= CAST(2017-12-11 09:30:00.0 AS TIMESTAMP))\"", "left" : "\"DECIMAL(3,0)\"", - "right" : "\"TIMESTAMP\"" + "right" : "\"TIMESTAMP\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(3,0)) <= CAST(2017-12-11 09:30:00.0 AS TIMESTAMP))\"" }, "queryContext" : [ { "objectType" : "", @@ -10881,9 +10881,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(5,0)) <= CAST(2017-12-11 09:30:00.0 AS TIMESTAMP))\"", "left" : "\"DECIMAL(5,0)\"", - "right" : "\"TIMESTAMP\"" + "right" : "\"TIMESTAMP\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(5,0)) <= CAST(2017-12-11 09:30:00.0 AS TIMESTAMP))\"" }, "queryContext" : [ { "objectType" : "", @@ -10905,9 +10905,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(10,0)) <= CAST(2017-12-11 09:30:00.0 AS TIMESTAMP))\"", "left" : "\"DECIMAL(10,0)\"", - "right" : "\"TIMESTAMP\"" + "right" : "\"TIMESTAMP\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(10,0)) <= CAST(2017-12-11 09:30:00.0 AS TIMESTAMP))\"" }, "queryContext" : [ { "objectType" : "", @@ -10929,9 +10929,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(20,0)) <= CAST(2017-12-11 09:30:00.0 AS TIMESTAMP))\"", "left" : "\"DECIMAL(20,0)\"", - "right" : "\"TIMESTAMP\"" + "right" : "\"TIMESTAMP\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(20,0)) <= CAST(2017-12-11 09:30:00.0 AS TIMESTAMP))\"" }, "queryContext" : [ { "objectType" : "", @@ -10953,9 +10953,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(3,0)) <= CAST(2017-12-11 09:30:00 AS DATE))\"", "left" : "\"DECIMAL(3,0)\"", - "right" : "\"DATE\"" + "right" : "\"DATE\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(3,0)) <= CAST(2017-12-11 09:30:00 AS DATE))\"" }, "queryContext" : [ { "objectType" : "", @@ -10977,9 +10977,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(5,0)) <= CAST(2017-12-11 09:30:00 AS DATE))\"", "left" : "\"DECIMAL(5,0)\"", - "right" : "\"DATE\"" + "right" : "\"DATE\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(5,0)) <= CAST(2017-12-11 09:30:00 AS DATE))\"" }, "queryContext" : [ { "objectType" : "", @@ -11001,9 +11001,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(10,0)) <= CAST(2017-12-11 09:30:00 AS DATE))\"", "left" : "\"DECIMAL(10,0)\"", - "right" : "\"DATE\"" + "right" : "\"DATE\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(10,0)) <= CAST(2017-12-11 09:30:00 AS DATE))\"" }, "queryContext" : [ { "objectType" : "", @@ -11025,9 +11025,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(20,0)) <= CAST(2017-12-11 09:30:00 AS DATE))\"", "left" : "\"DECIMAL(20,0)\"", - "right" : "\"DATE\"" + "right" : "\"DATE\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(20,0)) <= CAST(2017-12-11 09:30:00 AS DATE))\"" }, "queryContext" : [ { "objectType" : "", @@ -11273,9 +11273,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS BINARY) > CAST(1 AS DECIMAL(3,0)))\"", "left" : "\"BINARY\"", - "right" : "\"DECIMAL(3,0)\"" + "right" : "\"DECIMAL(3,0)\"", + "sqlExpr" : "\"(CAST(1 AS BINARY) > CAST(1 AS DECIMAL(3,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -11297,9 +11297,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS BINARY) > CAST(1 AS DECIMAL(5,0)))\"", "left" : "\"BINARY\"", - "right" : "\"DECIMAL(5,0)\"" + "right" : "\"DECIMAL(5,0)\"", + "sqlExpr" : "\"(CAST(1 AS BINARY) > CAST(1 AS DECIMAL(5,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -11321,9 +11321,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS BINARY) > CAST(1 AS DECIMAL(10,0)))\"", "left" : "\"BINARY\"", - "right" : "\"DECIMAL(10,0)\"" + "right" : "\"DECIMAL(10,0)\"", + "sqlExpr" : "\"(CAST(1 AS BINARY) > CAST(1 AS DECIMAL(10,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -11345,9 +11345,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS BINARY) > CAST(1 AS DECIMAL(20,0)))\"", "left" : "\"BINARY\"", - "right" : "\"DECIMAL(20,0)\"" + "right" : "\"DECIMAL(20,0)\"", + "sqlExpr" : "\"(CAST(1 AS BINARY) > CAST(1 AS DECIMAL(20,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -11369,9 +11369,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(2017-12-11 09:30:00.0 AS TIMESTAMP) > CAST(1 AS DECIMAL(3,0)))\"", "left" : "\"TIMESTAMP\"", - "right" : "\"DECIMAL(3,0)\"" + "right" : "\"DECIMAL(3,0)\"", + "sqlExpr" : "\"(CAST(2017-12-11 09:30:00.0 AS TIMESTAMP) > CAST(1 AS DECIMAL(3,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -11393,9 +11393,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(2017-12-11 09:30:00.0 AS TIMESTAMP) > CAST(1 AS DECIMAL(5,0)))\"", "left" : "\"TIMESTAMP\"", - "right" : "\"DECIMAL(5,0)\"" + "right" : "\"DECIMAL(5,0)\"", + "sqlExpr" : "\"(CAST(2017-12-11 09:30:00.0 AS TIMESTAMP) > CAST(1 AS DECIMAL(5,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -11417,9 +11417,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(2017-12-11 09:30:00.0 AS TIMESTAMP) > CAST(1 AS DECIMAL(10,0)))\"", "left" : "\"TIMESTAMP\"", - "right" : "\"DECIMAL(10,0)\"" + "right" : "\"DECIMAL(10,0)\"", + "sqlExpr" : "\"(CAST(2017-12-11 09:30:00.0 AS TIMESTAMP) > CAST(1 AS DECIMAL(10,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -11441,9 +11441,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(2017-12-11 09:30:00.0 AS TIMESTAMP) > CAST(1 AS DECIMAL(20,0)))\"", "left" : "\"TIMESTAMP\"", - "right" : "\"DECIMAL(20,0)\"" + "right" : "\"DECIMAL(20,0)\"", + "sqlExpr" : "\"(CAST(2017-12-11 09:30:00.0 AS TIMESTAMP) > CAST(1 AS DECIMAL(20,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -11465,9 +11465,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(2017-12-11 09:30:00 AS DATE) > CAST(1 AS DECIMAL(3,0)))\"", "left" : "\"DATE\"", - "right" : "\"DECIMAL(3,0)\"" + "right" : "\"DECIMAL(3,0)\"", + "sqlExpr" : "\"(CAST(2017-12-11 09:30:00 AS DATE) > CAST(1 AS DECIMAL(3,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -11489,9 +11489,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(2017-12-11 09:30:00 AS DATE) > CAST(1 AS DECIMAL(5,0)))\"", "left" : "\"DATE\"", - "right" : "\"DECIMAL(5,0)\"" + "right" : "\"DECIMAL(5,0)\"", + "sqlExpr" : "\"(CAST(2017-12-11 09:30:00 AS DATE) > CAST(1 AS DECIMAL(5,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -11513,9 +11513,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(2017-12-11 09:30:00 AS DATE) > CAST(1 AS DECIMAL(10,0)))\"", "left" : "\"DATE\"", - "right" : "\"DECIMAL(10,0)\"" + "right" : "\"DECIMAL(10,0)\"", + "sqlExpr" : "\"(CAST(2017-12-11 09:30:00 AS DATE) > CAST(1 AS DECIMAL(10,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -11537,9 +11537,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(2017-12-11 09:30:00 AS DATE) > CAST(1 AS DECIMAL(20,0)))\"", "left" : "\"DATE\"", - "right" : "\"DECIMAL(20,0)\"" + "right" : "\"DECIMAL(20,0)\"", + "sqlExpr" : "\"(CAST(2017-12-11 09:30:00 AS DATE) > CAST(1 AS DECIMAL(20,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -11817,9 +11817,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(3,0)) > CAST(1 AS BINARY))\"", "left" : "\"DECIMAL(3,0)\"", - "right" : "\"BINARY\"" + "right" : "\"BINARY\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(3,0)) > CAST(1 AS BINARY))\"" }, "queryContext" : [ { "objectType" : "", @@ -11841,9 +11841,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(5,0)) > CAST(1 AS BINARY))\"", "left" : "\"DECIMAL(5,0)\"", - "right" : "\"BINARY\"" + "right" : "\"BINARY\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(5,0)) > CAST(1 AS BINARY))\"" }, "queryContext" : [ { "objectType" : "", @@ -11865,9 +11865,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(10,0)) > CAST(1 AS BINARY))\"", "left" : "\"DECIMAL(10,0)\"", - "right" : "\"BINARY\"" + "right" : "\"BINARY\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(10,0)) > CAST(1 AS BINARY))\"" }, "queryContext" : [ { "objectType" : "", @@ -11889,9 +11889,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(20,0)) > CAST(1 AS BINARY))\"", "left" : "\"DECIMAL(20,0)\"", - "right" : "\"BINARY\"" + "right" : "\"BINARY\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(20,0)) > CAST(1 AS BINARY))\"" }, "queryContext" : [ { "objectType" : "", @@ -11913,9 +11913,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(3,0)) > CAST(1 AS BOOLEAN))\"", "left" : "\"DECIMAL(3,0)\"", - "right" : "\"BOOLEAN\"" + "right" : "\"BOOLEAN\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(3,0)) > CAST(1 AS BOOLEAN))\"" }, "queryContext" : [ { "objectType" : "", @@ -11937,9 +11937,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(5,0)) > CAST(1 AS BOOLEAN))\"", "left" : "\"DECIMAL(5,0)\"", - "right" : "\"BOOLEAN\"" + "right" : "\"BOOLEAN\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(5,0)) > CAST(1 AS BOOLEAN))\"" }, "queryContext" : [ { "objectType" : "", @@ -11961,9 +11961,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(10,0)) > CAST(1 AS BOOLEAN))\"", "left" : "\"DECIMAL(10,0)\"", - "right" : "\"BOOLEAN\"" + "right" : "\"BOOLEAN\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(10,0)) > CAST(1 AS BOOLEAN))\"" }, "queryContext" : [ { "objectType" : "", @@ -11985,9 +11985,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(20,0)) > CAST(1 AS BOOLEAN))\"", "left" : "\"DECIMAL(20,0)\"", - "right" : "\"BOOLEAN\"" + "right" : "\"BOOLEAN\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(20,0)) > CAST(1 AS BOOLEAN))\"" }, "queryContext" : [ { "objectType" : "", @@ -12009,9 +12009,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(3,0)) > CAST(2017-12-11 09:30:00.0 AS TIMESTAMP))\"", "left" : "\"DECIMAL(3,0)\"", - "right" : "\"TIMESTAMP\"" + "right" : "\"TIMESTAMP\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(3,0)) > CAST(2017-12-11 09:30:00.0 AS TIMESTAMP))\"" }, "queryContext" : [ { "objectType" : "", @@ -12033,9 +12033,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(5,0)) > CAST(2017-12-11 09:30:00.0 AS TIMESTAMP))\"", "left" : "\"DECIMAL(5,0)\"", - "right" : "\"TIMESTAMP\"" + "right" : "\"TIMESTAMP\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(5,0)) > CAST(2017-12-11 09:30:00.0 AS TIMESTAMP))\"" }, "queryContext" : [ { "objectType" : "", @@ -12057,9 +12057,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(10,0)) > CAST(2017-12-11 09:30:00.0 AS TIMESTAMP))\"", "left" : "\"DECIMAL(10,0)\"", - "right" : "\"TIMESTAMP\"" + "right" : "\"TIMESTAMP\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(10,0)) > CAST(2017-12-11 09:30:00.0 AS TIMESTAMP))\"" }, "queryContext" : [ { "objectType" : "", @@ -12081,9 +12081,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(20,0)) > CAST(2017-12-11 09:30:00.0 AS TIMESTAMP))\"", "left" : "\"DECIMAL(20,0)\"", - "right" : "\"TIMESTAMP\"" + "right" : "\"TIMESTAMP\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(20,0)) > CAST(2017-12-11 09:30:00.0 AS TIMESTAMP))\"" }, "queryContext" : [ { "objectType" : "", @@ -12105,9 +12105,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(3,0)) > CAST(2017-12-11 09:30:00 AS DATE))\"", "left" : "\"DECIMAL(3,0)\"", - "right" : "\"DATE\"" + "right" : "\"DATE\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(3,0)) > CAST(2017-12-11 09:30:00 AS DATE))\"" }, "queryContext" : [ { "objectType" : "", @@ -12129,9 +12129,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(5,0)) > CAST(2017-12-11 09:30:00 AS DATE))\"", "left" : "\"DECIMAL(5,0)\"", - "right" : "\"DATE\"" + "right" : "\"DATE\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(5,0)) > CAST(2017-12-11 09:30:00 AS DATE))\"" }, "queryContext" : [ { "objectType" : "", @@ -12153,9 +12153,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(10,0)) > CAST(2017-12-11 09:30:00 AS DATE))\"", "left" : "\"DECIMAL(10,0)\"", - "right" : "\"DATE\"" + "right" : "\"DATE\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(10,0)) > CAST(2017-12-11 09:30:00 AS DATE))\"" }, "queryContext" : [ { "objectType" : "", @@ -12177,9 +12177,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(20,0)) > CAST(2017-12-11 09:30:00 AS DATE))\"", "left" : "\"DECIMAL(20,0)\"", - "right" : "\"DATE\"" + "right" : "\"DATE\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(20,0)) > CAST(2017-12-11 09:30:00 AS DATE))\"" }, "queryContext" : [ { "objectType" : "", @@ -12425,9 +12425,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS BINARY) >= CAST(1 AS DECIMAL(3,0)))\"", "left" : "\"BINARY\"", - "right" : "\"DECIMAL(3,0)\"" + "right" : "\"DECIMAL(3,0)\"", + "sqlExpr" : "\"(CAST(1 AS BINARY) >= CAST(1 AS DECIMAL(3,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -12449,9 +12449,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS BINARY) >= CAST(1 AS DECIMAL(5,0)))\"", "left" : "\"BINARY\"", - "right" : "\"DECIMAL(5,0)\"" + "right" : "\"DECIMAL(5,0)\"", + "sqlExpr" : "\"(CAST(1 AS BINARY) >= CAST(1 AS DECIMAL(5,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -12473,9 +12473,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS BINARY) >= CAST(1 AS DECIMAL(10,0)))\"", "left" : "\"BINARY\"", - "right" : "\"DECIMAL(10,0)\"" + "right" : "\"DECIMAL(10,0)\"", + "sqlExpr" : "\"(CAST(1 AS BINARY) >= CAST(1 AS DECIMAL(10,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -12497,9 +12497,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS BINARY) >= CAST(1 AS DECIMAL(20,0)))\"", "left" : "\"BINARY\"", - "right" : "\"DECIMAL(20,0)\"" + "right" : "\"DECIMAL(20,0)\"", + "sqlExpr" : "\"(CAST(1 AS BINARY) >= CAST(1 AS DECIMAL(20,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -12521,9 +12521,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(2017-12-11 09:30:00.0 AS TIMESTAMP) >= CAST(1 AS DECIMAL(3,0)))\"", "left" : "\"TIMESTAMP\"", - "right" : "\"DECIMAL(3,0)\"" + "right" : "\"DECIMAL(3,0)\"", + "sqlExpr" : "\"(CAST(2017-12-11 09:30:00.0 AS TIMESTAMP) >= CAST(1 AS DECIMAL(3,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -12545,9 +12545,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(2017-12-11 09:30:00.0 AS TIMESTAMP) >= CAST(1 AS DECIMAL(5,0)))\"", "left" : "\"TIMESTAMP\"", - "right" : "\"DECIMAL(5,0)\"" + "right" : "\"DECIMAL(5,0)\"", + "sqlExpr" : "\"(CAST(2017-12-11 09:30:00.0 AS TIMESTAMP) >= CAST(1 AS DECIMAL(5,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -12569,9 +12569,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(2017-12-11 09:30:00.0 AS TIMESTAMP) >= CAST(1 AS DECIMAL(10,0)))\"", "left" : "\"TIMESTAMP\"", - "right" : "\"DECIMAL(10,0)\"" + "right" : "\"DECIMAL(10,0)\"", + "sqlExpr" : "\"(CAST(2017-12-11 09:30:00.0 AS TIMESTAMP) >= CAST(1 AS DECIMAL(10,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -12593,9 +12593,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(2017-12-11 09:30:00.0 AS TIMESTAMP) >= CAST(1 AS DECIMAL(20,0)))\"", "left" : "\"TIMESTAMP\"", - "right" : "\"DECIMAL(20,0)\"" + "right" : "\"DECIMAL(20,0)\"", + "sqlExpr" : "\"(CAST(2017-12-11 09:30:00.0 AS TIMESTAMP) >= CAST(1 AS DECIMAL(20,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -12617,9 +12617,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(2017-12-11 09:30:00 AS DATE) >= CAST(1 AS DECIMAL(3,0)))\"", "left" : "\"DATE\"", - "right" : "\"DECIMAL(3,0)\"" + "right" : "\"DECIMAL(3,0)\"", + "sqlExpr" : "\"(CAST(2017-12-11 09:30:00 AS DATE) >= CAST(1 AS DECIMAL(3,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -12641,9 +12641,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(2017-12-11 09:30:00 AS DATE) >= CAST(1 AS DECIMAL(5,0)))\"", "left" : "\"DATE\"", - "right" : "\"DECIMAL(5,0)\"" + "right" : "\"DECIMAL(5,0)\"", + "sqlExpr" : "\"(CAST(2017-12-11 09:30:00 AS DATE) >= CAST(1 AS DECIMAL(5,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -12665,9 +12665,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(2017-12-11 09:30:00 AS DATE) >= CAST(1 AS DECIMAL(10,0)))\"", "left" : "\"DATE\"", - "right" : "\"DECIMAL(10,0)\"" + "right" : "\"DECIMAL(10,0)\"", + "sqlExpr" : "\"(CAST(2017-12-11 09:30:00 AS DATE) >= CAST(1 AS DECIMAL(10,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -12689,9 +12689,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(2017-12-11 09:30:00 AS DATE) >= CAST(1 AS DECIMAL(20,0)))\"", "left" : "\"DATE\"", - "right" : "\"DECIMAL(20,0)\"" + "right" : "\"DECIMAL(20,0)\"", + "sqlExpr" : "\"(CAST(2017-12-11 09:30:00 AS DATE) >= CAST(1 AS DECIMAL(20,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -12969,9 +12969,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(3,0)) >= CAST(1 AS BINARY))\"", "left" : "\"DECIMAL(3,0)\"", - "right" : "\"BINARY\"" + "right" : "\"BINARY\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(3,0)) >= CAST(1 AS BINARY))\"" }, "queryContext" : [ { "objectType" : "", @@ -12993,9 +12993,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(5,0)) >= CAST(1 AS BINARY))\"", "left" : "\"DECIMAL(5,0)\"", - "right" : "\"BINARY\"" + "right" : "\"BINARY\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(5,0)) >= CAST(1 AS BINARY))\"" }, "queryContext" : [ { "objectType" : "", @@ -13017,9 +13017,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(10,0)) >= CAST(1 AS BINARY))\"", "left" : "\"DECIMAL(10,0)\"", - "right" : "\"BINARY\"" + "right" : "\"BINARY\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(10,0)) >= CAST(1 AS BINARY))\"" }, "queryContext" : [ { "objectType" : "", @@ -13041,9 +13041,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(20,0)) >= CAST(1 AS BINARY))\"", "left" : "\"DECIMAL(20,0)\"", - "right" : "\"BINARY\"" + "right" : "\"BINARY\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(20,0)) >= CAST(1 AS BINARY))\"" }, "queryContext" : [ { "objectType" : "", @@ -13065,9 +13065,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(3,0)) >= CAST(1 AS BOOLEAN))\"", "left" : "\"DECIMAL(3,0)\"", - "right" : "\"BOOLEAN\"" + "right" : "\"BOOLEAN\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(3,0)) >= CAST(1 AS BOOLEAN))\"" }, "queryContext" : [ { "objectType" : "", @@ -13089,9 +13089,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(5,0)) >= CAST(1 AS BOOLEAN))\"", "left" : "\"DECIMAL(5,0)\"", - "right" : "\"BOOLEAN\"" + "right" : "\"BOOLEAN\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(5,0)) >= CAST(1 AS BOOLEAN))\"" }, "queryContext" : [ { "objectType" : "", @@ -13113,9 +13113,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(10,0)) >= CAST(1 AS BOOLEAN))\"", "left" : "\"DECIMAL(10,0)\"", - "right" : "\"BOOLEAN\"" + "right" : "\"BOOLEAN\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(10,0)) >= CAST(1 AS BOOLEAN))\"" }, "queryContext" : [ { "objectType" : "", @@ -13137,9 +13137,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(20,0)) >= CAST(1 AS BOOLEAN))\"", "left" : "\"DECIMAL(20,0)\"", - "right" : "\"BOOLEAN\"" + "right" : "\"BOOLEAN\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(20,0)) >= CAST(1 AS BOOLEAN))\"" }, "queryContext" : [ { "objectType" : "", @@ -13161,9 +13161,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(3,0)) >= CAST(2017-12-11 09:30:00.0 AS TIMESTAMP))\"", "left" : "\"DECIMAL(3,0)\"", - "right" : "\"TIMESTAMP\"" + "right" : "\"TIMESTAMP\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(3,0)) >= CAST(2017-12-11 09:30:00.0 AS TIMESTAMP))\"" }, "queryContext" : [ { "objectType" : "", @@ -13185,9 +13185,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(5,0)) >= CAST(2017-12-11 09:30:00.0 AS TIMESTAMP))\"", "left" : "\"DECIMAL(5,0)\"", - "right" : "\"TIMESTAMP\"" + "right" : "\"TIMESTAMP\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(5,0)) >= CAST(2017-12-11 09:30:00.0 AS TIMESTAMP))\"" }, "queryContext" : [ { "objectType" : "", @@ -13209,9 +13209,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(10,0)) >= CAST(2017-12-11 09:30:00.0 AS TIMESTAMP))\"", "left" : "\"DECIMAL(10,0)\"", - "right" : "\"TIMESTAMP\"" + "right" : "\"TIMESTAMP\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(10,0)) >= CAST(2017-12-11 09:30:00.0 AS TIMESTAMP))\"" }, "queryContext" : [ { "objectType" : "", @@ -13233,9 +13233,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(20,0)) >= CAST(2017-12-11 09:30:00.0 AS TIMESTAMP))\"", "left" : "\"DECIMAL(20,0)\"", - "right" : "\"TIMESTAMP\"" + "right" : "\"TIMESTAMP\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(20,0)) >= CAST(2017-12-11 09:30:00.0 AS TIMESTAMP))\"" }, "queryContext" : [ { "objectType" : "", @@ -13257,9 +13257,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(3,0)) >= CAST(2017-12-11 09:30:00 AS DATE))\"", "left" : "\"DECIMAL(3,0)\"", - "right" : "\"DATE\"" + "right" : "\"DATE\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(3,0)) >= CAST(2017-12-11 09:30:00 AS DATE))\"" }, "queryContext" : [ { "objectType" : "", @@ -13281,9 +13281,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(5,0)) >= CAST(2017-12-11 09:30:00 AS DATE))\"", "left" : "\"DECIMAL(5,0)\"", - "right" : "\"DATE\"" + "right" : "\"DATE\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(5,0)) >= CAST(2017-12-11 09:30:00 AS DATE))\"" }, "queryContext" : [ { "objectType" : "", @@ -13305,9 +13305,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(10,0)) >= CAST(2017-12-11 09:30:00 AS DATE))\"", "left" : "\"DECIMAL(10,0)\"", - "right" : "\"DATE\"" + "right" : "\"DATE\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(10,0)) >= CAST(2017-12-11 09:30:00 AS DATE))\"" }, "queryContext" : [ { "objectType" : "", @@ -13329,9 +13329,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(20,0)) >= CAST(2017-12-11 09:30:00 AS DATE))\"", "left" : "\"DECIMAL(20,0)\"", - "right" : "\"DATE\"" + "right" : "\"DATE\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(20,0)) >= CAST(2017-12-11 09:30:00 AS DATE))\"" }, "queryContext" : [ { "objectType" : "", @@ -13577,9 +13577,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS BINARY) = CAST(1 AS DECIMAL(3,0)))\"", "left" : "\"BINARY\"", - "right" : "\"DECIMAL(3,0)\"" + "right" : "\"DECIMAL(3,0)\"", + "sqlExpr" : "\"(CAST(1 AS BINARY) = CAST(1 AS DECIMAL(3,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -13601,9 +13601,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS BINARY) = CAST(1 AS DECIMAL(5,0)))\"", "left" : "\"BINARY\"", - "right" : "\"DECIMAL(5,0)\"" + "right" : "\"DECIMAL(5,0)\"", + "sqlExpr" : "\"(CAST(1 AS BINARY) = CAST(1 AS DECIMAL(5,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -13625,9 +13625,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS BINARY) = CAST(1 AS DECIMAL(10,0)))\"", "left" : "\"BINARY\"", - "right" : "\"DECIMAL(10,0)\"" + "right" : "\"DECIMAL(10,0)\"", + "sqlExpr" : "\"(CAST(1 AS BINARY) = CAST(1 AS DECIMAL(10,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -13649,9 +13649,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS BINARY) = CAST(1 AS DECIMAL(20,0)))\"", "left" : "\"BINARY\"", - "right" : "\"DECIMAL(20,0)\"" + "right" : "\"DECIMAL(20,0)\"", + "sqlExpr" : "\"(CAST(1 AS BINARY) = CAST(1 AS DECIMAL(20,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -13673,9 +13673,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(2017-12-11 09:30:00.0 AS TIMESTAMP) = CAST(1 AS DECIMAL(3,0)))\"", "left" : "\"TIMESTAMP\"", - "right" : "\"DECIMAL(3,0)\"" + "right" : "\"DECIMAL(3,0)\"", + "sqlExpr" : "\"(CAST(2017-12-11 09:30:00.0 AS TIMESTAMP) = CAST(1 AS DECIMAL(3,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -13697,9 +13697,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(2017-12-11 09:30:00.0 AS TIMESTAMP) = CAST(1 AS DECIMAL(5,0)))\"", "left" : "\"TIMESTAMP\"", - "right" : "\"DECIMAL(5,0)\"" + "right" : "\"DECIMAL(5,0)\"", + "sqlExpr" : "\"(CAST(2017-12-11 09:30:00.0 AS TIMESTAMP) = CAST(1 AS DECIMAL(5,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -13721,9 +13721,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(2017-12-11 09:30:00.0 AS TIMESTAMP) = CAST(1 AS DECIMAL(10,0)))\"", "left" : "\"TIMESTAMP\"", - "right" : "\"DECIMAL(10,0)\"" + "right" : "\"DECIMAL(10,0)\"", + "sqlExpr" : "\"(CAST(2017-12-11 09:30:00.0 AS TIMESTAMP) = CAST(1 AS DECIMAL(10,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -13745,9 +13745,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(2017-12-11 09:30:00.0 AS TIMESTAMP) = CAST(1 AS DECIMAL(20,0)))\"", "left" : "\"TIMESTAMP\"", - "right" : "\"DECIMAL(20,0)\"" + "right" : "\"DECIMAL(20,0)\"", + "sqlExpr" : "\"(CAST(2017-12-11 09:30:00.0 AS TIMESTAMP) = CAST(1 AS DECIMAL(20,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -13769,9 +13769,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(2017-12-11 09:30:00 AS DATE) = CAST(1 AS DECIMAL(3,0)))\"", "left" : "\"DATE\"", - "right" : "\"DECIMAL(3,0)\"" + "right" : "\"DECIMAL(3,0)\"", + "sqlExpr" : "\"(CAST(2017-12-11 09:30:00 AS DATE) = CAST(1 AS DECIMAL(3,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -13793,9 +13793,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(2017-12-11 09:30:00 AS DATE) = CAST(1 AS DECIMAL(5,0)))\"", "left" : "\"DATE\"", - "right" : "\"DECIMAL(5,0)\"" + "right" : "\"DECIMAL(5,0)\"", + "sqlExpr" : "\"(CAST(2017-12-11 09:30:00 AS DATE) = CAST(1 AS DECIMAL(5,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -13817,9 +13817,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(2017-12-11 09:30:00 AS DATE) = CAST(1 AS DECIMAL(10,0)))\"", "left" : "\"DATE\"", - "right" : "\"DECIMAL(10,0)\"" + "right" : "\"DECIMAL(10,0)\"", + "sqlExpr" : "\"(CAST(2017-12-11 09:30:00 AS DATE) = CAST(1 AS DECIMAL(10,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -13841,9 +13841,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(2017-12-11 09:30:00 AS DATE) = CAST(1 AS DECIMAL(20,0)))\"", "left" : "\"DATE\"", - "right" : "\"DECIMAL(20,0)\"" + "right" : "\"DECIMAL(20,0)\"", + "sqlExpr" : "\"(CAST(2017-12-11 09:30:00 AS DATE) = CAST(1 AS DECIMAL(20,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -14121,9 +14121,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(3,0)) = CAST(1 AS BINARY))\"", "left" : "\"DECIMAL(3,0)\"", - "right" : "\"BINARY\"" + "right" : "\"BINARY\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(3,0)) = CAST(1 AS BINARY))\"" }, "queryContext" : [ { "objectType" : "", @@ -14145,9 +14145,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(5,0)) = CAST(1 AS BINARY))\"", "left" : "\"DECIMAL(5,0)\"", - "right" : "\"BINARY\"" + "right" : "\"BINARY\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(5,0)) = CAST(1 AS BINARY))\"" }, "queryContext" : [ { "objectType" : "", @@ -14169,9 +14169,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(10,0)) = CAST(1 AS BINARY))\"", "left" : "\"DECIMAL(10,0)\"", - "right" : "\"BINARY\"" + "right" : "\"BINARY\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(10,0)) = CAST(1 AS BINARY))\"" }, "queryContext" : [ { "objectType" : "", @@ -14193,9 +14193,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(20,0)) = CAST(1 AS BINARY))\"", "left" : "\"DECIMAL(20,0)\"", - "right" : "\"BINARY\"" + "right" : "\"BINARY\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(20,0)) = CAST(1 AS BINARY))\"" }, "queryContext" : [ { "objectType" : "", @@ -14249,9 +14249,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(3,0)) = CAST(2017-12-11 09:30:00.0 AS TIMESTAMP))\"", "left" : "\"DECIMAL(3,0)\"", - "right" : "\"TIMESTAMP\"" + "right" : "\"TIMESTAMP\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(3,0)) = CAST(2017-12-11 09:30:00.0 AS TIMESTAMP))\"" }, "queryContext" : [ { "objectType" : "", @@ -14273,9 +14273,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(5,0)) = CAST(2017-12-11 09:30:00.0 AS TIMESTAMP))\"", "left" : "\"DECIMAL(5,0)\"", - "right" : "\"TIMESTAMP\"" + "right" : "\"TIMESTAMP\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(5,0)) = CAST(2017-12-11 09:30:00.0 AS TIMESTAMP))\"" }, "queryContext" : [ { "objectType" : "", @@ -14297,9 +14297,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(10,0)) = CAST(2017-12-11 09:30:00.0 AS TIMESTAMP))\"", "left" : "\"DECIMAL(10,0)\"", - "right" : "\"TIMESTAMP\"" + "right" : "\"TIMESTAMP\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(10,0)) = CAST(2017-12-11 09:30:00.0 AS TIMESTAMP))\"" }, "queryContext" : [ { "objectType" : "", @@ -14321,9 +14321,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(20,0)) = CAST(2017-12-11 09:30:00.0 AS TIMESTAMP))\"", "left" : "\"DECIMAL(20,0)\"", - "right" : "\"TIMESTAMP\"" + "right" : "\"TIMESTAMP\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(20,0)) = CAST(2017-12-11 09:30:00.0 AS TIMESTAMP))\"" }, "queryContext" : [ { "objectType" : "", @@ -14345,9 +14345,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(3,0)) = CAST(2017-12-11 09:30:00 AS DATE))\"", "left" : "\"DECIMAL(3,0)\"", - "right" : "\"DATE\"" + "right" : "\"DATE\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(3,0)) = CAST(2017-12-11 09:30:00 AS DATE))\"" }, "queryContext" : [ { "objectType" : "", @@ -14369,9 +14369,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(5,0)) = CAST(2017-12-11 09:30:00 AS DATE))\"", "left" : "\"DECIMAL(5,0)\"", - "right" : "\"DATE\"" + "right" : "\"DATE\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(5,0)) = CAST(2017-12-11 09:30:00 AS DATE))\"" }, "queryContext" : [ { "objectType" : "", @@ -14393,9 +14393,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(10,0)) = CAST(2017-12-11 09:30:00 AS DATE))\"", "left" : "\"DECIMAL(10,0)\"", - "right" : "\"DATE\"" + "right" : "\"DATE\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(10,0)) = CAST(2017-12-11 09:30:00 AS DATE))\"" }, "queryContext" : [ { "objectType" : "", @@ -14417,9 +14417,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(20,0)) = CAST(2017-12-11 09:30:00 AS DATE))\"", "left" : "\"DECIMAL(20,0)\"", - "right" : "\"DATE\"" + "right" : "\"DATE\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(20,0)) = CAST(2017-12-11 09:30:00 AS DATE))\"" }, "queryContext" : [ { "objectType" : "", diff --git a/sql/core/src/test/resources/sql-tests/results/typeCoercion/native/division.sql.out b/sql/core/src/test/resources/sql-tests/results/typeCoercion/native/division.sql.out index 934ae2b0cdcc9..6a889ff469811 100644 --- a/sql/core/src/test/resources/sql-tests/results/typeCoercion/native/division.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/typeCoercion/native/division.sql.out @@ -81,9 +81,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS TINYINT) / CAST(1 AS BINARY))\"", "left" : "\"TINYINT\"", - "right" : "\"BINARY\"" + "right" : "\"BINARY\"", + "sqlExpr" : "\"(CAST(1 AS TINYINT) / CAST(1 AS BINARY))\"" }, "queryContext" : [ { "objectType" : "", @@ -105,9 +105,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS TINYINT) / CAST(1 AS BOOLEAN))\"", "left" : "\"TINYINT\"", - "right" : "\"BOOLEAN\"" + "right" : "\"BOOLEAN\"", + "sqlExpr" : "\"(CAST(1 AS TINYINT) / CAST(1 AS BOOLEAN))\"" }, "queryContext" : [ { "objectType" : "", @@ -129,9 +129,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS TINYINT) / CAST(2017-12-11 09:30:00.0 AS TIMESTAMP))\"", "left" : "\"TINYINT\"", - "right" : "\"TIMESTAMP\"" + "right" : "\"TIMESTAMP\"", + "sqlExpr" : "\"(CAST(1 AS TINYINT) / CAST(2017-12-11 09:30:00.0 AS TIMESTAMP))\"" }, "queryContext" : [ { "objectType" : "", @@ -153,9 +153,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS TINYINT) / CAST(2017-12-11 09:30:00 AS DATE))\"", "left" : "\"TINYINT\"", - "right" : "\"DATE\"" + "right" : "\"DATE\"", + "sqlExpr" : "\"(CAST(1 AS TINYINT) / CAST(2017-12-11 09:30:00 AS DATE))\"" }, "queryContext" : [ { "objectType" : "", @@ -241,9 +241,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS SMALLINT) / CAST(1 AS BINARY))\"", "left" : "\"SMALLINT\"", - "right" : "\"BINARY\"" + "right" : "\"BINARY\"", + "sqlExpr" : "\"(CAST(1 AS SMALLINT) / CAST(1 AS BINARY))\"" }, "queryContext" : [ { "objectType" : "", @@ -265,9 +265,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS SMALLINT) / CAST(1 AS BOOLEAN))\"", "left" : "\"SMALLINT\"", - "right" : "\"BOOLEAN\"" + "right" : "\"BOOLEAN\"", + "sqlExpr" : "\"(CAST(1 AS SMALLINT) / CAST(1 AS BOOLEAN))\"" }, "queryContext" : [ { "objectType" : "", @@ -289,9 +289,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS SMALLINT) / CAST(2017-12-11 09:30:00.0 AS TIMESTAMP))\"", "left" : "\"SMALLINT\"", - "right" : "\"TIMESTAMP\"" + "right" : "\"TIMESTAMP\"", + "sqlExpr" : "\"(CAST(1 AS SMALLINT) / CAST(2017-12-11 09:30:00.0 AS TIMESTAMP))\"" }, "queryContext" : [ { "objectType" : "", @@ -313,9 +313,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS SMALLINT) / CAST(2017-12-11 09:30:00 AS DATE))\"", "left" : "\"SMALLINT\"", - "right" : "\"DATE\"" + "right" : "\"DATE\"", + "sqlExpr" : "\"(CAST(1 AS SMALLINT) / CAST(2017-12-11 09:30:00 AS DATE))\"" }, "queryContext" : [ { "objectType" : "", @@ -401,9 +401,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS INT) / CAST(1 AS BINARY))\"", "left" : "\"INT\"", - "right" : "\"BINARY\"" + "right" : "\"BINARY\"", + "sqlExpr" : "\"(CAST(1 AS INT) / CAST(1 AS BINARY))\"" }, "queryContext" : [ { "objectType" : "", @@ -425,9 +425,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS INT) / CAST(1 AS BOOLEAN))\"", "left" : "\"INT\"", - "right" : "\"BOOLEAN\"" + "right" : "\"BOOLEAN\"", + "sqlExpr" : "\"(CAST(1 AS INT) / CAST(1 AS BOOLEAN))\"" }, "queryContext" : [ { "objectType" : "", @@ -449,9 +449,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS INT) / CAST(2017-12-11 09:30:00.0 AS TIMESTAMP))\"", "left" : "\"INT\"", - "right" : "\"TIMESTAMP\"" + "right" : "\"TIMESTAMP\"", + "sqlExpr" : "\"(CAST(1 AS INT) / CAST(2017-12-11 09:30:00.0 AS TIMESTAMP))\"" }, "queryContext" : [ { "objectType" : "", @@ -473,9 +473,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS INT) / CAST(2017-12-11 09:30:00 AS DATE))\"", "left" : "\"INT\"", - "right" : "\"DATE\"" + "right" : "\"DATE\"", + "sqlExpr" : "\"(CAST(1 AS INT) / CAST(2017-12-11 09:30:00 AS DATE))\"" }, "queryContext" : [ { "objectType" : "", @@ -561,9 +561,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS BIGINT) / CAST(1 AS BINARY))\"", "left" : "\"BIGINT\"", - "right" : "\"BINARY\"" + "right" : "\"BINARY\"", + "sqlExpr" : "\"(CAST(1 AS BIGINT) / CAST(1 AS BINARY))\"" }, "queryContext" : [ { "objectType" : "", @@ -585,9 +585,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS BIGINT) / CAST(1 AS BOOLEAN))\"", "left" : "\"BIGINT\"", - "right" : "\"BOOLEAN\"" + "right" : "\"BOOLEAN\"", + "sqlExpr" : "\"(CAST(1 AS BIGINT) / CAST(1 AS BOOLEAN))\"" }, "queryContext" : [ { "objectType" : "", @@ -609,9 +609,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS BIGINT) / CAST(2017-12-11 09:30:00.0 AS TIMESTAMP))\"", "left" : "\"BIGINT\"", - "right" : "\"TIMESTAMP\"" + "right" : "\"TIMESTAMP\"", + "sqlExpr" : "\"(CAST(1 AS BIGINT) / CAST(2017-12-11 09:30:00.0 AS TIMESTAMP))\"" }, "queryContext" : [ { "objectType" : "", @@ -633,9 +633,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS BIGINT) / CAST(2017-12-11 09:30:00 AS DATE))\"", "left" : "\"BIGINT\"", - "right" : "\"DATE\"" + "right" : "\"DATE\"", + "sqlExpr" : "\"(CAST(1 AS BIGINT) / CAST(2017-12-11 09:30:00 AS DATE))\"" }, "queryContext" : [ { "objectType" : "", @@ -721,9 +721,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS FLOAT) / CAST(1 AS BINARY))\"", "left" : "\"FLOAT\"", - "right" : "\"BINARY\"" + "right" : "\"BINARY\"", + "sqlExpr" : "\"(CAST(1 AS FLOAT) / CAST(1 AS BINARY))\"" }, "queryContext" : [ { "objectType" : "", @@ -745,9 +745,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS FLOAT) / CAST(1 AS BOOLEAN))\"", "left" : "\"FLOAT\"", - "right" : "\"BOOLEAN\"" + "right" : "\"BOOLEAN\"", + "sqlExpr" : "\"(CAST(1 AS FLOAT) / CAST(1 AS BOOLEAN))\"" }, "queryContext" : [ { "objectType" : "", @@ -769,9 +769,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS FLOAT) / CAST(2017-12-11 09:30:00.0 AS TIMESTAMP))\"", "left" : "\"FLOAT\"", - "right" : "\"TIMESTAMP\"" + "right" : "\"TIMESTAMP\"", + "sqlExpr" : "\"(CAST(1 AS FLOAT) / CAST(2017-12-11 09:30:00.0 AS TIMESTAMP))\"" }, "queryContext" : [ { "objectType" : "", @@ -793,9 +793,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS FLOAT) / CAST(2017-12-11 09:30:00 AS DATE))\"", "left" : "\"FLOAT\"", - "right" : "\"DATE\"" + "right" : "\"DATE\"", + "sqlExpr" : "\"(CAST(1 AS FLOAT) / CAST(2017-12-11 09:30:00 AS DATE))\"" }, "queryContext" : [ { "objectType" : "", @@ -881,9 +881,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DOUBLE) / CAST(1 AS BINARY))\"", "left" : "\"DOUBLE\"", - "right" : "\"BINARY\"" + "right" : "\"BINARY\"", + "sqlExpr" : "\"(CAST(1 AS DOUBLE) / CAST(1 AS BINARY))\"" }, "queryContext" : [ { "objectType" : "", @@ -905,9 +905,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DOUBLE) / CAST(1 AS BOOLEAN))\"", "left" : "\"DOUBLE\"", - "right" : "\"BOOLEAN\"" + "right" : "\"BOOLEAN\"", + "sqlExpr" : "\"(CAST(1 AS DOUBLE) / CAST(1 AS BOOLEAN))\"" }, "queryContext" : [ { "objectType" : "", @@ -929,9 +929,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DOUBLE) / CAST(2017-12-11 09:30:00.0 AS TIMESTAMP))\"", "left" : "\"DOUBLE\"", - "right" : "\"TIMESTAMP\"" + "right" : "\"TIMESTAMP\"", + "sqlExpr" : "\"(CAST(1 AS DOUBLE) / CAST(2017-12-11 09:30:00.0 AS TIMESTAMP))\"" }, "queryContext" : [ { "objectType" : "", @@ -953,9 +953,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DOUBLE) / CAST(2017-12-11 09:30:00 AS DATE))\"", "left" : "\"DOUBLE\"", - "right" : "\"DATE\"" + "right" : "\"DATE\"", + "sqlExpr" : "\"(CAST(1 AS DOUBLE) / CAST(2017-12-11 09:30:00 AS DATE))\"" }, "queryContext" : [ { "objectType" : "", @@ -1041,9 +1041,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(10,0)) / CAST(1 AS BINARY))\"", "left" : "\"DECIMAL(10,0)\"", - "right" : "\"BINARY\"" + "right" : "\"BINARY\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(10,0)) / CAST(1 AS BINARY))\"" }, "queryContext" : [ { "objectType" : "", @@ -1065,9 +1065,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(10,0)) / CAST(1 AS BOOLEAN))\"", "left" : "\"DECIMAL(10,0)\"", - "right" : "\"BOOLEAN\"" + "right" : "\"BOOLEAN\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(10,0)) / CAST(1 AS BOOLEAN))\"" }, "queryContext" : [ { "objectType" : "", @@ -1089,9 +1089,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(10,0)) / CAST(2017-12-11 09:30:00.0 AS TIMESTAMP))\"", "left" : "\"DECIMAL(10,0)\"", - "right" : "\"TIMESTAMP\"" + "right" : "\"TIMESTAMP\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(10,0)) / CAST(2017-12-11 09:30:00.0 AS TIMESTAMP))\"" }, "queryContext" : [ { "objectType" : "", @@ -1113,9 +1113,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS DECIMAL(10,0)) / CAST(2017-12-11 09:30:00 AS DATE))\"", "left" : "\"DECIMAL(10,0)\"", - "right" : "\"DATE\"" + "right" : "\"DATE\"", + "sqlExpr" : "\"(CAST(1 AS DECIMAL(10,0)) / CAST(2017-12-11 09:30:00 AS DATE))\"" }, "queryContext" : [ { "objectType" : "", @@ -1201,9 +1201,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS STRING) / CAST(1 AS BINARY))\"", "left" : "\"DOUBLE\"", - "right" : "\"BINARY\"" + "right" : "\"BINARY\"", + "sqlExpr" : "\"(CAST(1 AS STRING) / CAST(1 AS BINARY))\"" }, "queryContext" : [ { "objectType" : "", @@ -1225,9 +1225,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS STRING) / CAST(1 AS BOOLEAN))\"", "left" : "\"DOUBLE\"", - "right" : "\"BOOLEAN\"" + "right" : "\"BOOLEAN\"", + "sqlExpr" : "\"(CAST(1 AS STRING) / CAST(1 AS BOOLEAN))\"" }, "queryContext" : [ { "objectType" : "", @@ -1249,9 +1249,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS STRING) / CAST(2017-12-11 09:30:00.0 AS TIMESTAMP))\"", "left" : "\"DOUBLE\"", - "right" : "\"TIMESTAMP\"" + "right" : "\"TIMESTAMP\"", + "sqlExpr" : "\"(CAST(1 AS STRING) / CAST(2017-12-11 09:30:00.0 AS TIMESTAMP))\"" }, "queryContext" : [ { "objectType" : "", @@ -1273,9 +1273,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS STRING) / CAST(2017-12-11 09:30:00 AS DATE))\"", "left" : "\"DOUBLE\"", - "right" : "\"DATE\"" + "right" : "\"DATE\"", + "sqlExpr" : "\"(CAST(1 AS STRING) / CAST(2017-12-11 09:30:00 AS DATE))\"" }, "queryContext" : [ { "objectType" : "", @@ -1297,9 +1297,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS BINARY) / CAST(1 AS TINYINT))\"", "left" : "\"BINARY\"", - "right" : "\"TINYINT\"" + "right" : "\"TINYINT\"", + "sqlExpr" : "\"(CAST(1 AS BINARY) / CAST(1 AS TINYINT))\"" }, "queryContext" : [ { "objectType" : "", @@ -1321,9 +1321,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS BINARY) / CAST(1 AS SMALLINT))\"", "left" : "\"BINARY\"", - "right" : "\"SMALLINT\"" + "right" : "\"SMALLINT\"", + "sqlExpr" : "\"(CAST(1 AS BINARY) / CAST(1 AS SMALLINT))\"" }, "queryContext" : [ { "objectType" : "", @@ -1345,9 +1345,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS BINARY) / CAST(1 AS INT))\"", "left" : "\"BINARY\"", - "right" : "\"INT\"" + "right" : "\"INT\"", + "sqlExpr" : "\"(CAST(1 AS BINARY) / CAST(1 AS INT))\"" }, "queryContext" : [ { "objectType" : "", @@ -1369,9 +1369,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS BINARY) / CAST(1 AS BIGINT))\"", "left" : "\"BINARY\"", - "right" : "\"BIGINT\"" + "right" : "\"BIGINT\"", + "sqlExpr" : "\"(CAST(1 AS BINARY) / CAST(1 AS BIGINT))\"" }, "queryContext" : [ { "objectType" : "", @@ -1393,9 +1393,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS BINARY) / CAST(1 AS FLOAT))\"", "left" : "\"BINARY\"", - "right" : "\"FLOAT\"" + "right" : "\"FLOAT\"", + "sqlExpr" : "\"(CAST(1 AS BINARY) / CAST(1 AS FLOAT))\"" }, "queryContext" : [ { "objectType" : "", @@ -1417,9 +1417,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS BINARY) / CAST(1 AS DOUBLE))\"", "left" : "\"BINARY\"", - "right" : "\"DOUBLE\"" + "right" : "\"DOUBLE\"", + "sqlExpr" : "\"(CAST(1 AS BINARY) / CAST(1 AS DOUBLE))\"" }, "queryContext" : [ { "objectType" : "", @@ -1441,9 +1441,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS BINARY) / CAST(1 AS DECIMAL(10,0)))\"", "left" : "\"BINARY\"", - "right" : "\"DECIMAL(10,0)\"" + "right" : "\"DECIMAL(10,0)\"", + "sqlExpr" : "\"(CAST(1 AS BINARY) / CAST(1 AS DECIMAL(10,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -1465,9 +1465,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS BINARY) / CAST(1 AS STRING))\"", "left" : "\"BINARY\"", - "right" : "\"DOUBLE\"" + "right" : "\"DOUBLE\"", + "sqlExpr" : "\"(CAST(1 AS BINARY) / CAST(1 AS STRING))\"" }, "queryContext" : [ { "objectType" : "", @@ -1489,9 +1489,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_WRONG_TYPE", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS BINARY) / CAST(1 AS BINARY))\"", + "actualDataType" : "\"BINARY\"", "inputType" : "(\"DOUBLE\" or \"DECIMAL\")", - "actualDataType" : "\"BINARY\"" + "sqlExpr" : "\"(CAST(1 AS BINARY) / CAST(1 AS BINARY))\"" }, "queryContext" : [ { "objectType" : "", @@ -1513,9 +1513,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS BINARY) / CAST(1 AS BOOLEAN))\"", "left" : "\"BINARY\"", - "right" : "\"BOOLEAN\"" + "right" : "\"BOOLEAN\"", + "sqlExpr" : "\"(CAST(1 AS BINARY) / CAST(1 AS BOOLEAN))\"" }, "queryContext" : [ { "objectType" : "", @@ -1537,9 +1537,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS BINARY) / CAST(2017-12-11 09:30:00.0 AS TIMESTAMP))\"", "left" : "\"BINARY\"", - "right" : "\"TIMESTAMP\"" + "right" : "\"TIMESTAMP\"", + "sqlExpr" : "\"(CAST(1 AS BINARY) / CAST(2017-12-11 09:30:00.0 AS TIMESTAMP))\"" }, "queryContext" : [ { "objectType" : "", @@ -1561,9 +1561,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS BINARY) / CAST(2017-12-11 09:30:00 AS DATE))\"", "left" : "\"BINARY\"", - "right" : "\"DATE\"" + "right" : "\"DATE\"", + "sqlExpr" : "\"(CAST(1 AS BINARY) / CAST(2017-12-11 09:30:00 AS DATE))\"" }, "queryContext" : [ { "objectType" : "", @@ -1585,9 +1585,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS BOOLEAN) / CAST(1 AS TINYINT))\"", "left" : "\"BOOLEAN\"", - "right" : "\"TINYINT\"" + "right" : "\"TINYINT\"", + "sqlExpr" : "\"(CAST(1 AS BOOLEAN) / CAST(1 AS TINYINT))\"" }, "queryContext" : [ { "objectType" : "", @@ -1609,9 +1609,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS BOOLEAN) / CAST(1 AS SMALLINT))\"", "left" : "\"BOOLEAN\"", - "right" : "\"SMALLINT\"" + "right" : "\"SMALLINT\"", + "sqlExpr" : "\"(CAST(1 AS BOOLEAN) / CAST(1 AS SMALLINT))\"" }, "queryContext" : [ { "objectType" : "", @@ -1633,9 +1633,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS BOOLEAN) / CAST(1 AS INT))\"", "left" : "\"BOOLEAN\"", - "right" : "\"INT\"" + "right" : "\"INT\"", + "sqlExpr" : "\"(CAST(1 AS BOOLEAN) / CAST(1 AS INT))\"" }, "queryContext" : [ { "objectType" : "", @@ -1657,9 +1657,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS BOOLEAN) / CAST(1 AS BIGINT))\"", "left" : "\"BOOLEAN\"", - "right" : "\"BIGINT\"" + "right" : "\"BIGINT\"", + "sqlExpr" : "\"(CAST(1 AS BOOLEAN) / CAST(1 AS BIGINT))\"" }, "queryContext" : [ { "objectType" : "", @@ -1681,9 +1681,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS BOOLEAN) / CAST(1 AS FLOAT))\"", "left" : "\"BOOLEAN\"", - "right" : "\"FLOAT\"" + "right" : "\"FLOAT\"", + "sqlExpr" : "\"(CAST(1 AS BOOLEAN) / CAST(1 AS FLOAT))\"" }, "queryContext" : [ { "objectType" : "", @@ -1705,9 +1705,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS BOOLEAN) / CAST(1 AS DOUBLE))\"", "left" : "\"BOOLEAN\"", - "right" : "\"DOUBLE\"" + "right" : "\"DOUBLE\"", + "sqlExpr" : "\"(CAST(1 AS BOOLEAN) / CAST(1 AS DOUBLE))\"" }, "queryContext" : [ { "objectType" : "", @@ -1729,9 +1729,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS BOOLEAN) / CAST(1 AS DECIMAL(10,0)))\"", "left" : "\"BOOLEAN\"", - "right" : "\"DECIMAL(10,0)\"" + "right" : "\"DECIMAL(10,0)\"", + "sqlExpr" : "\"(CAST(1 AS BOOLEAN) / CAST(1 AS DECIMAL(10,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -1753,9 +1753,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS BOOLEAN) / CAST(1 AS STRING))\"", "left" : "\"BOOLEAN\"", - "right" : "\"DOUBLE\"" + "right" : "\"DOUBLE\"", + "sqlExpr" : "\"(CAST(1 AS BOOLEAN) / CAST(1 AS STRING))\"" }, "queryContext" : [ { "objectType" : "", @@ -1777,9 +1777,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS BOOLEAN) / CAST(1 AS BINARY))\"", "left" : "\"BOOLEAN\"", - "right" : "\"BINARY\"" + "right" : "\"BINARY\"", + "sqlExpr" : "\"(CAST(1 AS BOOLEAN) / CAST(1 AS BINARY))\"" }, "queryContext" : [ { "objectType" : "", @@ -1801,9 +1801,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_WRONG_TYPE", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS BOOLEAN) / CAST(1 AS BOOLEAN))\"", + "actualDataType" : "\"BOOLEAN\"", "inputType" : "(\"DOUBLE\" or \"DECIMAL\")", - "actualDataType" : "\"BOOLEAN\"" + "sqlExpr" : "\"(CAST(1 AS BOOLEAN) / CAST(1 AS BOOLEAN))\"" }, "queryContext" : [ { "objectType" : "", @@ -1825,9 +1825,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS BOOLEAN) / CAST(2017-12-11 09:30:00.0 AS TIMESTAMP))\"", "left" : "\"BOOLEAN\"", - "right" : "\"TIMESTAMP\"" + "right" : "\"TIMESTAMP\"", + "sqlExpr" : "\"(CAST(1 AS BOOLEAN) / CAST(2017-12-11 09:30:00.0 AS TIMESTAMP))\"" }, "queryContext" : [ { "objectType" : "", @@ -1849,9 +1849,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS BOOLEAN) / CAST(2017-12-11 09:30:00 AS DATE))\"", "left" : "\"BOOLEAN\"", - "right" : "\"DATE\"" + "right" : "\"DATE\"", + "sqlExpr" : "\"(CAST(1 AS BOOLEAN) / CAST(2017-12-11 09:30:00 AS DATE))\"" }, "queryContext" : [ { "objectType" : "", @@ -1873,9 +1873,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(2017-12-11 09:30:00.0 AS TIMESTAMP) / CAST(1 AS TINYINT))\"", "left" : "\"TIMESTAMP\"", - "right" : "\"TINYINT\"" + "right" : "\"TINYINT\"", + "sqlExpr" : "\"(CAST(2017-12-11 09:30:00.0 AS TIMESTAMP) / CAST(1 AS TINYINT))\"" }, "queryContext" : [ { "objectType" : "", @@ -1897,9 +1897,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(2017-12-11 09:30:00.0 AS TIMESTAMP) / CAST(1 AS SMALLINT))\"", "left" : "\"TIMESTAMP\"", - "right" : "\"SMALLINT\"" + "right" : "\"SMALLINT\"", + "sqlExpr" : "\"(CAST(2017-12-11 09:30:00.0 AS TIMESTAMP) / CAST(1 AS SMALLINT))\"" }, "queryContext" : [ { "objectType" : "", @@ -1921,9 +1921,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(2017-12-11 09:30:00.0 AS TIMESTAMP) / CAST(1 AS INT))\"", "left" : "\"TIMESTAMP\"", - "right" : "\"INT\"" + "right" : "\"INT\"", + "sqlExpr" : "\"(CAST(2017-12-11 09:30:00.0 AS TIMESTAMP) / CAST(1 AS INT))\"" }, "queryContext" : [ { "objectType" : "", @@ -1945,9 +1945,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(2017-12-11 09:30:00.0 AS TIMESTAMP) / CAST(1 AS BIGINT))\"", "left" : "\"TIMESTAMP\"", - "right" : "\"BIGINT\"" + "right" : "\"BIGINT\"", + "sqlExpr" : "\"(CAST(2017-12-11 09:30:00.0 AS TIMESTAMP) / CAST(1 AS BIGINT))\"" }, "queryContext" : [ { "objectType" : "", @@ -1969,9 +1969,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(2017-12-11 09:30:00.0 AS TIMESTAMP) / CAST(1 AS FLOAT))\"", "left" : "\"TIMESTAMP\"", - "right" : "\"FLOAT\"" + "right" : "\"FLOAT\"", + "sqlExpr" : "\"(CAST(2017-12-11 09:30:00.0 AS TIMESTAMP) / CAST(1 AS FLOAT))\"" }, "queryContext" : [ { "objectType" : "", @@ -1993,9 +1993,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(2017-12-11 09:30:00.0 AS TIMESTAMP) / CAST(1 AS DOUBLE))\"", "left" : "\"TIMESTAMP\"", - "right" : "\"DOUBLE\"" + "right" : "\"DOUBLE\"", + "sqlExpr" : "\"(CAST(2017-12-11 09:30:00.0 AS TIMESTAMP) / CAST(1 AS DOUBLE))\"" }, "queryContext" : [ { "objectType" : "", @@ -2017,9 +2017,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(2017-12-11 09:30:00.0 AS TIMESTAMP) / CAST(1 AS DECIMAL(10,0)))\"", "left" : "\"TIMESTAMP\"", - "right" : "\"DECIMAL(10,0)\"" + "right" : "\"DECIMAL(10,0)\"", + "sqlExpr" : "\"(CAST(2017-12-11 09:30:00.0 AS TIMESTAMP) / CAST(1 AS DECIMAL(10,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -2041,9 +2041,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(2017-12-11 09:30:00.0 AS TIMESTAMP) / CAST(1 AS STRING))\"", "left" : "\"TIMESTAMP\"", - "right" : "\"DOUBLE\"" + "right" : "\"DOUBLE\"", + "sqlExpr" : "\"(CAST(2017-12-11 09:30:00.0 AS TIMESTAMP) / CAST(1 AS STRING))\"" }, "queryContext" : [ { "objectType" : "", @@ -2065,9 +2065,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(2017-12-11 09:30:00.0 AS TIMESTAMP) / CAST(1 AS BINARY))\"", "left" : "\"TIMESTAMP\"", - "right" : "\"BINARY\"" + "right" : "\"BINARY\"", + "sqlExpr" : "\"(CAST(2017-12-11 09:30:00.0 AS TIMESTAMP) / CAST(1 AS BINARY))\"" }, "queryContext" : [ { "objectType" : "", @@ -2089,9 +2089,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(2017-12-11 09:30:00.0 AS TIMESTAMP) / CAST(1 AS BOOLEAN))\"", "left" : "\"TIMESTAMP\"", - "right" : "\"BOOLEAN\"" + "right" : "\"BOOLEAN\"", + "sqlExpr" : "\"(CAST(2017-12-11 09:30:00.0 AS TIMESTAMP) / CAST(1 AS BOOLEAN))\"" }, "queryContext" : [ { "objectType" : "", @@ -2113,9 +2113,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_WRONG_TYPE", "messageParameters" : { - "sqlExpr" : "\"(CAST(2017-12-11 09:30:00.0 AS TIMESTAMP) / CAST(2017-12-11 09:30:00.0 AS TIMESTAMP))\"", + "actualDataType" : "\"TIMESTAMP\"", "inputType" : "(\"DOUBLE\" or \"DECIMAL\")", - "actualDataType" : "\"TIMESTAMP\"" + "sqlExpr" : "\"(CAST(2017-12-11 09:30:00.0 AS TIMESTAMP) / CAST(2017-12-11 09:30:00.0 AS TIMESTAMP))\"" }, "queryContext" : [ { "objectType" : "", @@ -2137,9 +2137,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(2017-12-11 09:30:00.0 AS TIMESTAMP) / CAST(2017-12-11 09:30:00 AS DATE))\"", "left" : "\"TIMESTAMP\"", - "right" : "\"DATE\"" + "right" : "\"DATE\"", + "sqlExpr" : "\"(CAST(2017-12-11 09:30:00.0 AS TIMESTAMP) / CAST(2017-12-11 09:30:00 AS DATE))\"" }, "queryContext" : [ { "objectType" : "", @@ -2161,9 +2161,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(2017-12-11 09:30:00 AS DATE) / CAST(1 AS TINYINT))\"", "left" : "\"DATE\"", - "right" : "\"TINYINT\"" + "right" : "\"TINYINT\"", + "sqlExpr" : "\"(CAST(2017-12-11 09:30:00 AS DATE) / CAST(1 AS TINYINT))\"" }, "queryContext" : [ { "objectType" : "", @@ -2185,9 +2185,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(2017-12-11 09:30:00 AS DATE) / CAST(1 AS SMALLINT))\"", "left" : "\"DATE\"", - "right" : "\"SMALLINT\"" + "right" : "\"SMALLINT\"", + "sqlExpr" : "\"(CAST(2017-12-11 09:30:00 AS DATE) / CAST(1 AS SMALLINT))\"" }, "queryContext" : [ { "objectType" : "", @@ -2209,9 +2209,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(2017-12-11 09:30:00 AS DATE) / CAST(1 AS INT))\"", "left" : "\"DATE\"", - "right" : "\"INT\"" + "right" : "\"INT\"", + "sqlExpr" : "\"(CAST(2017-12-11 09:30:00 AS DATE) / CAST(1 AS INT))\"" }, "queryContext" : [ { "objectType" : "", @@ -2233,9 +2233,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(2017-12-11 09:30:00 AS DATE) / CAST(1 AS BIGINT))\"", "left" : "\"DATE\"", - "right" : "\"BIGINT\"" + "right" : "\"BIGINT\"", + "sqlExpr" : "\"(CAST(2017-12-11 09:30:00 AS DATE) / CAST(1 AS BIGINT))\"" }, "queryContext" : [ { "objectType" : "", @@ -2257,9 +2257,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(2017-12-11 09:30:00 AS DATE) / CAST(1 AS FLOAT))\"", "left" : "\"DATE\"", - "right" : "\"FLOAT\"" + "right" : "\"FLOAT\"", + "sqlExpr" : "\"(CAST(2017-12-11 09:30:00 AS DATE) / CAST(1 AS FLOAT))\"" }, "queryContext" : [ { "objectType" : "", @@ -2281,9 +2281,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(2017-12-11 09:30:00 AS DATE) / CAST(1 AS DOUBLE))\"", "left" : "\"DATE\"", - "right" : "\"DOUBLE\"" + "right" : "\"DOUBLE\"", + "sqlExpr" : "\"(CAST(2017-12-11 09:30:00 AS DATE) / CAST(1 AS DOUBLE))\"" }, "queryContext" : [ { "objectType" : "", @@ -2305,9 +2305,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(2017-12-11 09:30:00 AS DATE) / CAST(1 AS DECIMAL(10,0)))\"", "left" : "\"DATE\"", - "right" : "\"DECIMAL(10,0)\"" + "right" : "\"DECIMAL(10,0)\"", + "sqlExpr" : "\"(CAST(2017-12-11 09:30:00 AS DATE) / CAST(1 AS DECIMAL(10,0)))\"" }, "queryContext" : [ { "objectType" : "", @@ -2329,9 +2329,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(2017-12-11 09:30:00 AS DATE) / CAST(1 AS STRING))\"", "left" : "\"DATE\"", - "right" : "\"DOUBLE\"" + "right" : "\"DOUBLE\"", + "sqlExpr" : "\"(CAST(2017-12-11 09:30:00 AS DATE) / CAST(1 AS STRING))\"" }, "queryContext" : [ { "objectType" : "", @@ -2353,9 +2353,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(2017-12-11 09:30:00 AS DATE) / CAST(1 AS BINARY))\"", "left" : "\"DATE\"", - "right" : "\"BINARY\"" + "right" : "\"BINARY\"", + "sqlExpr" : "\"(CAST(2017-12-11 09:30:00 AS DATE) / CAST(1 AS BINARY))\"" }, "queryContext" : [ { "objectType" : "", @@ -2377,9 +2377,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(2017-12-11 09:30:00 AS DATE) / CAST(1 AS BOOLEAN))\"", "left" : "\"DATE\"", - "right" : "\"BOOLEAN\"" + "right" : "\"BOOLEAN\"", + "sqlExpr" : "\"(CAST(2017-12-11 09:30:00 AS DATE) / CAST(1 AS BOOLEAN))\"" }, "queryContext" : [ { "objectType" : "", @@ -2401,9 +2401,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(2017-12-11 09:30:00 AS DATE) / CAST(2017-12-11 09:30:00.0 AS TIMESTAMP))\"", "left" : "\"DATE\"", - "right" : "\"TIMESTAMP\"" + "right" : "\"TIMESTAMP\"", + "sqlExpr" : "\"(CAST(2017-12-11 09:30:00 AS DATE) / CAST(2017-12-11 09:30:00.0 AS TIMESTAMP))\"" }, "queryContext" : [ { "objectType" : "", @@ -2425,9 +2425,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_WRONG_TYPE", "messageParameters" : { - "sqlExpr" : "\"(CAST(2017-12-11 09:30:00 AS DATE) / CAST(2017-12-11 09:30:00 AS DATE))\"", + "actualDataType" : "\"DATE\"", "inputType" : "(\"DOUBLE\" or \"DECIMAL\")", - "actualDataType" : "\"DATE\"" + "sqlExpr" : "\"(CAST(2017-12-11 09:30:00 AS DATE) / CAST(2017-12-11 09:30:00 AS DATE))\"" }, "queryContext" : [ { "objectType" : "", diff --git a/sql/core/src/test/resources/sql-tests/results/typeCoercion/native/promoteStrings.sql.out b/sql/core/src/test/resources/sql-tests/results/typeCoercion/native/promoteStrings.sql.out index 7f953de75bf9c..edb8151ae794b 100644 --- a/sql/core/src/test/resources/sql-tests/results/typeCoercion/native/promoteStrings.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/typeCoercion/native/promoteStrings.sql.out @@ -81,9 +81,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(1 + CAST(1 AS BINARY))\"", "left" : "\"DOUBLE\"", - "right" : "\"BINARY\"" + "right" : "\"BINARY\"", + "sqlExpr" : "\"(1 + CAST(1 AS BINARY))\"" }, "queryContext" : [ { "objectType" : "", @@ -105,9 +105,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(1 + CAST(1 AS BOOLEAN))\"", "left" : "\"DOUBLE\"", - "right" : "\"BOOLEAN\"" + "right" : "\"BOOLEAN\"", + "sqlExpr" : "\"(1 + CAST(1 AS BOOLEAN))\"" }, "queryContext" : [ { "objectType" : "", @@ -129,9 +129,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(1 + CAST(2017-12-11 09:30:00.0 AS TIMESTAMP))\"", "left" : "\"DOUBLE\"", - "right" : "\"TIMESTAMP\"" + "right" : "\"TIMESTAMP\"", + "sqlExpr" : "\"(1 + CAST(2017-12-11 09:30:00.0 AS TIMESTAMP))\"" }, "queryContext" : [ { "objectType" : "", @@ -226,9 +226,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(1 - CAST(1 AS BINARY))\"", "left" : "\"DOUBLE\"", - "right" : "\"BINARY\"" + "right" : "\"BINARY\"", + "sqlExpr" : "\"(1 - CAST(1 AS BINARY))\"" }, "queryContext" : [ { "objectType" : "", @@ -250,9 +250,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(1 - CAST(1 AS BOOLEAN))\"", "left" : "\"DOUBLE\"", - "right" : "\"BOOLEAN\"" + "right" : "\"BOOLEAN\"", + "sqlExpr" : "\"(1 - CAST(1 AS BOOLEAN))\"" }, "queryContext" : [ { "objectType" : "", @@ -355,9 +355,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(1 * CAST(1 AS BINARY))\"", "left" : "\"DOUBLE\"", - "right" : "\"BINARY\"" + "right" : "\"BINARY\"", + "sqlExpr" : "\"(1 * CAST(1 AS BINARY))\"" }, "queryContext" : [ { "objectType" : "", @@ -379,9 +379,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(1 * CAST(1 AS BOOLEAN))\"", "left" : "\"DOUBLE\"", - "right" : "\"BOOLEAN\"" + "right" : "\"BOOLEAN\"", + "sqlExpr" : "\"(1 * CAST(1 AS BOOLEAN))\"" }, "queryContext" : [ { "objectType" : "", @@ -403,9 +403,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(1 * CAST(2017-12-11 09:30:00.0 AS TIMESTAMP))\"", "left" : "\"DOUBLE\"", - "right" : "\"TIMESTAMP\"" + "right" : "\"TIMESTAMP\"", + "sqlExpr" : "\"(1 * CAST(2017-12-11 09:30:00.0 AS TIMESTAMP))\"" }, "queryContext" : [ { "objectType" : "", @@ -427,9 +427,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(1 * CAST(2017-12-11 09:30:00 AS DATE))\"", "left" : "\"DOUBLE\"", - "right" : "\"DATE\"" + "right" : "\"DATE\"", + "sqlExpr" : "\"(1 * CAST(2017-12-11 09:30:00 AS DATE))\"" }, "queryContext" : [ { "objectType" : "", @@ -515,9 +515,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(1 / CAST(1 AS BINARY))\"", "left" : "\"DOUBLE\"", - "right" : "\"BINARY\"" + "right" : "\"BINARY\"", + "sqlExpr" : "\"(1 / CAST(1 AS BINARY))\"" }, "queryContext" : [ { "objectType" : "", @@ -539,9 +539,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(1 / CAST(1 AS BOOLEAN))\"", "left" : "\"DOUBLE\"", - "right" : "\"BOOLEAN\"" + "right" : "\"BOOLEAN\"", + "sqlExpr" : "\"(1 / CAST(1 AS BOOLEAN))\"" }, "queryContext" : [ { "objectType" : "", @@ -563,9 +563,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(1 / CAST(2017-12-11 09:30:00.0 AS TIMESTAMP))\"", "left" : "\"DOUBLE\"", - "right" : "\"TIMESTAMP\"" + "right" : "\"TIMESTAMP\"", + "sqlExpr" : "\"(1 / CAST(2017-12-11 09:30:00.0 AS TIMESTAMP))\"" }, "queryContext" : [ { "objectType" : "", @@ -587,9 +587,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(1 / CAST(2017-12-11 09:30:00 AS DATE))\"", "left" : "\"DOUBLE\"", - "right" : "\"DATE\"" + "right" : "\"DATE\"", + "sqlExpr" : "\"(1 / CAST(2017-12-11 09:30:00 AS DATE))\"" }, "queryContext" : [ { "objectType" : "", @@ -675,9 +675,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(1 % CAST(1 AS BINARY))\"", "left" : "\"DOUBLE\"", - "right" : "\"BINARY\"" + "right" : "\"BINARY\"", + "sqlExpr" : "\"(1 % CAST(1 AS BINARY))\"" }, "queryContext" : [ { "objectType" : "", @@ -699,9 +699,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(1 % CAST(1 AS BOOLEAN))\"", "left" : "\"DOUBLE\"", - "right" : "\"BOOLEAN\"" + "right" : "\"BOOLEAN\"", + "sqlExpr" : "\"(1 % CAST(1 AS BOOLEAN))\"" }, "queryContext" : [ { "objectType" : "", @@ -723,9 +723,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(1 % CAST(2017-12-11 09:30:00.0 AS TIMESTAMP))\"", "left" : "\"DOUBLE\"", - "right" : "\"TIMESTAMP\"" + "right" : "\"TIMESTAMP\"", + "sqlExpr" : "\"(1 % CAST(2017-12-11 09:30:00.0 AS TIMESTAMP))\"" }, "queryContext" : [ { "objectType" : "", @@ -747,9 +747,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(1 % CAST(2017-12-11 09:30:00 AS DATE))\"", "left" : "\"DOUBLE\"", - "right" : "\"DATE\"" + "right" : "\"DATE\"", + "sqlExpr" : "\"(1 % CAST(2017-12-11 09:30:00 AS DATE))\"" }, "queryContext" : [ { "objectType" : "", @@ -835,9 +835,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"pmod(1, CAST(1 AS BINARY))\"", "left" : "\"DOUBLE\"", - "right" : "\"BINARY\"" + "right" : "\"BINARY\"", + "sqlExpr" : "\"pmod(1, CAST(1 AS BINARY))\"" }, "queryContext" : [ { "objectType" : "", @@ -859,9 +859,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"pmod(1, CAST(1 AS BOOLEAN))\"", "left" : "\"DOUBLE\"", - "right" : "\"BOOLEAN\"" + "right" : "\"BOOLEAN\"", + "sqlExpr" : "\"pmod(1, CAST(1 AS BOOLEAN))\"" }, "queryContext" : [ { "objectType" : "", @@ -883,9 +883,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"pmod(1, CAST(2017-12-11 09:30:00.0 AS TIMESTAMP))\"", "left" : "\"DOUBLE\"", - "right" : "\"TIMESTAMP\"" + "right" : "\"TIMESTAMP\"", + "sqlExpr" : "\"pmod(1, CAST(2017-12-11 09:30:00.0 AS TIMESTAMP))\"" }, "queryContext" : [ { "objectType" : "", @@ -907,9 +907,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"pmod(1, CAST(2017-12-11 09:30:00 AS DATE))\"", "left" : "\"DOUBLE\"", - "right" : "\"DATE\"" + "right" : "\"DATE\"", + "sqlExpr" : "\"pmod(1, CAST(2017-12-11 09:30:00 AS DATE))\"" }, "queryContext" : [ { "objectType" : "", @@ -987,9 +987,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS BINARY) + 1)\"", "left" : "\"BINARY\"", - "right" : "\"DOUBLE\"" + "right" : "\"DOUBLE\"", + "sqlExpr" : "\"(CAST(1 AS BINARY) + 1)\"" }, "queryContext" : [ { "objectType" : "", @@ -1011,9 +1011,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS BOOLEAN) + 1)\"", "left" : "\"BOOLEAN\"", - "right" : "\"DOUBLE\"" + "right" : "\"DOUBLE\"", + "sqlExpr" : "\"(CAST(1 AS BOOLEAN) + 1)\"" }, "queryContext" : [ { "objectType" : "", @@ -1035,9 +1035,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(2017-12-11 09:30:00.0 AS TIMESTAMP) + 1)\"", "left" : "\"TIMESTAMP\"", - "right" : "\"DOUBLE\"" + "right" : "\"DOUBLE\"", + "sqlExpr" : "\"(CAST(2017-12-11 09:30:00.0 AS TIMESTAMP) + 1)\"" }, "queryContext" : [ { "objectType" : "", @@ -1124,9 +1124,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS BINARY) - 1)\"", "left" : "\"BINARY\"", - "right" : "\"DOUBLE\"" + "right" : "\"DOUBLE\"", + "sqlExpr" : "\"(CAST(1 AS BINARY) - 1)\"" }, "queryContext" : [ { "objectType" : "", @@ -1148,9 +1148,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS BOOLEAN) - 1)\"", "left" : "\"BOOLEAN\"", - "right" : "\"DOUBLE\"" + "right" : "\"DOUBLE\"", + "sqlExpr" : "\"(CAST(1 AS BOOLEAN) - 1)\"" }, "queryContext" : [ { "objectType" : "", @@ -1246,9 +1246,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS BINARY) * 1)\"", "left" : "\"BINARY\"", - "right" : "\"DOUBLE\"" + "right" : "\"DOUBLE\"", + "sqlExpr" : "\"(CAST(1 AS BINARY) * 1)\"" }, "queryContext" : [ { "objectType" : "", @@ -1270,9 +1270,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS BOOLEAN) * 1)\"", "left" : "\"BOOLEAN\"", - "right" : "\"DOUBLE\"" + "right" : "\"DOUBLE\"", + "sqlExpr" : "\"(CAST(1 AS BOOLEAN) * 1)\"" }, "queryContext" : [ { "objectType" : "", @@ -1294,9 +1294,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(2017-12-11 09:30:00.0 AS TIMESTAMP) * 1)\"", "left" : "\"TIMESTAMP\"", - "right" : "\"DOUBLE\"" + "right" : "\"DOUBLE\"", + "sqlExpr" : "\"(CAST(2017-12-11 09:30:00.0 AS TIMESTAMP) * 1)\"" }, "queryContext" : [ { "objectType" : "", @@ -1318,9 +1318,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(2017-12-11 09:30:00 AS DATE) * 1)\"", "left" : "\"DATE\"", - "right" : "\"DOUBLE\"" + "right" : "\"DOUBLE\"", + "sqlExpr" : "\"(CAST(2017-12-11 09:30:00 AS DATE) * 1)\"" }, "queryContext" : [ { "objectType" : "", @@ -1398,9 +1398,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS BINARY) / 1)\"", "left" : "\"BINARY\"", - "right" : "\"DOUBLE\"" + "right" : "\"DOUBLE\"", + "sqlExpr" : "\"(CAST(1 AS BINARY) / 1)\"" }, "queryContext" : [ { "objectType" : "", @@ -1422,9 +1422,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS BOOLEAN) / 1)\"", "left" : "\"BOOLEAN\"", - "right" : "\"DOUBLE\"" + "right" : "\"DOUBLE\"", + "sqlExpr" : "\"(CAST(1 AS BOOLEAN) / 1)\"" }, "queryContext" : [ { "objectType" : "", @@ -1446,9 +1446,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(2017-12-11 09:30:00.0 AS TIMESTAMP) / 1)\"", "left" : "\"TIMESTAMP\"", - "right" : "\"DOUBLE\"" + "right" : "\"DOUBLE\"", + "sqlExpr" : "\"(CAST(2017-12-11 09:30:00.0 AS TIMESTAMP) / 1)\"" }, "queryContext" : [ { "objectType" : "", @@ -1470,9 +1470,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(2017-12-11 09:30:00 AS DATE) / 1)\"", "left" : "\"DATE\"", - "right" : "\"DOUBLE\"" + "right" : "\"DOUBLE\"", + "sqlExpr" : "\"(CAST(2017-12-11 09:30:00 AS DATE) / 1)\"" }, "queryContext" : [ { "objectType" : "", @@ -1550,9 +1550,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS BINARY) % 1)\"", "left" : "\"BINARY\"", - "right" : "\"DOUBLE\"" + "right" : "\"DOUBLE\"", + "sqlExpr" : "\"(CAST(1 AS BINARY) % 1)\"" }, "queryContext" : [ { "objectType" : "", @@ -1574,9 +1574,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(1 AS BOOLEAN) % 1)\"", "left" : "\"BOOLEAN\"", - "right" : "\"DOUBLE\"" + "right" : "\"DOUBLE\"", + "sqlExpr" : "\"(CAST(1 AS BOOLEAN) % 1)\"" }, "queryContext" : [ { "objectType" : "", @@ -1598,9 +1598,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(2017-12-11 09:30:00.0 AS TIMESTAMP) % 1)\"", "left" : "\"TIMESTAMP\"", - "right" : "\"DOUBLE\"" + "right" : "\"DOUBLE\"", + "sqlExpr" : "\"(CAST(2017-12-11 09:30:00.0 AS TIMESTAMP) % 1)\"" }, "queryContext" : [ { "objectType" : "", @@ -1622,9 +1622,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"(CAST(2017-12-11 09:30:00 AS DATE) % 1)\"", "left" : "\"DATE\"", - "right" : "\"DOUBLE\"" + "right" : "\"DOUBLE\"", + "sqlExpr" : "\"(CAST(2017-12-11 09:30:00 AS DATE) % 1)\"" }, "queryContext" : [ { "objectType" : "", @@ -1702,9 +1702,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"pmod(CAST(1 AS BINARY), 1)\"", "left" : "\"BINARY\"", - "right" : "\"DOUBLE\"" + "right" : "\"DOUBLE\"", + "sqlExpr" : "\"pmod(CAST(1 AS BINARY), 1)\"" }, "queryContext" : [ { "objectType" : "", @@ -1726,9 +1726,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"pmod(CAST(1 AS BOOLEAN), 1)\"", "left" : "\"BOOLEAN\"", - "right" : "\"DOUBLE\"" + "right" : "\"DOUBLE\"", + "sqlExpr" : "\"pmod(CAST(1 AS BOOLEAN), 1)\"" }, "queryContext" : [ { "objectType" : "", @@ -1750,9 +1750,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"pmod(CAST(2017-12-11 09:30:00.0 AS TIMESTAMP), 1)\"", "left" : "\"TIMESTAMP\"", - "right" : "\"DOUBLE\"" + "right" : "\"DOUBLE\"", + "sqlExpr" : "\"pmod(CAST(2017-12-11 09:30:00.0 AS TIMESTAMP), 1)\"" }, "queryContext" : [ { "objectType" : "", @@ -1774,9 +1774,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "DATATYPE_MISMATCH", "errorSubClass" : "BINARY_OP_DIFF_TYPES", "messageParameters" : { - "sqlExpr" : "\"pmod(CAST(2017-12-11 09:30:00 AS DATE), 1)\"", "left" : "\"DATE\"", - "right" : "\"DOUBLE\"" + "right" : "\"DOUBLE\"", + "sqlExpr" : "\"pmod(CAST(2017-12-11 09:30:00 AS DATE), 1)\"" }, "queryContext" : [ { "objectType" : "", diff --git a/sql/core/src/test/resources/sql-tests/results/typeCoercion/native/stringCastAndExpressions.sql.out b/sql/core/src/test/resources/sql-tests/results/typeCoercion/native/stringCastAndExpressions.sql.out index 45e2cf187a31e..bd5e33ef9f7f9 100644 --- a/sql/core/src/test/resources/sql-tests/results/typeCoercion/native/stringCastAndExpressions.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/typeCoercion/native/stringCastAndExpressions.sql.out @@ -140,8 +140,8 @@ org.apache.spark.SparkUpgradeException "errorClass" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION", "errorSubClass" : "DATETIME_PATTERN_RECOGNITION", "messageParameters" : { - "pattern" : "'aa'", - "config" : "\"spark.sql.legacy.timeParserPolicy\"" + "config" : "\"spark.sql.legacy.timeParserPolicy\"", + "pattern" : "'aa'" } } @@ -164,8 +164,8 @@ org.apache.spark.SparkUpgradeException "errorClass" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION", "errorSubClass" : "DATETIME_PATTERN_RECOGNITION", "messageParameters" : { - "pattern" : "'aa'", - "config" : "\"spark.sql.legacy.timeParserPolicy\"" + "config" : "\"spark.sql.legacy.timeParserPolicy\"", + "pattern" : "'aa'" } } @@ -188,8 +188,8 @@ org.apache.spark.SparkUpgradeException "errorClass" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION", "errorSubClass" : "DATETIME_PATTERN_RECOGNITION", "messageParameters" : { - "pattern" : "'aa'", - "config" : "\"spark.sql.legacy.timeParserPolicy\"" + "config" : "\"spark.sql.legacy.timeParserPolicy\"", + "pattern" : "'aa'" } } diff --git a/sql/core/src/test/resources/sql-tests/results/udf/udf-pivot.sql.out b/sql/core/src/test/resources/sql-tests/results/udf/udf-pivot.sql.out index 1cd1db599ad06..3390b139c5b8c 100644 --- a/sql/core/src/test/resources/sql-tests/results/udf/udf-pivot.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/udf/udf-pivot.sql.out @@ -329,9 +329,9 @@ org.apache.spark.sql.AnalysisException "errorClass" : "PIVOT_VALUE_DATA_TYPE_MISMATCH", "sqlState" : "42000", "messageParameters" : { + "pivotType" : "struct", "value" : "dotNET", - "valueType" : "string", - "pivotType" : "struct" + "valueType" : "string" } } From 3a0589a3f9b7600f4ea733763d09ff7a9889f831 Mon Sep 17 00:00:00 2001 From: Max Gekk Date: Wed, 14 Sep 2022 11:19:16 +0300 Subject: [PATCH 3/3] Trigger build