diff --git a/common/utils/src/main/scala/org/apache/spark/SparkThrowableHelper.scala b/common/utils/src/main/scala/org/apache/spark/SparkThrowableHelper.scala index b6c2b176de62..47486fb57015 100644 --- a/common/utils/src/main/scala/org/apache/spark/SparkThrowableHelper.scala +++ b/common/utils/src/main/scala/org/apache/spark/SparkThrowableHelper.scala @@ -85,20 +85,20 @@ private[spark] object SparkThrowableHelper { toJsonString { generator => val g = generator.useDefaultPrettyPrinter() g.writeStartObject() - g.writeStringField("errorClass", "LEGACY") + g.writeStringField("condition", "LEGACY") g.writeObjectFieldStart("messageParameters") g.writeStringField("message", e.getMessage) g.writeEndObject() g.writeEndObject() } case MINIMAL | STANDARD => - val errorClass = e.getCondition + val condition = e.getCondition toJsonString { generator => val g = generator.useDefaultPrettyPrinter() g.writeStartObject() - g.writeStringField("errorClass", errorClass) + g.writeStringField("condition", condition) if (format == STANDARD) { - g.writeStringField("messageTemplate", errorReader.getMessageTemplate(errorClass)) + g.writeStringField("messageTemplate", errorReader.getMessageTemplate(condition)) } val sqlState = e.getSqlState if (sqlState != null) g.writeStringField("sqlState", sqlState) diff --git a/core/src/test/scala/org/apache/spark/SparkThrowableSuite.scala b/core/src/test/scala/org/apache/spark/SparkThrowableSuite.scala index ea845c0f93a4..1dea1500136c 100644 --- a/core/src/test/scala/org/apache/spark/SparkThrowableSuite.scala +++ b/core/src/test/scala/org/apache/spark/SparkThrowableSuite.scala @@ -314,7 +314,7 @@ class SparkThrowableSuite extends SparkFunSuite { // scalastyle:off line.size.limit assert(SparkThrowableHelper.getMessage(e, MINIMAL) === """{ - | "errorClass" : "DIVIDE_BY_ZERO", + | "condition" : "DIVIDE_BY_ZERO", | "sqlState" : "22012", | "messageParameters" : { | "config" : "CONFIG" @@ -328,7 +328,7 @@ class SparkThrowableSuite extends SparkFunSuite { |}""".stripMargin) assert(SparkThrowableHelper.getMessage(e, STANDARD) === """{ - | "errorClass" : "DIVIDE_BY_ZERO", + | "condition" : "DIVIDE_BY_ZERO", | "messageTemplate" : "Division by zero. Use `try_divide` to tolerate divisor being 0 and return NULL instead. If necessary set to \"false\" to bypass this error.", | "sqlState" : "22012", | "messageParameters" : { @@ -348,7 +348,7 @@ class SparkThrowableSuite extends SparkFunSuite { messageParameters = Map("saveMode" -> "UNSUPPORTED_MODE")) assert(SparkThrowableHelper.getMessage(e2, STANDARD) === """{ - | "errorClass" : "UNSUPPORTED_SAVE_MODE.EXISTENT_PATH", + | "condition" : "UNSUPPORTED_SAVE_MODE.EXISTENT_PATH", | "messageTemplate" : "The save mode is not supported for: an existent path.", | "sqlState" : "0A000", | "messageParameters" : { @@ -363,7 +363,7 @@ class SparkThrowableSuite extends SparkFunSuite { val e3 = new LegacyException assert(SparkThrowableHelper.getMessage(e3, MINIMAL) === """{ - | "errorClass" : "LEGACY", + | "condition" : "LEGACY", | "messageParameters" : { | "message" : "Test message" | } @@ -392,7 +392,7 @@ class SparkThrowableSuite extends SparkFunSuite { // scalastyle:off line.size.limit assert(SparkThrowableHelper.getMessage(e4, MINIMAL) === """{ - | "errorClass" : "DIVIDE_BY_ZERO", + | "condition" : "DIVIDE_BY_ZERO", | "sqlState" : "22012", | "messageParameters" : { | "config" : "CONFIG" @@ -404,7 +404,7 @@ class SparkThrowableSuite extends SparkFunSuite { |}""".stripMargin) assert(SparkThrowableHelper.getMessage(e4, STANDARD) === """{ - | "errorClass" : "DIVIDE_BY_ZERO", + | "condition" : "DIVIDE_BY_ZERO", | "messageTemplate" : "Division by zero. Use `try_divide` to tolerate divisor being 0 and return NULL instead. If necessary set to \"false\" to bypass this error.", | "sqlState" : "22012", | "messageParameters" : { diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/array.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/array.sql.out index 53595d1b8a3e..7e6202b92bfa 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/array.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/array.sql.out @@ -172,7 +172,7 @@ select sort_array(array('b', 'd'), '1') -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"1\"", @@ -338,7 +338,7 @@ select array_size(map('a', 1, 'b', 2)) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"map(a, 1, b, 2)\"", @@ -460,7 +460,7 @@ select array_insert(array(1, 2, 3), 3, "4") -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.ARRAY_FUNCTION_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.ARRAY_FUNCTION_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "\"ARRAY\"", @@ -552,7 +552,7 @@ select array_compact(id) from values (1) as t(id) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"id\"", diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/bitwise.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/bitwise.sql.out index 1267a984565a..568ecf179097 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/bitwise.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/bitwise.sql.out @@ -130,7 +130,7 @@ select bit_count("bit count") -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"bit count\"", @@ -154,7 +154,7 @@ select bit_count('a') -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"a\"", @@ -334,7 +334,7 @@ SELECT 20181117 > > 2 -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'>'", @@ -348,7 +348,7 @@ SELECT 20181117 < < 2 -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'<'", @@ -362,7 +362,7 @@ SELECT 20181117 > >> 2 -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'>>'", @@ -376,7 +376,7 @@ SELECT 20181117 <<< 2 -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'<'", @@ -390,7 +390,7 @@ SELECT 20181117 >>>> 2 -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'>'", diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/cast.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/cast.sql.out index 643dfd3771ff..50caf59ff561 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/cast.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/cast.sql.out @@ -207,7 +207,7 @@ SELECT HEX(CAST(CAST(123 AS byte) AS binary)) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", + "condition" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", "sqlState" : "42K09", "messageParameters" : { "config" : "\"spark.sql.ansi.enabled\"", @@ -231,7 +231,7 @@ SELECT HEX(CAST(CAST(-123 AS byte) AS binary)) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", + "condition" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", "sqlState" : "42K09", "messageParameters" : { "config" : "\"spark.sql.ansi.enabled\"", @@ -255,7 +255,7 @@ SELECT HEX(CAST(123S AS binary)) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", + "condition" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", "sqlState" : "42K09", "messageParameters" : { "config" : "\"spark.sql.ansi.enabled\"", @@ -279,7 +279,7 @@ SELECT HEX(CAST(-123S AS binary)) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", + "condition" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", "sqlState" : "42K09", "messageParameters" : { "config" : "\"spark.sql.ansi.enabled\"", @@ -303,7 +303,7 @@ SELECT HEX(CAST(123 AS binary)) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", + "condition" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", "sqlState" : "42K09", "messageParameters" : { "config" : "\"spark.sql.ansi.enabled\"", @@ -327,7 +327,7 @@ SELECT HEX(CAST(-123 AS binary)) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", + "condition" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", "sqlState" : "42K09", "messageParameters" : { "config" : "\"spark.sql.ansi.enabled\"", @@ -351,7 +351,7 @@ SELECT HEX(CAST(123L AS binary)) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", + "condition" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", "sqlState" : "42K09", "messageParameters" : { "config" : "\"spark.sql.ansi.enabled\"", @@ -375,7 +375,7 @@ SELECT HEX(CAST(-123L AS binary)) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", + "condition" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", "sqlState" : "42K09", "messageParameters" : { "config" : "\"spark.sql.ansi.enabled\"", @@ -432,7 +432,7 @@ SELECT CAST(interval 3 month 1 hour AS string) -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0029", + "condition" : "_LEGACY_ERROR_TEMP_0029", "messageParameters" : { "literal" : "interval 3 month 1 hour" }, @@ -942,7 +942,7 @@ SELECT HEX((123 :: byte) :: binary) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", + "condition" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", "sqlState" : "42K09", "messageParameters" : { "config" : "\"spark.sql.ansi.enabled\"", @@ -1066,7 +1066,7 @@ SELECT '2147483648' :: BINT -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "UNSUPPORTED_DATATYPE", + "condition" : "UNSUPPORTED_DATATYPE", "sqlState" : "0A000", "messageParameters" : { "typeName" : "\"BINT\"" @@ -1086,7 +1086,7 @@ SELECT '2147483648' :: SELECT -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "UNSUPPORTED_DATATYPE", + "condition" : "UNSUPPORTED_DATATYPE", "sqlState" : "0A000", "messageParameters" : { "typeName" : "\"SELECT\"" @@ -1106,7 +1106,7 @@ SELECT FALSE IS NOT NULL :: string -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'::'", diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/ceil-floor-with-scale-param.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/ceil-floor-with-scale-param.sql.out index e4087d5108ed..d769f6771147 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/ceil-floor-with-scale-param.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/ceil-floor-with-scale-param.sql.out @@ -81,7 +81,7 @@ SELECT CEIL(2.5, null) -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "NON_FOLDABLE_ARGUMENT", + "condition" : "NON_FOLDABLE_ARGUMENT", "sqlState" : "42K08", "messageParameters" : { "funcName" : "`ceil`", @@ -103,7 +103,7 @@ SELECT CEIL(2.5, 'a') -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "NON_FOLDABLE_ARGUMENT", + "condition" : "NON_FOLDABLE_ARGUMENT", "sqlState" : "42K08", "messageParameters" : { "funcName" : "`ceil`", @@ -125,7 +125,7 @@ SELECT CEIL(2.5, 0, 0) -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", + "condition" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", "sqlState" : "42605", "messageParameters" : { "actualNum" : "3", @@ -225,7 +225,7 @@ SELECT FLOOR(2.5, null) -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "NON_FOLDABLE_ARGUMENT", + "condition" : "NON_FOLDABLE_ARGUMENT", "sqlState" : "42K08", "messageParameters" : { "funcName" : "`floor`", @@ -247,7 +247,7 @@ SELECT FLOOR(2.5, 'a') -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "NON_FOLDABLE_ARGUMENT", + "condition" : "NON_FOLDABLE_ARGUMENT", "sqlState" : "42K08", "messageParameters" : { "funcName" : "`floor`", @@ -269,7 +269,7 @@ SELECT FLOOR(2.5, 0, 0) -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", + "condition" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", "sqlState" : "42605", "messageParameters" : { "actualNum" : "3", diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/change-column.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/change-column.sql.out index 64ae845550f2..98060de3ac58 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/change-column.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/change-column.sql.out @@ -16,7 +16,7 @@ ALTER TABLE test_change CHANGE a -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0035", + "condition" : "_LEGACY_ERROR_TEMP_0035", "messageParameters" : { "message" : "ALTER TABLE table CHANGE COLUMN requires a TYPE, a SET/DROP, a COMMENT, or a FIRST/AFTER" }, @@ -41,7 +41,7 @@ ALTER TABLE test_change RENAME COLUMN a TO a1 -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "UNSUPPORTED_FEATURE.TABLE_OPERATION", + "condition" : "UNSUPPORTED_FEATURE.TABLE_OPERATION", "sqlState" : "0A000", "messageParameters" : { "operation" : "RENAME COLUMN", @@ -61,7 +61,7 @@ ALTER TABLE test_change CHANGE a TYPE STRING -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "NOT_SUPPORTED_CHANGE_COLUMN", + "condition" : "NOT_SUPPORTED_CHANGE_COLUMN", "sqlState" : "0A000", "messageParameters" : { "newName" : "`a`", @@ -91,7 +91,7 @@ ALTER TABLE test_change CHANGE a AFTER b -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "UNSUPPORTED_FEATURE.TABLE_OPERATION", + "condition" : "UNSUPPORTED_FEATURE.TABLE_OPERATION", "sqlState" : "0A000", "messageParameters" : { "operation" : "ALTER COLUMN ... FIRST | AFTER", @@ -105,7 +105,7 @@ ALTER TABLE test_change CHANGE b FIRST -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "UNSUPPORTED_FEATURE.TABLE_OPERATION", + "condition" : "UNSUPPORTED_FEATURE.TABLE_OPERATION", "sqlState" : "0A000", "messageParameters" : { "operation" : "ALTER COLUMN ... FIRST | AFTER", @@ -167,7 +167,7 @@ ALTER TABLE test_change CHANGE invalid_col TYPE INT -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`invalid_col`", @@ -214,7 +214,7 @@ ALTER TABLE temp_view CHANGE a TYPE INT -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "EXPECT_TABLE_NOT_VIEW.NO_ALTERNATIVE", + "condition" : "EXPECT_TABLE_NOT_VIEW.NO_ALTERNATIVE", "sqlState" : "42809", "messageParameters" : { "operation" : "ALTER TABLE ... CHANGE COLUMN", @@ -243,7 +243,7 @@ ALTER TABLE global_temp.global_temp_view CHANGE a TYPE INT -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "EXPECT_TABLE_NOT_VIEW.NO_ALTERNATIVE", + "condition" : "EXPECT_TABLE_NOT_VIEW.NO_ALTERNATIVE", "sqlState" : "42809", "messageParameters" : { "operation" : "ALTER TABLE ... CHANGE COLUMN", diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/charvarchar.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/charvarchar.sql.out index 524797015a2f..3ab1505d8f2a 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/charvarchar.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/charvarchar.sql.out @@ -82,7 +82,7 @@ create view char_view as select * from char_tbl -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "TABLE_OR_VIEW_ALREADY_EXISTS", + "condition" : "TABLE_OR_VIEW_ALREADY_EXISTS", "sqlState" : "42P07", "messageParameters" : { "relationName" : "`spark_catalog`.`default`.`char_view`" @@ -125,7 +125,7 @@ alter table char_tbl1 change column c type char(6) -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "NOT_SUPPORTED_CHANGE_COLUMN", + "condition" : "NOT_SUPPORTED_CHANGE_COLUMN", "sqlState" : "0A000", "messageParameters" : { "newName" : "`c`", @@ -249,7 +249,7 @@ create table char_part(c1 char(5), c2 char(2), v1 varchar(6), v2 varchar(2)) usi -- !query analysis org.apache.spark.sql.catalyst.analysis.TableAlreadyExistsException { - "errorClass" : "TABLE_OR_VIEW_ALREADY_EXISTS", + "condition" : "TABLE_OR_VIEW_ALREADY_EXISTS", "sqlState" : "42P07", "messageParameters" : { "relationName" : "`spark_catalog`.`default`.`char_part`" @@ -280,7 +280,7 @@ alter table char_part add partition (v2='ke', c2='nt') location 'loc1' -- !query analysis org.apache.spark.sql.catalyst.analysis.PartitionsAlreadyExistException { - "errorClass" : "PARTITIONS_ALREADY_EXIST", + "condition" : "PARTITIONS_ALREADY_EXIST", "sqlState" : "428FT", "messageParameters" : { "partitionList" : "PARTITION (`v2` = ke, `c2` = nt)", @@ -300,7 +300,7 @@ alter table char_part partition (v2='ke') rename to partition (v2='nt') -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "_LEGACY_ERROR_TEMP_1232", + "condition" : "_LEGACY_ERROR_TEMP_1232", "messageParameters" : { "partitionColumnNames" : "v2, c2", "specKeys" : "v2", diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/collations.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/collations.sql.out index 7a4777c34fed..ca51aa50ac1b 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/collations.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/collations.sql.out @@ -211,7 +211,7 @@ select col1 collate utf8_lcase from values ('aaa'), ('AAA'), ('bbb'), ('BBB'), ( -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -236,7 +236,7 @@ select col1 collate utf8_lcase from values ('aaa'), ('AAA'), ('bbb'), ('BBB'), ( -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -261,7 +261,7 @@ select col1 collate utf8_lcase from values ('aaa'), ('AAA'), ('bbb'), ('BBB'), ( -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -286,7 +286,7 @@ select col1 collate utf8_lcase from values ('aaa'), ('AAA'), ('bbb'), ('BBB'), ( -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -311,7 +311,7 @@ select col1 collate utf8_lcase from values ('aaa'), ('bbb'), ('BBB'), ('zzz'), ( -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -535,7 +535,7 @@ select str_to_map(text, pairDelim, keyValueDelim) from t4 -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "COLLATION_MISMATCH.IMPLICIT", + "condition" : "COLLATION_MISMATCH.IMPLICIT", "sqlState" : "42P21", "messageParameters" : { "implicitTypes" : "\"STRING\", \"STRING COLLATE UTF8_LCASE\"" @@ -548,7 +548,7 @@ select str_to_map(text collate utf8_binary, pairDelim collate utf8_lcase, keyVal -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "COLLATION_MISMATCH.EXPLICIT", + "condition" : "COLLATION_MISMATCH.EXPLICIT", "sqlState" : "42P21", "messageParameters" : { "explicitTypes" : "\"STRING\", \"STRING COLLATE UTF8_LCASE\"" @@ -569,7 +569,7 @@ select str_to_map(text collate unicode_ai, pairDelim collate unicode_ai, keyValu -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"collate(text, unicode_ai)\"", @@ -836,7 +836,7 @@ select concat_ws(' ', utf8_binary, utf8_lcase) from t5 -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "COLLATION_MISMATCH.IMPLICIT", + "condition" : "COLLATION_MISMATCH.IMPLICIT", "sqlState" : "42P21", "messageParameters" : { "implicitTypes" : "\"STRING\", \"STRING COLLATE UTF8_LCASE\"" @@ -849,7 +849,7 @@ select concat_ws(' ' collate utf8_binary, utf8_binary, 'SQL' collate utf8_lcase) -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "COLLATION_MISMATCH.EXPLICIT", + "condition" : "COLLATION_MISMATCH.EXPLICIT", "sqlState" : "42P21", "messageParameters" : { "explicitTypes" : "\"STRING\", \"STRING COLLATE UTF8_LCASE\"" @@ -894,7 +894,7 @@ select elt(2, utf8_binary, utf8_lcase, s) from t5 -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "COLLATION_MISMATCH.IMPLICIT", + "condition" : "COLLATION_MISMATCH.IMPLICIT", "sqlState" : "42P21", "messageParameters" : { "implicitTypes" : "\"STRING\", \"STRING COLLATE UTF8_LCASE\"" @@ -907,7 +907,7 @@ select elt(1, utf8_binary collate utf8_binary, utf8_lcase collate utf8_lcase) fr -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "COLLATION_MISMATCH.EXPLICIT", + "condition" : "COLLATION_MISMATCH.EXPLICIT", "sqlState" : "42P21", "messageParameters" : { "explicitTypes" : "\"STRING\", \"STRING COLLATE UTF8_LCASE\"" @@ -952,7 +952,7 @@ select split_part(utf8_binary, utf8_lcase, 3) from t5 -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "COLLATION_MISMATCH.IMPLICIT", + "condition" : "COLLATION_MISMATCH.IMPLICIT", "sqlState" : "42P21", "messageParameters" : { "implicitTypes" : "\"STRING\", \"STRING COLLATE UTF8_LCASE\"" @@ -973,7 +973,7 @@ select split_part(utf8_binary collate utf8_binary, s collate utf8_lcase, 1) from -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "COLLATION_MISMATCH.EXPLICIT", + "condition" : "COLLATION_MISMATCH.EXPLICIT", "sqlState" : "42P21", "messageParameters" : { "explicitTypes" : "\"STRING\", \"STRING COLLATE UTF8_LCASE\"" @@ -1002,7 +1002,7 @@ select split_part(utf8_binary collate unicode_ai, utf8_lcase collate unicode_ai, -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"collate(utf8_binary, unicode_ai)\"", @@ -1050,7 +1050,7 @@ select contains(utf8_binary, utf8_lcase) from t5 -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "COLLATION_MISMATCH.IMPLICIT", + "condition" : "COLLATION_MISMATCH.IMPLICIT", "sqlState" : "42P21", "messageParameters" : { "implicitTypes" : "\"STRING\", \"STRING COLLATE UTF8_LCASE\"" @@ -1071,7 +1071,7 @@ select contains(utf8_binary collate utf8_binary, s collate utf8_lcase) from t5 -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "COLLATION_MISMATCH.EXPLICIT", + "condition" : "COLLATION_MISMATCH.EXPLICIT", "sqlState" : "42P21", "messageParameters" : { "explicitTypes" : "\"STRING\", \"STRING COLLATE UTF8_LCASE\"" @@ -1100,7 +1100,7 @@ select contains(utf8_binary collate unicode_ai, utf8_lcase collate unicode_ai) f -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"collate(utf8_binary, unicode_ai)\"", @@ -1148,7 +1148,7 @@ select substring_index(utf8_binary, utf8_lcase, 2) from t5 -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "COLLATION_MISMATCH.IMPLICIT", + "condition" : "COLLATION_MISMATCH.IMPLICIT", "sqlState" : "42P21", "messageParameters" : { "implicitTypes" : "\"STRING\", \"STRING COLLATE UTF8_LCASE\"" @@ -1169,7 +1169,7 @@ select substring_index(utf8_binary collate utf8_binary, s collate utf8_lcase, 3) -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "COLLATION_MISMATCH.EXPLICIT", + "condition" : "COLLATION_MISMATCH.EXPLICIT", "sqlState" : "42P21", "messageParameters" : { "explicitTypes" : "\"STRING\", \"STRING COLLATE UTF8_LCASE\"" @@ -1198,7 +1198,7 @@ select substring_index(utf8_binary collate unicode_ai, utf8_lcase collate unicod -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"collate(utf8_binary, unicode_ai)\"", @@ -1246,7 +1246,7 @@ select instr(utf8_binary, utf8_lcase) from t5 -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "COLLATION_MISMATCH.IMPLICIT", + "condition" : "COLLATION_MISMATCH.IMPLICIT", "sqlState" : "42P21", "messageParameters" : { "implicitTypes" : "\"STRING\", \"STRING COLLATE UTF8_LCASE\"" @@ -1267,7 +1267,7 @@ select instr(utf8_binary collate utf8_binary, s collate utf8_lcase) from t5 -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "COLLATION_MISMATCH.EXPLICIT", + "condition" : "COLLATION_MISMATCH.EXPLICIT", "sqlState" : "42P21", "messageParameters" : { "explicitTypes" : "\"STRING\", \"STRING COLLATE UTF8_LCASE\"" @@ -1296,7 +1296,7 @@ select instr(utf8_binary collate unicode_ai, utf8_lcase collate unicode_ai) from -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"collate(utf8_binary, unicode_ai)\"", @@ -1336,7 +1336,7 @@ select find_in_set(utf8_binary, utf8_lcase) from t5 -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "COLLATION_MISMATCH.IMPLICIT", + "condition" : "COLLATION_MISMATCH.IMPLICIT", "sqlState" : "42P21", "messageParameters" : { "implicitTypes" : "\"STRING\", \"STRING COLLATE UTF8_LCASE\"" @@ -1357,7 +1357,7 @@ select find_in_set(utf8_binary collate utf8_binary, s collate utf8_lcase) from t -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "COLLATION_MISMATCH.EXPLICIT", + "condition" : "COLLATION_MISMATCH.EXPLICIT", "sqlState" : "42P21", "messageParameters" : { "explicitTypes" : "\"STRING\", \"STRING COLLATE UTF8_LCASE\"" @@ -1410,7 +1410,7 @@ select startswith(utf8_binary, utf8_lcase) from t5 -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "COLLATION_MISMATCH.IMPLICIT", + "condition" : "COLLATION_MISMATCH.IMPLICIT", "sqlState" : "42P21", "messageParameters" : { "implicitTypes" : "\"STRING\", \"STRING COLLATE UTF8_LCASE\"" @@ -1431,7 +1431,7 @@ select startswith(utf8_binary collate utf8_binary, s collate utf8_lcase) from t5 -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "COLLATION_MISMATCH.EXPLICIT", + "condition" : "COLLATION_MISMATCH.EXPLICIT", "sqlState" : "42P21", "messageParameters" : { "explicitTypes" : "\"STRING\", \"STRING COLLATE UTF8_LCASE\"" @@ -1460,7 +1460,7 @@ select startswith(utf8_binary collate unicode_ai, utf8_lcase collate unicode_ai) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"collate(utf8_binary, unicode_ai)\"", @@ -1516,7 +1516,7 @@ select translate(utf8_binary, utf8_lcase, '12345') from t5 -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "COLLATION_MISMATCH.IMPLICIT", + "condition" : "COLLATION_MISMATCH.IMPLICIT", "sqlState" : "42P21", "messageParameters" : { "implicitTypes" : "\"STRING\", \"STRING COLLATE UTF8_LCASE\"" @@ -1529,7 +1529,7 @@ select translate(utf8_binary, 'aBc' collate utf8_lcase, '12345' collate utf8_bin -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "COLLATION_MISMATCH.EXPLICIT", + "condition" : "COLLATION_MISMATCH.EXPLICIT", "sqlState" : "42P21", "messageParameters" : { "explicitTypes" : "\"STRING COLLATE UTF8_LCASE\", \"STRING\"" @@ -1550,7 +1550,7 @@ select translate(utf8_binary, 'SQL' collate unicode_ai, '12345' collate unicode_ -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"utf8_binary\"", @@ -1598,7 +1598,7 @@ select replace(utf8_binary, utf8_lcase, 'abc') from t5 -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "COLLATION_MISMATCH.IMPLICIT", + "condition" : "COLLATION_MISMATCH.IMPLICIT", "sqlState" : "42P21", "messageParameters" : { "implicitTypes" : "\"STRING\", \"STRING COLLATE UTF8_LCASE\"" @@ -1619,7 +1619,7 @@ select replace(utf8_binary collate utf8_binary, s collate utf8_lcase, 'abc') fro -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "COLLATION_MISMATCH.EXPLICIT", + "condition" : "COLLATION_MISMATCH.EXPLICIT", "sqlState" : "42P21", "messageParameters" : { "explicitTypes" : "\"STRING\", \"STRING COLLATE UTF8_LCASE\"" @@ -1648,7 +1648,7 @@ select replace(utf8_binary collate unicode_ai, utf8_lcase collate unicode_ai, 'a -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"collate(utf8_binary, unicode_ai)\"", @@ -1696,7 +1696,7 @@ select endswith(utf8_binary, utf8_lcase) from t5 -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "COLLATION_MISMATCH.IMPLICIT", + "condition" : "COLLATION_MISMATCH.IMPLICIT", "sqlState" : "42P21", "messageParameters" : { "implicitTypes" : "\"STRING\", \"STRING COLLATE UTF8_LCASE\"" @@ -1717,7 +1717,7 @@ select endswith(utf8_binary collate utf8_binary, s collate utf8_lcase) from t5 -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "COLLATION_MISMATCH.EXPLICIT", + "condition" : "COLLATION_MISMATCH.EXPLICIT", "sqlState" : "42P21", "messageParameters" : { "explicitTypes" : "\"STRING\", \"STRING COLLATE UTF8_LCASE\"" @@ -1746,7 +1746,7 @@ select endswith(utf8_binary collate unicode_ai, utf8_lcase collate unicode_ai) f -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"collate(utf8_binary, unicode_ai)\"", @@ -1994,7 +1994,7 @@ select overlay(utf8_binary, utf8_lcase, 2) from t5 -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "COLLATION_MISMATCH.IMPLICIT", + "condition" : "COLLATION_MISMATCH.IMPLICIT", "sqlState" : "42P21", "messageParameters" : { "implicitTypes" : "\"STRING\", \"STRING COLLATE UTF8_LCASE\"" @@ -2015,7 +2015,7 @@ select overlay(utf8_binary collate utf8_binary, s collate utf8_lcase, 3) from t5 -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "COLLATION_MISMATCH.EXPLICIT", + "condition" : "COLLATION_MISMATCH.EXPLICIT", "sqlState" : "42P21", "messageParameters" : { "explicitTypes" : "\"STRING\", \"STRING COLLATE UTF8_LCASE\"" @@ -2164,7 +2164,7 @@ select levenshtein(utf8_binary, utf8_lcase) from t5 -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "COLLATION_MISMATCH.IMPLICIT", + "condition" : "COLLATION_MISMATCH.IMPLICIT", "sqlState" : "42P21", "messageParameters" : { "implicitTypes" : "\"STRING\", \"STRING COLLATE UTF8_LCASE\"" @@ -2185,7 +2185,7 @@ select levenshtein(utf8_binary collate utf8_binary, s collate utf8_lcase) from t -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "COLLATION_MISMATCH.EXPLICIT", + "condition" : "COLLATION_MISMATCH.EXPLICIT", "sqlState" : "42P21", "messageParameters" : { "explicitTypes" : "\"STRING\", \"STRING COLLATE UTF8_LCASE\"" @@ -2374,7 +2374,7 @@ select rpad(utf8_binary, 8, utf8_lcase) from t5 -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "COLLATION_MISMATCH.IMPLICIT", + "condition" : "COLLATION_MISMATCH.IMPLICIT", "sqlState" : "42P21", "messageParameters" : { "implicitTypes" : "\"STRING\", \"STRING COLLATE UTF8_LCASE\"" @@ -2395,7 +2395,7 @@ select rpad(utf8_binary collate utf8_binary, 8, s collate utf8_lcase) from t5 -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "COLLATION_MISMATCH.EXPLICIT", + "condition" : "COLLATION_MISMATCH.EXPLICIT", "sqlState" : "42P21", "messageParameters" : { "explicitTypes" : "\"STRING\", \"STRING COLLATE UTF8_LCASE\"" @@ -2448,7 +2448,7 @@ select lpad(utf8_binary, 8, utf8_lcase) from t5 -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "COLLATION_MISMATCH.IMPLICIT", + "condition" : "COLLATION_MISMATCH.IMPLICIT", "sqlState" : "42P21", "messageParameters" : { "implicitTypes" : "\"STRING\", \"STRING COLLATE UTF8_LCASE\"" @@ -2469,7 +2469,7 @@ select lpad(utf8_binary collate utf8_binary, 8, s collate utf8_lcase) from t5 -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "COLLATION_MISMATCH.EXPLICIT", + "condition" : "COLLATION_MISMATCH.EXPLICIT", "sqlState" : "42P21", "messageParameters" : { "explicitTypes" : "\"STRING\", \"STRING COLLATE UTF8_LCASE\"" @@ -2522,7 +2522,7 @@ select locate(utf8_binary, utf8_lcase) from t5 -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "COLLATION_MISMATCH.IMPLICIT", + "condition" : "COLLATION_MISMATCH.IMPLICIT", "sqlState" : "42P21", "messageParameters" : { "implicitTypes" : "\"STRING\", \"STRING COLLATE UTF8_LCASE\"" @@ -2543,7 +2543,7 @@ select locate(utf8_binary collate utf8_binary, s collate utf8_lcase) from t5 -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "COLLATION_MISMATCH.EXPLICIT", + "condition" : "COLLATION_MISMATCH.EXPLICIT", "sqlState" : "42P21", "messageParameters" : { "explicitTypes" : "\"STRING\", \"STRING COLLATE UTF8_LCASE\"" @@ -2572,7 +2572,7 @@ select locate(utf8_binary collate unicode_ai, utf8_lcase collate unicode_ai, 3) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"collate(utf8_binary, unicode_ai)\"", @@ -2620,7 +2620,7 @@ select TRIM(utf8_binary, utf8_lcase) from t5 -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "COLLATION_MISMATCH.IMPLICIT", + "condition" : "COLLATION_MISMATCH.IMPLICIT", "sqlState" : "42P21", "messageParameters" : { "implicitTypes" : "\"STRING COLLATE UTF8_LCASE\", \"STRING\"" @@ -2641,7 +2641,7 @@ select TRIM(utf8_binary collate utf8_binary, s collate utf8_lcase) from t5 -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "COLLATION_MISMATCH.EXPLICIT", + "condition" : "COLLATION_MISMATCH.EXPLICIT", "sqlState" : "42P21", "messageParameters" : { "explicitTypes" : "\"STRING COLLATE UTF8_LCASE\", \"STRING\"" @@ -2670,7 +2670,7 @@ select TRIM(utf8_binary collate unicode_ai, utf8_lcase collate unicode_ai) from -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"collate(utf8_lcase, unicode_ai)\"", @@ -2718,7 +2718,7 @@ select BTRIM(utf8_binary, utf8_lcase) from t5 -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "COLLATION_MISMATCH.IMPLICIT", + "condition" : "COLLATION_MISMATCH.IMPLICIT", "sqlState" : "42P21", "messageParameters" : { "implicitTypes" : "\"STRING\", \"STRING COLLATE UTF8_LCASE\"" @@ -2739,7 +2739,7 @@ select BTRIM(utf8_binary collate utf8_binary, s collate utf8_lcase) from t5 -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "COLLATION_MISMATCH.EXPLICIT", + "condition" : "COLLATION_MISMATCH.EXPLICIT", "sqlState" : "42P21", "messageParameters" : { "explicitTypes" : "\"STRING\", \"STRING COLLATE UTF8_LCASE\"" @@ -2768,7 +2768,7 @@ select BTRIM(utf8_binary collate unicode_ai, utf8_lcase collate unicode_ai) from -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"collate(utf8_binary, unicode_ai)\"", @@ -2816,7 +2816,7 @@ select LTRIM(utf8_binary, utf8_lcase) from t5 -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "COLLATION_MISMATCH.IMPLICIT", + "condition" : "COLLATION_MISMATCH.IMPLICIT", "sqlState" : "42P21", "messageParameters" : { "implicitTypes" : "\"STRING COLLATE UTF8_LCASE\", \"STRING\"" @@ -2837,7 +2837,7 @@ select LTRIM(utf8_binary collate utf8_binary, s collate utf8_lcase) from t5 -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "COLLATION_MISMATCH.EXPLICIT", + "condition" : "COLLATION_MISMATCH.EXPLICIT", "sqlState" : "42P21", "messageParameters" : { "explicitTypes" : "\"STRING COLLATE UTF8_LCASE\", \"STRING\"" @@ -2866,7 +2866,7 @@ select LTRIM(utf8_binary collate unicode_ai, utf8_lcase collate unicode_ai) from -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"collate(utf8_lcase, unicode_ai)\"", @@ -2914,7 +2914,7 @@ select RTRIM(utf8_binary, utf8_lcase) from t5 -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "COLLATION_MISMATCH.IMPLICIT", + "condition" : "COLLATION_MISMATCH.IMPLICIT", "sqlState" : "42P21", "messageParameters" : { "implicitTypes" : "\"STRING COLLATE UTF8_LCASE\", \"STRING\"" @@ -2935,7 +2935,7 @@ select RTRIM(utf8_binary collate utf8_binary, s collate utf8_lcase) from t5 -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "COLLATION_MISMATCH.EXPLICIT", + "condition" : "COLLATION_MISMATCH.EXPLICIT", "sqlState" : "42P21", "messageParameters" : { "explicitTypes" : "\"STRING COLLATE UTF8_LCASE\", \"STRING\"" @@ -2964,7 +2964,7 @@ select RTRIM(utf8_binary collate unicode_ai, utf8_lcase collate unicode_ai) from -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"collate(utf8_lcase, unicode_ai)\"", diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/column-resolution-aggregate.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/column-resolution-aggregate.sql.out index b3bfec1fe3a8..ba5a9c635142 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/column-resolution-aggregate.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/column-resolution-aggregate.sql.out @@ -20,7 +20,7 @@ SELECT max(a) AS b, b FROM v1 GROUP BY k -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "MISSING_AGGREGATION", + "condition" : "MISSING_AGGREGATION", "sqlState" : "42803", "messageParameters" : { "expression" : "\"b\"", @@ -54,7 +54,7 @@ SELECT a AS k FROM v1 GROUP BY k -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "MISSING_AGGREGATION", + "condition" : "MISSING_AGGREGATION", "sqlState" : "42803", "messageParameters" : { "expression" : "\"a\"", @@ -68,7 +68,7 @@ SELECT x FROM v2 GROUP BY all -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "MISSING_AGGREGATION", + "condition" : "MISSING_AGGREGATION", "sqlState" : "42803", "messageParameters" : { "expression" : "\"x\"", @@ -82,7 +82,7 @@ SELECT a AS all, b FROM v1 GROUP BY all -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "MISSING_AGGREGATION", + "condition" : "MISSING_AGGREGATION", "sqlState" : "42803", "messageParameters" : { "expression" : "\"b\"", @@ -96,7 +96,7 @@ SELECT k AS lca, lca + 1 AS col FROM v1 GROUP BY k, col -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "UNSUPPORTED_FEATURE.LATERAL_COLUMN_ALIAS_IN_GROUP_BY", + "condition" : "UNSUPPORTED_FEATURE.LATERAL_COLUMN_ALIAS_IN_GROUP_BY", "sqlState" : "0A000" } @@ -106,7 +106,7 @@ SELECT k AS lca, lca + 1 AS col FROM v1 GROUP BY all -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "UNSUPPORTED_FEATURE.LATERAL_COLUMN_ALIAS_IN_GROUP_BY", + "condition" : "UNSUPPORTED_FEATURE.LATERAL_COLUMN_ALIAS_IN_GROUP_BY", "sqlState" : "0A000" } diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/columnresolution-negative.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/columnresolution-negative.sql.out index f16d42fac722..6b29655c05cc 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/columnresolution-negative.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/columnresolution-negative.sql.out @@ -61,7 +61,7 @@ SELECT i1 FROM t1, mydb1.t1 -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "AMBIGUOUS_REFERENCE", + "condition" : "AMBIGUOUS_REFERENCE", "sqlState" : "42704", "messageParameters" : { "name" : "`i1`", @@ -82,7 +82,7 @@ SELECT t1.i1 FROM t1, mydb1.t1 -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "AMBIGUOUS_REFERENCE", + "condition" : "AMBIGUOUS_REFERENCE", "sqlState" : "42704", "messageParameters" : { "name" : "`t1`.`i1`", @@ -103,7 +103,7 @@ SELECT mydb1.t1.i1 FROM t1, mydb1.t1 -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "AMBIGUOUS_REFERENCE", + "condition" : "AMBIGUOUS_REFERENCE", "sqlState" : "42704", "messageParameters" : { "name" : "`mydb1`.`t1`.`i1`", @@ -124,7 +124,7 @@ SELECT i1 FROM t1, mydb2.t1 -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "AMBIGUOUS_REFERENCE", + "condition" : "AMBIGUOUS_REFERENCE", "sqlState" : "42704", "messageParameters" : { "name" : "`i1`", @@ -145,7 +145,7 @@ SELECT t1.i1 FROM t1, mydb2.t1 -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "AMBIGUOUS_REFERENCE", + "condition" : "AMBIGUOUS_REFERENCE", "sqlState" : "42704", "messageParameters" : { "name" : "`t1`.`i1`", @@ -173,7 +173,7 @@ SELECT i1 FROM t1, mydb1.t1 -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "AMBIGUOUS_REFERENCE", + "condition" : "AMBIGUOUS_REFERENCE", "sqlState" : "42704", "messageParameters" : { "name" : "`i1`", @@ -194,7 +194,7 @@ SELECT t1.i1 FROM t1, mydb1.t1 -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "AMBIGUOUS_REFERENCE", + "condition" : "AMBIGUOUS_REFERENCE", "sqlState" : "42704", "messageParameters" : { "name" : "`t1`.`i1`", @@ -215,7 +215,7 @@ SELECT i1 FROM t1, mydb2.t1 -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "AMBIGUOUS_REFERENCE", + "condition" : "AMBIGUOUS_REFERENCE", "sqlState" : "42704", "messageParameters" : { "name" : "`i1`", @@ -236,7 +236,7 @@ SELECT t1.i1 FROM t1, mydb2.t1 -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "AMBIGUOUS_REFERENCE", + "condition" : "AMBIGUOUS_REFERENCE", "sqlState" : "42704", "messageParameters" : { "name" : "`t1`.`i1`", @@ -257,7 +257,7 @@ SELECT db1.t1.i1 FROM t1, mydb2.t1 -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`db1`.`t1`.`i1`", @@ -291,7 +291,7 @@ SELECT mydb1.t1 FROM t1 -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`mydb1`.`t1`", @@ -312,7 +312,7 @@ SELECT t1.x.y.* FROM t1 -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "CANNOT_RESOLVE_STAR_EXPAND", + "condition" : "CANNOT_RESOLVE_STAR_EXPAND", "sqlState" : "42704", "messageParameters" : { "columns" : "`i1`", @@ -333,7 +333,7 @@ SELECT t1 FROM mydb1.t1 -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`t1`", @@ -361,7 +361,7 @@ SELECT mydb1.t1.i1 FROM t1 -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`mydb1`.`t1`.`i1`", @@ -413,7 +413,7 @@ SELECT * FROM v1 -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "INCOMPATIBLE_VIEW_SCHEMA_CHANGE", + "condition" : "INCOMPATIBLE_VIEW_SCHEMA_CHANGE", "sqlState" : "51024", "messageParameters" : { "actualCols" : "[]", @@ -461,7 +461,7 @@ SELECT * FROM v2 -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "INCOMPATIBLE_VIEW_SCHEMA_CHANGE", + "condition" : "INCOMPATIBLE_VIEW_SCHEMA_CHANGE", "sqlState" : "51024", "messageParameters" : { "actualCols" : "[]", diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/comments.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/comments.sql.out index 30df13ace66b..af081b8c69b1 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/comments.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/comments.sql.out @@ -121,7 +121,7 @@ select 1 as a -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "UNCLOSED_BRACKETED_COMMENT", + "condition" : "UNCLOSED_BRACKETED_COMMENT", "sqlState" : "42601" } @@ -140,6 +140,6 @@ select 4 as d -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "UNCLOSED_BRACKETED_COMMENT", + "condition" : "UNCLOSED_BRACKETED_COMMENT", "sqlState" : "42601" } diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/count.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/count.sql.out index 732b71461579..def204d4d69e 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/count.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/count.sql.out @@ -181,7 +181,7 @@ SELECT count() FROM testData -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "WRONG_NUM_ARGS.WITH_SUGGESTION", + "condition" : "WRONG_NUM_ARGS.WITH_SUGGESTION", "sqlState" : "42605", "messageParameters" : { "actualNum" : "0", @@ -223,7 +223,7 @@ SELECT count(testData.*) FROM testData -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "_LEGACY_ERROR_TEMP_1021", + "condition" : "_LEGACY_ERROR_TEMP_1021", "messageParameters" : { "targetString" : "testData" } diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/csv-functions.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/csv-functions.sql.out index 691864ef8b1c..9b04afa2108f 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/csv-functions.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/csv-functions.sql.out @@ -18,7 +18,7 @@ select from_csv('1', 1) -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_SCHEMA.NON_STRING_LITERAL", + "condition" : "INVALID_SCHEMA.NON_STRING_LITERAL", "sqlState" : "42K07", "messageParameters" : { "inputSchema" : "\"1\"" @@ -38,7 +38,7 @@ select from_csv('1', 'a InvalidType') -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'InvalidType'", @@ -59,7 +59,7 @@ select from_csv('1', 'Array') -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_SCHEMA.NON_STRUCT_TYPE", + "condition" : "INVALID_SCHEMA.NON_STRUCT_TYPE", "sqlState" : "42K07", "messageParameters" : { "dataType" : "\"ARRAY\"", @@ -80,7 +80,7 @@ select from_csv('1', 'a INT', named_struct('mode', 'PERMISSIVE')) -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_OPTIONS.NON_MAP_FUNCTION", + "condition" : "INVALID_OPTIONS.NON_MAP_FUNCTION", "sqlState" : "42K06", "queryContext" : [ { "objectType" : "", @@ -97,7 +97,7 @@ select from_csv('1', 'a INT', map('mode', 1)) -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_OPTIONS.NON_STRING_TYPE", + "condition" : "INVALID_OPTIONS.NON_STRING_TYPE", "sqlState" : "42K06", "messageParameters" : { "mapType" : "\"MAP\"" @@ -117,7 +117,7 @@ select from_csv() -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", + "condition" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", "sqlState" : "42605", "messageParameters" : { "actualNum" : "0", @@ -154,7 +154,7 @@ select schema_of_csv(null) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_NULL", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_NULL", "sqlState" : "42K09", "messageParameters" : { "exprName" : "csv", @@ -183,7 +183,7 @@ SELECT schema_of_csv(csvField) FROM csvTable -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.NON_FOLDABLE_INPUT", + "condition" : "DATATYPE_MISMATCH.NON_FOLDABLE_INPUT", "sqlState" : "42K09", "messageParameters" : { "inputExpr" : "\"csvField\"", @@ -226,7 +226,7 @@ select to_csv(named_struct('a', 1, 'b', 2), named_struct('mode', 'PERMISSIVE')) -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_OPTIONS.NON_MAP_FUNCTION", + "condition" : "INVALID_OPTIONS.NON_MAP_FUNCTION", "sqlState" : "42K06", "queryContext" : [ { "objectType" : "", @@ -243,7 +243,7 @@ select to_csv(named_struct('a', 1, 'b', 2), map('mode', 1)) -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_OPTIONS.NON_STRING_TYPE", + "condition" : "INVALID_OPTIONS.NON_STRING_TYPE", "sqlState" : "42K06", "messageParameters" : { "mapType" : "\"MAP\"" diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/cte-legacy.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/cte-legacy.sql.out index f9b78e94236f..84b97db63ea9 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/cte-legacy.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/cte-legacy.sql.out @@ -78,7 +78,7 @@ SELECT * FROM cte -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "TABLE_OR_VIEW_NOT_FOUND", + "condition" : "TABLE_OR_VIEW_NOT_FOUND", "sqlState" : "42P01", "messageParameters" : { "relationName" : "`cte`" @@ -353,7 +353,7 @@ SELECT * FROM t2 -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "TABLE_OR_VIEW_NOT_FOUND", + "condition" : "TABLE_OR_VIEW_NOT_FOUND", "sqlState" : "42P01", "messageParameters" : { "relationName" : "`t1`" @@ -407,7 +407,7 @@ SELECT * FROM ( -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "TABLE_OR_VIEW_NOT_FOUND", + "condition" : "TABLE_OR_VIEW_NOT_FOUND", "sqlState" : "42P01", "messageParameters" : { "relationName" : "`cte_outer`" @@ -438,7 +438,7 @@ SELECT * FROM ( -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "TABLE_OR_VIEW_NOT_FOUND", + "condition" : "TABLE_OR_VIEW_NOT_FOUND", "sqlState" : "42P01", "messageParameters" : { "relationName" : "`cte_invisible_inner`" @@ -471,7 +471,7 @@ SELECT * FROM ( -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "TABLE_OR_VIEW_NOT_FOUND", + "condition" : "TABLE_OR_VIEW_NOT_FOUND", "sqlState" : "42P01", "messageParameters" : { "relationName" : "`cte_invisible_inner`" diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/cte-nested.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/cte-nested.sql.out index 1b968d0431f4..0d39ff7ad510 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/cte-nested.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/cte-nested.sql.out @@ -103,7 +103,7 @@ SELECT * FROM cte -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "TABLE_OR_VIEW_NOT_FOUND", + "condition" : "TABLE_OR_VIEW_NOT_FOUND", "sqlState" : "42P01", "messageParameters" : { "relationName" : "`cte`" @@ -625,7 +625,7 @@ SELECT * FROM ( -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "TABLE_OR_VIEW_NOT_FOUND", + "condition" : "TABLE_OR_VIEW_NOT_FOUND", "sqlState" : "42P01", "messageParameters" : { "relationName" : "`cte_invisible_inner`" @@ -658,7 +658,7 @@ SELECT * FROM ( -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "TABLE_OR_VIEW_NOT_FOUND", + "condition" : "TABLE_OR_VIEW_NOT_FOUND", "sqlState" : "42P01", "messageParameters" : { "relationName" : "`cte_invisible_inner`" diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/cte-nonlegacy.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/cte-nonlegacy.sql.out index fcfe64097069..633352a8a3b6 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/cte-nonlegacy.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/cte-nonlegacy.sql.out @@ -103,7 +103,7 @@ SELECT * FROM cte -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "TABLE_OR_VIEW_NOT_FOUND", + "condition" : "TABLE_OR_VIEW_NOT_FOUND", "sqlState" : "42P01", "messageParameters" : { "relationName" : "`cte`" @@ -129,7 +129,7 @@ SELECT * FROM t2 -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "AMBIGUOUS_ALIAS_IN_NESTED_CTE", + "condition" : "AMBIGUOUS_ALIAS_IN_NESTED_CTE", "sqlState" : "42KD0", "messageParameters" : { "config" : "\"spark.sql.legacy.ctePrecedencePolicy\"", @@ -193,7 +193,7 @@ SELECT * FROM t2 -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "AMBIGUOUS_ALIAS_IN_NESTED_CTE", + "condition" : "AMBIGUOUS_ALIAS_IN_NESTED_CTE", "sqlState" : "42KD0", "messageParameters" : { "config" : "\"spark.sql.legacy.ctePrecedencePolicy\"", @@ -301,7 +301,7 @@ SELECT ( -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "AMBIGUOUS_ALIAS_IN_NESTED_CTE", + "condition" : "AMBIGUOUS_ALIAS_IN_NESTED_CTE", "sqlState" : "42KD0", "messageParameters" : { "config" : "\"spark.sql.legacy.ctePrecedencePolicy\"", @@ -322,7 +322,7 @@ SELECT ( -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "AMBIGUOUS_ALIAS_IN_NESTED_CTE", + "condition" : "AMBIGUOUS_ALIAS_IN_NESTED_CTE", "sqlState" : "42KD0", "messageParameters" : { "config" : "\"spark.sql.legacy.ctePrecedencePolicy\"", @@ -344,7 +344,7 @@ SELECT ( -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "AMBIGUOUS_ALIAS_IN_NESTED_CTE", + "condition" : "AMBIGUOUS_ALIAS_IN_NESTED_CTE", "sqlState" : "42KD0", "messageParameters" : { "config" : "\"spark.sql.legacy.ctePrecedencePolicy\"", @@ -364,7 +364,7 @@ WHERE c IN ( -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "AMBIGUOUS_ALIAS_IN_NESTED_CTE", + "condition" : "AMBIGUOUS_ALIAS_IN_NESTED_CTE", "sqlState" : "42KD0", "messageParameters" : { "config" : "\"spark.sql.legacy.ctePrecedencePolicy\"", @@ -413,7 +413,7 @@ SELECT * FROM t -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "AMBIGUOUS_ALIAS_IN_NESTED_CTE", + "condition" : "AMBIGUOUS_ALIAS_IN_NESTED_CTE", "sqlState" : "42KD0", "messageParameters" : { "config" : "\"spark.sql.legacy.ctePrecedencePolicy\"", @@ -432,7 +432,7 @@ SELECT ( -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "AMBIGUOUS_ALIAS_IN_NESTED_CTE", + "condition" : "AMBIGUOUS_ALIAS_IN_NESTED_CTE", "sqlState" : "42KD0", "messageParameters" : { "config" : "\"spark.sql.legacy.ctePrecedencePolicy\"", @@ -557,7 +557,7 @@ SELECT * FROM ( -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "TABLE_OR_VIEW_NOT_FOUND", + "condition" : "TABLE_OR_VIEW_NOT_FOUND", "sqlState" : "42P01", "messageParameters" : { "relationName" : "`cte_invisible_inner`" @@ -590,7 +590,7 @@ SELECT * FROM ( -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "TABLE_OR_VIEW_NOT_FOUND", + "condition" : "TABLE_OR_VIEW_NOT_FOUND", "sqlState" : "42P01", "messageParameters" : { "relationName" : "`cte_invisible_inner`" diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/cte.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/cte.sql.out index cdd3698ce9af..ded612ec8f8b 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/cte.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/cte.sql.out @@ -35,7 +35,7 @@ WITH s AS (SELECT 1 FROM s) SELECT * FROM s -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "TABLE_OR_VIEW_NOT_FOUND", + "condition" : "TABLE_OR_VIEW_NOT_FOUND", "sqlState" : "42P01", "messageParameters" : { "relationName" : "`s`" @@ -56,7 +56,7 @@ SELECT * FROM r -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "TABLE_OR_VIEW_NOT_FOUND", + "condition" : "TABLE_OR_VIEW_NOT_FOUND", "sqlState" : "42P01", "messageParameters" : { "relationName" : "`r`" @@ -115,7 +115,7 @@ WITH s1 AS (SELECT 1 FROM s2), s2 AS (SELECT 1 FROM s1) SELECT * FROM s1, s2 -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "TABLE_OR_VIEW_NOT_FOUND", + "condition" : "TABLE_OR_VIEW_NOT_FOUND", "sqlState" : "42P01", "messageParameters" : { "relationName" : "`s2`" @@ -250,7 +250,7 @@ SELECT * FROM t -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "')'", @@ -267,7 +267,7 @@ SELECT * FROM t -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "DUPLICATED_CTE_NAMES", + "condition" : "DUPLICATED_CTE_NAMES", "sqlState" : "42602", "messageParameters" : { "duplicateNames" : "`t`" @@ -288,7 +288,7 @@ SELECT 2 -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "TABLE_OR_VIEW_NOT_FOUND", + "condition" : "TABLE_OR_VIEW_NOT_FOUND", "sqlState" : "42P01", "messageParameters" : { "relationName" : "`non_existing_table`" @@ -393,7 +393,7 @@ FROM t -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_INLINE_TABLE.CANNOT_EVALUATE_EXPRESSION_IN_INLINE_TABLE", + "condition" : "INVALID_INLINE_TABLE.CANNOT_EVALUATE_EXPRESSION_IN_INLINE_TABLE", "sqlState" : "42000", "messageParameters" : { "expr" : "\"outer(t.id)\"" @@ -432,7 +432,7 @@ SELECT same_name.x FROM (SELECT 10) AS same_name(x), same_name -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "AMBIGUOUS_REFERENCE", + "condition" : "AMBIGUOUS_REFERENCE", "sqlState" : "42704", "messageParameters" : { "name" : "`same_name`.`x`", @@ -468,7 +468,7 @@ select id from cte -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`id`", @@ -491,7 +491,7 @@ select 2 -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "TABLE_OR_VIEW_NOT_FOUND", + "condition" : "TABLE_OR_VIEW_NOT_FOUND", "sqlState" : "42P01", "messageParameters" : { "relationName" : "`r2`" @@ -559,7 +559,7 @@ SELECT * FROM -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "TABLE_OR_VIEW_NOT_FOUND", + "condition" : "TABLE_OR_VIEW_NOT_FOUND", "sqlState" : "42P01", "messageParameters" : { "relationName" : "`q`" @@ -723,7 +723,7 @@ with cte1 as (select 42), cte1 as (select 42) select * FROM cte1 -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "DUPLICATED_CTE_NAMES", + "condition" : "DUPLICATED_CTE_NAMES", "sqlState" : "42602", "messageParameters" : { "duplicateNames" : "`cte1`" @@ -767,7 +767,7 @@ with cte AS (SELECT * FROM va) SELECT * FROM cte -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "TABLE_OR_VIEW_NOT_FOUND", + "condition" : "TABLE_OR_VIEW_NOT_FOUND", "sqlState" : "42P01", "messageParameters" : { "relationName" : "`va`" @@ -787,7 +787,7 @@ with cte as (select * from cte) select * from cte -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "TABLE_OR_VIEW_NOT_FOUND", + "condition" : "TABLE_OR_VIEW_NOT_FOUND", "sqlState" : "42P01", "messageParameters" : { "relationName" : "`cte`" diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/date.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/date.sql.out index 0e4d2d4e99e2..a1c53df3f7c9 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/date.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/date.sql.out @@ -18,7 +18,7 @@ select date '2020-01-01中文' -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_TYPED_LITERAL", + "condition" : "INVALID_TYPED_LITERAL", "sqlState" : "42604", "messageParameters" : { "value" : "'2020-01-01中文'", @@ -60,7 +60,7 @@ select date'015' -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_TYPED_LITERAL", + "condition" : "INVALID_TYPED_LITERAL", "sqlState" : "42604", "messageParameters" : { "value" : "'015'", @@ -81,7 +81,7 @@ select date'2021-4294967297-11' -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_TYPED_LITERAL", + "condition" : "INVALID_TYPED_LITERAL", "sqlState" : "42604", "messageParameters" : { "value" : "'2021-4294967297-11'", @@ -114,7 +114,7 @@ select curdate(1) -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", + "condition" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", "sqlState" : "42605", "messageParameters" : { "actualNum" : "1", @@ -285,7 +285,7 @@ select date_add('2011-11-11', 1L) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"1\"", @@ -309,7 +309,7 @@ select date_add('2011-11-11', 1.0) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"1.0\"", @@ -333,7 +333,7 @@ select date_add('2011-11-11', 1E1) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"10.0\"", @@ -423,7 +423,7 @@ select date_sub('2011-11-11', 1L) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"1\"", @@ -447,7 +447,7 @@ select date_sub('2011-11-11', 1.0) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"1.0\"", @@ -471,7 +471,7 @@ select date_sub('2011-11-11', 1E1) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"10.0\"", @@ -576,7 +576,7 @@ select date '2011-11-11' + 1E1 -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"10.0\"", @@ -660,7 +660,7 @@ select date'2011-11-11' + '1' -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"1\"", @@ -684,7 +684,7 @@ select '1' + date'2011-11-11' -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"DATE '2011-11-11'\"", @@ -812,7 +812,7 @@ select dateadd('MICROSECOND', 1001, timestamp'2022-02-25 01:02:03.123') -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", + "condition" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", "sqlState" : "22023", "messageParameters" : { "functionName" : "`dateadd`", @@ -834,7 +834,7 @@ select date_add('QUARTER', 5, date'2022-02-25') -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", + "condition" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", "sqlState" : "22023", "messageParameters" : { "functionName" : "`date_add`", @@ -916,7 +916,7 @@ select date_diff('MILLISECOND', timestamp'2022-02-25 01:02:03.456', timestamp'20 -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", + "condition" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", "sqlState" : "22023", "messageParameters" : { "functionName" : "`date_diff`", @@ -938,7 +938,7 @@ select datediff('YEAR', date'2022-02-25', date'2023-02-25') -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", + "condition" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", "sqlState" : "22023", "messageParameters" : { "functionName" : "`datediff`", diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/datetime-legacy.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/datetime-legacy.sql.out index 3681a5dfd390..500852d6a9de 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/datetime-legacy.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/datetime-legacy.sql.out @@ -18,7 +18,7 @@ select date '2020-01-01中文' -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_TYPED_LITERAL", + "condition" : "INVALID_TYPED_LITERAL", "sqlState" : "42604", "messageParameters" : { "value" : "'2020-01-01中文'", @@ -60,7 +60,7 @@ select date'015' -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_TYPED_LITERAL", + "condition" : "INVALID_TYPED_LITERAL", "sqlState" : "42604", "messageParameters" : { "value" : "'015'", @@ -81,7 +81,7 @@ select date'2021-4294967297-11' -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_TYPED_LITERAL", + "condition" : "INVALID_TYPED_LITERAL", "sqlState" : "42604", "messageParameters" : { "value" : "'2021-4294967297-11'", @@ -114,7 +114,7 @@ select curdate(1) -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", + "condition" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", "sqlState" : "42605", "messageParameters" : { "actualNum" : "1", @@ -285,7 +285,7 @@ select date_add('2011-11-11', 1L) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"1\"", @@ -309,7 +309,7 @@ select date_add('2011-11-11', 1.0) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"1.0\"", @@ -333,7 +333,7 @@ select date_add('2011-11-11', 1E1) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"10.0\"", @@ -423,7 +423,7 @@ select date_sub('2011-11-11', 1L) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"1\"", @@ -447,7 +447,7 @@ select date_sub('2011-11-11', 1.0) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"1.0\"", @@ -471,7 +471,7 @@ select date_sub('2011-11-11', 1E1) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"10.0\"", @@ -576,7 +576,7 @@ select date '2011-11-11' + 1E1 -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"10.0\"", @@ -660,7 +660,7 @@ select date'2011-11-11' + '1' -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"1\"", @@ -684,7 +684,7 @@ select '1' + date'2011-11-11' -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"DATE '2011-11-11'\"", @@ -812,7 +812,7 @@ select dateadd('MICROSECOND', 1001, timestamp'2022-02-25 01:02:03.123') -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", + "condition" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", "sqlState" : "22023", "messageParameters" : { "functionName" : "`dateadd`", @@ -834,7 +834,7 @@ select date_add('QUARTER', 5, date'2022-02-25') -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", + "condition" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", "sqlState" : "22023", "messageParameters" : { "functionName" : "`date_add`", @@ -916,7 +916,7 @@ select date_diff('MILLISECOND', timestamp'2022-02-25 01:02:03.456', timestamp'20 -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", + "condition" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", "sqlState" : "22023", "messageParameters" : { "functionName" : "`date_diff`", @@ -938,7 +938,7 @@ select datediff('YEAR', date'2022-02-25', date'2023-02-25') -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", + "condition" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", "sqlState" : "22023", "messageParameters" : { "functionName" : "`datediff`", @@ -966,7 +966,7 @@ select timestamp '2019-01-01中文' -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_TYPED_LITERAL", + "condition" : "INVALID_TYPED_LITERAL", "sqlState" : "42604", "messageParameters" : { "value" : "'2019-01-01中文'", @@ -987,7 +987,7 @@ select timestamp'4294967297' -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_TYPED_LITERAL", + "condition" : "INVALID_TYPED_LITERAL", "sqlState" : "42604", "messageParameters" : { "value" : "'4294967297'", @@ -1008,7 +1008,7 @@ select timestamp'2021-01-01T12:30:4294967297.123456' -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_TYPED_LITERAL", + "condition" : "INVALID_TYPED_LITERAL", "sqlState" : "42604", "messageParameters" : { "value" : "'2021-01-01T12:30:4294967297.123456'", @@ -1553,7 +1553,7 @@ select timestamp'2011-11-11 11:11:11' + '1' -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"TIMESTAMP\"", @@ -1575,7 +1575,7 @@ select '1' + timestamp'2011-11-11 11:11:11' -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"TIMESTAMP\"", @@ -1597,7 +1597,7 @@ select timestamp'2011-11-11 11:11:11' + null -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP\"", @@ -1619,7 +1619,7 @@ select null + timestamp'2011-11-11 11:11:11' -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"VOID\"", @@ -1726,7 +1726,7 @@ select timestampadd('MONTH', -1, timestamp'2022-02-14 01:02:03') -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", + "condition" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", "sqlState" : "22023", "messageParameters" : { "functionName" : "`timestampadd`", @@ -1748,7 +1748,7 @@ select timestampadd('SECOND', -1, date'2022-02-15') -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", + "condition" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", "sqlState" : "22023", "messageParameters" : { "functionName" : "`timestampadd`", @@ -1794,7 +1794,7 @@ select timestampdiff('MINUTE', timestamp'2022-02-14 01:02:03', timestamp'2022-02 -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", + "condition" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", "sqlState" : "22023", "messageParameters" : { "functionName" : "`timestampdiff`", @@ -1816,7 +1816,7 @@ select timestampdiff('YEAR', date'2022-02-15', date'2023-02-15') -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", + "condition" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", "sqlState" : "22023", "messageParameters" : { "functionName" : "`timestampdiff`", @@ -1862,7 +1862,7 @@ select timediff('MINUTE', timestamp'2023-02-14 01:02:03', timestamp'2023-02-14 0 -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", + "condition" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", "sqlState" : "22023", "messageParameters" : { "functionName" : "`timediff`", @@ -1884,7 +1884,7 @@ select timediff('YEAR', date'2020-02-15', date'2023-02-15') -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", + "condition" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", "sqlState" : "22023", "messageParameters" : { "functionName" : "`timediff`", diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/describe-query.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/describe-query.sql.out index 5119d6a08f29..03bb7d447bb8 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/describe-query.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/describe-query.sql.out @@ -76,7 +76,7 @@ DESCRIBE INSERT INTO desc_temp1 values (1, 'val1') -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'desc_temp1'", @@ -90,7 +90,7 @@ DESCRIBE INSERT INTO desc_temp1 SELECT * FROM desc_temp2 -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'desc_temp1'", @@ -107,7 +107,7 @@ DESCRIBE -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'insert'", diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/describe.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/describe.sql.out index 307a0a3e25fb..99f7326e5ef8 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/describe.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/describe.sql.out @@ -68,7 +68,7 @@ DESCRIBE t AS JSON -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "DESCRIBE_JSON_NOT_EXTENDED", + "condition" : "DESCRIBE_JSON_NOT_EXTENDED", "sqlState" : "0A000", "messageParameters" : { "tableName" : "t" @@ -81,7 +81,7 @@ DESC FORMATTED t a AS JSON -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "UNSUPPORTED_FEATURE.DESC_TABLE_COLUMN_JSON", + "condition" : "UNSUPPORTED_FEATURE.DESC_TABLE_COLUMN_JSON", "sqlState" : "0A000" } @@ -170,7 +170,7 @@ DESC t PARTITION (c='Us', d=2) -- !query analysis org.apache.spark.sql.catalyst.analysis.NoSuchPartitionException { - "errorClass" : "PARTITIONS_NOT_FOUND", + "condition" : "PARTITIONS_NOT_FOUND", "sqlState" : "428FT", "messageParameters" : { "partitionList" : "PARTITION (`c` = Us, `d` = 2)", @@ -184,7 +184,7 @@ DESC t PARTITION (c='Us') -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "_LEGACY_ERROR_TEMP_1232", + "condition" : "_LEGACY_ERROR_TEMP_1232", "messageParameters" : { "partitionColumnNames" : "c, d", "specKeys" : "c", @@ -198,7 +198,7 @@ DESC t PARTITION (c='Us', d) -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_SQL_SYNTAX.EMPTY_PARTITION_VALUE", + "condition" : "INVALID_SQL_SYNTAX.EMPTY_PARTITION_VALUE", "sqlState" : "42000", "messageParameters" : { "partKey" : "`d`" @@ -248,7 +248,7 @@ DESC temp_v PARTITION (c='Us', d=1) -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "FORBIDDEN_OPERATION", + "condition" : "FORBIDDEN_OPERATION", "sqlState" : "42809", "messageParameters" : { "objectName" : "`temp_v`", @@ -287,7 +287,7 @@ DESC v PARTITION (c='Us', d=1) -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "FORBIDDEN_OPERATION", + "condition" : "FORBIDDEN_OPERATION", "sqlState" : "42809", "messageParameters" : { "objectName" : "`v`", @@ -363,7 +363,7 @@ CREATE TABLE d (a STRING DEFAULT 'default-value', b INT DEFAULT 42) USING parque -- !query analysis org.apache.spark.sql.catalyst.analysis.TableAlreadyExistsException { - "errorClass" : "TABLE_OR_VIEW_ALREADY_EXISTS", + "condition" : "TABLE_OR_VIEW_ALREADY_EXISTS", "sqlState" : "42P07", "messageParameters" : { "relationName" : "`spark_catalog`.`default`.`d`" @@ -400,7 +400,7 @@ CREATE TABLE e (a STRING DEFAULT CONCAT('a\n b\n ', 'c\n d'), b INT DEFAULT 42) -- !query analysis org.apache.spark.sql.catalyst.analysis.TableAlreadyExistsException { - "errorClass" : "TABLE_OR_VIEW_ALREADY_EXISTS", + "condition" : "TABLE_OR_VIEW_ALREADY_EXISTS", "sqlState" : "42P07", "messageParameters" : { "relationName" : "`spark_catalog`.`default`.`e`" diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/double-quoted-identifiers-enabled.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/double-quoted-identifiers-enabled.sql.out index f2b1d91d7e44..f9f0067648fc 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/double-quoted-identifiers-enabled.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/double-quoted-identifiers-enabled.sql.out @@ -4,7 +4,7 @@ SELECT 1 FROM "not_exist" -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "TABLE_OR_VIEW_NOT_FOUND", + "condition" : "TABLE_OR_VIEW_NOT_FOUND", "sqlState" : "42P01", "messageParameters" : { "relationName" : "`not_exist`" @@ -24,7 +24,7 @@ USE SCHEMA "not_exist" -- !query analysis org.apache.spark.sql.catalyst.analysis.NoSuchNamespaceException { - "errorClass" : "SCHEMA_NOT_FOUND", + "condition" : "SCHEMA_NOT_FOUND", "sqlState" : "42704", "messageParameters" : { "schemaName" : "`spark_catalog`.`not_exist`" @@ -37,7 +37,7 @@ ALTER TABLE "not_exist" ADD COLUMN not_exist int -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "TABLE_OR_VIEW_NOT_FOUND", + "condition" : "TABLE_OR_VIEW_NOT_FOUND", "sqlState" : "42P01", "messageParameters" : { "relationName" : "`not_exist`" @@ -57,7 +57,7 @@ ALTER TABLE not_exist ADD COLUMN "not_exist" int -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "TABLE_OR_VIEW_NOT_FOUND", + "condition" : "TABLE_OR_VIEW_NOT_FOUND", "sqlState" : "42P01", "messageParameters" : { "relationName" : "`not_exist`" @@ -77,7 +77,7 @@ SELECT 1 AS "not_exist" FROM not_exist -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "TABLE_OR_VIEW_NOT_FOUND", + "condition" : "TABLE_OR_VIEW_NOT_FOUND", "sqlState" : "42P01", "messageParameters" : { "relationName" : "`not_exist`" @@ -97,7 +97,7 @@ SELECT 1 FROM not_exist AS X("hello") -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "TABLE_OR_VIEW_NOT_FOUND", + "condition" : "TABLE_OR_VIEW_NOT_FOUND", "sqlState" : "42P01", "messageParameters" : { "relationName" : "`not_exist`" @@ -117,7 +117,7 @@ SELECT "not_exist"() -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "UNRESOLVED_ROUTINE", + "condition" : "UNRESOLVED_ROUTINE", "sqlState" : "42883", "messageParameters" : { "routineName" : "`not_exist`", @@ -138,7 +138,7 @@ SELECT "not_exist".not_exist() -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "UNRESOLVED_ROUTINE", + "condition" : "UNRESOLVED_ROUTINE", "sqlState" : "42883", "messageParameters" : { "routineName" : "`not_exist`.`not_exist`", @@ -159,7 +159,7 @@ SELECT 1 FROM `hello` -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "TABLE_OR_VIEW_NOT_FOUND", + "condition" : "TABLE_OR_VIEW_NOT_FOUND", "sqlState" : "42P01", "messageParameters" : { "relationName" : "`hello`" @@ -179,7 +179,7 @@ USE SCHEMA `not_exist` -- !query analysis org.apache.spark.sql.catalyst.analysis.NoSuchNamespaceException { - "errorClass" : "SCHEMA_NOT_FOUND", + "condition" : "SCHEMA_NOT_FOUND", "sqlState" : "42704", "messageParameters" : { "schemaName" : "`spark_catalog`.`not_exist`" @@ -192,7 +192,7 @@ ALTER TABLE `not_exist` ADD COLUMN not_exist int -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "TABLE_OR_VIEW_NOT_FOUND", + "condition" : "TABLE_OR_VIEW_NOT_FOUND", "sqlState" : "42P01", "messageParameters" : { "relationName" : "`not_exist`" @@ -212,7 +212,7 @@ ALTER TABLE not_exist ADD COLUMN `not_exist` int -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "TABLE_OR_VIEW_NOT_FOUND", + "condition" : "TABLE_OR_VIEW_NOT_FOUND", "sqlState" : "42P01", "messageParameters" : { "relationName" : "`not_exist`" @@ -232,7 +232,7 @@ SELECT 1 AS `not_exist` FROM `not_exist` -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "TABLE_OR_VIEW_NOT_FOUND", + "condition" : "TABLE_OR_VIEW_NOT_FOUND", "sqlState" : "42P01", "messageParameters" : { "relationName" : "`not_exist`" @@ -252,7 +252,7 @@ SELECT 1 FROM not_exist AS X(`hello`) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "TABLE_OR_VIEW_NOT_FOUND", + "condition" : "TABLE_OR_VIEW_NOT_FOUND", "sqlState" : "42P01", "messageParameters" : { "relationName" : "`not_exist`" @@ -272,7 +272,7 @@ SELECT `not_exist`() -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "UNRESOLVED_ROUTINE", + "condition" : "UNRESOLVED_ROUTINE", "sqlState" : "42883", "messageParameters" : { "routineName" : "`not_exist`", @@ -293,7 +293,7 @@ SELECT `not_exist`.not_exist() -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "UNRESOLVED_ROUTINE", + "condition" : "UNRESOLVED_ROUTINE", "sqlState" : "42883", "messageParameters" : { "routineName" : "`not_exist`.`not_exist`", @@ -314,7 +314,7 @@ SELECT "hello" -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITHOUT_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITHOUT_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`hello`" @@ -334,7 +334,7 @@ CREATE TEMPORARY VIEW v(c1 COMMENT "hello") AS SELECT 1 -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'\"hello\"'", @@ -348,7 +348,7 @@ DROP VIEW v -- !query analysis org.apache.spark.sql.catalyst.analysis.NoSuchTableException { - "errorClass" : "TABLE_OR_VIEW_NOT_FOUND", + "condition" : "TABLE_OR_VIEW_NOT_FOUND", "sqlState" : "42P01", "messageParameters" : { "relationName" : "`spark_catalog`.`default`.`v`" @@ -361,7 +361,7 @@ SELECT INTERVAL "1" YEAR -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'\"1\"'", diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/double-quoted-identifiers.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/double-quoted-identifiers.sql.out index a02bf525f947..f2ee4fb48e63 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/double-quoted-identifiers.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/double-quoted-identifiers.sql.out @@ -4,7 +4,7 @@ SELECT 1 FROM "not_exist" -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'\"not_exist\"'", @@ -18,7 +18,7 @@ USE SCHEMA "not_exist" -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'\"not_exist\"'", @@ -32,7 +32,7 @@ ALTER TABLE "not_exist" ADD COLUMN not_exist int -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'\"not_exist\"'", @@ -46,7 +46,7 @@ ALTER TABLE not_exist ADD COLUMN "not_exist" int -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'\"not_exist\"'", @@ -60,7 +60,7 @@ SELECT 1 AS "not_exist" FROM not_exist -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'\"not_exist\"'", @@ -74,7 +74,7 @@ SELECT 1 FROM not_exist AS X("hello") -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'\"hello\"'", @@ -88,7 +88,7 @@ SELECT "not_exist"() -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'\"not_exist\"'", @@ -102,7 +102,7 @@ SELECT "not_exist".not_exist() -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'\"not_exist\"'", @@ -116,7 +116,7 @@ SELECT 1 FROM `hello` -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "TABLE_OR_VIEW_NOT_FOUND", + "condition" : "TABLE_OR_VIEW_NOT_FOUND", "sqlState" : "42P01", "messageParameters" : { "relationName" : "`hello`" @@ -136,7 +136,7 @@ USE SCHEMA `not_exist` -- !query analysis org.apache.spark.sql.catalyst.analysis.NoSuchNamespaceException { - "errorClass" : "SCHEMA_NOT_FOUND", + "condition" : "SCHEMA_NOT_FOUND", "sqlState" : "42704", "messageParameters" : { "schemaName" : "`spark_catalog`.`not_exist`" @@ -149,7 +149,7 @@ ALTER TABLE `not_exist` ADD COLUMN not_exist int -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "TABLE_OR_VIEW_NOT_FOUND", + "condition" : "TABLE_OR_VIEW_NOT_FOUND", "sqlState" : "42P01", "messageParameters" : { "relationName" : "`not_exist`" @@ -169,7 +169,7 @@ ALTER TABLE not_exist ADD COLUMN `not_exist` int -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "TABLE_OR_VIEW_NOT_FOUND", + "condition" : "TABLE_OR_VIEW_NOT_FOUND", "sqlState" : "42P01", "messageParameters" : { "relationName" : "`not_exist`" @@ -189,7 +189,7 @@ SELECT 1 AS `not_exist` FROM `not_exist` -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "TABLE_OR_VIEW_NOT_FOUND", + "condition" : "TABLE_OR_VIEW_NOT_FOUND", "sqlState" : "42P01", "messageParameters" : { "relationName" : "`not_exist`" @@ -209,7 +209,7 @@ SELECT 1 FROM not_exist AS X(`hello`) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "TABLE_OR_VIEW_NOT_FOUND", + "condition" : "TABLE_OR_VIEW_NOT_FOUND", "sqlState" : "42P01", "messageParameters" : { "relationName" : "`not_exist`" @@ -229,7 +229,7 @@ SELECT `not_exist`() -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "UNRESOLVED_ROUTINE", + "condition" : "UNRESOLVED_ROUTINE", "sqlState" : "42883", "messageParameters" : { "routineName" : "`not_exist`", @@ -250,7 +250,7 @@ SELECT `not_exist`.not_exist() -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "UNRESOLVED_ROUTINE", + "condition" : "UNRESOLVED_ROUTINE", "sqlState" : "42883", "messageParameters" : { "routineName" : "`not_exist`.`not_exist`", @@ -327,7 +327,7 @@ CREATE SCHEMA "myschema" -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'\"myschema\"'", @@ -342,7 +342,7 @@ CREATE TEMPORARY VIEW "myview"("c1") AS -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'\"myview\"'", @@ -356,7 +356,7 @@ SELECT "a1" AS "a2" FROM "myview" AS "atab"("a1") -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'\"a2\"'", @@ -370,7 +370,7 @@ DROP TABLE "myview" -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'\"myview\"'", @@ -384,7 +384,7 @@ DROP SCHEMA "myschema" -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'\"myschema\"'", diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/except-all.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/except-all.sql.out index 3972cde51bb7..0f7508eb53b7 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/except-all.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/except-all.sql.out @@ -201,7 +201,7 @@ SELECT array(1) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -378,7 +378,7 @@ SELECT k, v FROM tab4 -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "NUM_COLUMNS_MISMATCH", + "condition" : "NUM_COLUMNS_MISMATCH", "sqlState" : "42826", "messageParameters" : { "firstNumColumns" : "1", diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/execute-immediate.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/execute-immediate.sql.out index ce510527c878..d28401118441 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/execute-immediate.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/execute-immediate.sql.out @@ -468,7 +468,7 @@ EXECUTE IMMEDIATE 'SELECT \'invalid_cast_error_expected\'' INTO res_id -- !query analysis org.apache.spark.SparkNumberFormatException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'invalid_cast_error_expected'", @@ -490,7 +490,7 @@ EXECUTE IMMEDIATE 'INSERT INTO x VALUES (?)' INTO res_id USING 1 -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_STATEMENT_FOR_EXECUTE_INTO", + "condition" : "INVALID_STATEMENT_FOR_EXECUTE_INTO", "sqlState" : "07501", "messageParameters" : { "sqlString" : "INSERT INTO X VALUES (?)" @@ -503,7 +503,7 @@ EXECUTE IMMEDIATE 'SELECT * FROM tbl_view WHERE ? = id' USING id -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "UNRESOLVED_VARIABLE", + "condition" : "UNRESOLVED_VARIABLE", "sqlState" : "42883", "messageParameters" : { "searchPath" : "`system`.`session`", @@ -524,7 +524,7 @@ EXECUTE IMMEDIATE 'SELECT * FROM tbl_view where ? = id and :first = name' USING -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_QUERY_MIXED_QUERY_PARAMETERS", + "condition" : "INVALID_QUERY_MIXED_QUERY_PARAMETERS", "sqlState" : "42613" } @@ -534,7 +534,7 @@ EXECUTE IMMEDIATE 'SELECT * FROM tbl_view where :first = name' USING 1, 'name2' -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "ALL_PARAMETERS_MUST_BE_NAMED", + "condition" : "ALL_PARAMETERS_MUST_BE_NAMED", "sqlState" : "07001", "messageParameters" : { "exprs" : "\"1\"" @@ -547,7 +547,7 @@ EXECUTE IMMEDIATE 'SELCT Fa' -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'SELCT'", @@ -568,7 +568,7 @@ EXECUTE IMMEDIATE 'SELCT Fa' INTO res_id -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'SELCT'", @@ -589,7 +589,7 @@ EXECUTE IMMEDIATE b -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_VARIABLE_TYPE_FOR_QUERY_EXECUTE_IMMEDIATE", + "condition" : "INVALID_VARIABLE_TYPE_FOR_QUERY_EXECUTE_IMMEDIATE", "sqlState" : "42K09", "messageParameters" : { "varType" : "\"INT\"" @@ -618,7 +618,7 @@ EXECUTE IMMEDIATE 'SELECT * from tbl_view where name = :first' USING CONCAT(a , -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "UNSUPPORTED_EXPR_FOR_PARAMETER", + "condition" : "UNSUPPORTED_EXPR_FOR_PARAMETER", "sqlState" : "42K0E", "messageParameters" : { "invalidExprSql" : "\"CONCAT(a, me1)\"" @@ -638,7 +638,7 @@ EXECUTE IMMEDIATE 'SELECT * from tbl_view where name = :first' USING (SELECT 42) -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "UNSUPPORTED_EXPR_FOR_PARAMETER", + "condition" : "UNSUPPORTED_EXPR_FOR_PARAMETER", "sqlState" : "42K0E", "messageParameters" : { "invalidExprSql" : "\"scalarsubquery()\"" @@ -658,7 +658,7 @@ EXECUTE IMMEDIATE 'SELECT id, name FROM tbl_view WHERE id = ?' INTO a, b USING 1 -- !query analysis org.apache.spark.SparkNumberFormatException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'name1'", @@ -680,7 +680,7 @@ EXECUTE IMMEDIATE 'SELECT id, name FROM tbl_view WHERE id = ?' INTO (a, b) USING -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'('", @@ -694,7 +694,7 @@ EXECUTE IMMEDIATE 'SELECT id FROM tbl_view' INTO res_id -- !query analysis org.apache.spark.SparkException { - "errorClass" : "ROW_SUBQUERY_TOO_MANY_ROWS", + "condition" : "ROW_SUBQUERY_TOO_MANY_ROWS", "sqlState" : "21000" } @@ -704,7 +704,7 @@ EXECUTE IMMEDIATE 'SELECT id, data.f1 FROM tbl_view' INTO res_id -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "ASSIGNMENT_ARITY_MISMATCH", + "condition" : "ASSIGNMENT_ARITY_MISMATCH", "sqlState" : "42802", "messageParameters" : { "numExpr" : "2", @@ -718,7 +718,7 @@ EXECUTE IMMEDIATE 'SELECT id FROM tbl_view' INTO res_id, b -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "ASSIGNMENT_ARITY_MISMATCH", + "condition" : "ASSIGNMENT_ARITY_MISMATCH", "sqlState" : "42802", "messageParameters" : { "numExpr" : "1", @@ -732,7 +732,7 @@ EXECUTE IMMEDIATE 'SELECT id FROM tbl_view WHERE id = :first' USING 10 as first, -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "EXEC_IMMEDIATE_DUPLICATE_ARGUMENT_ALIASES", + "condition" : "EXEC_IMMEDIATE_DUPLICATE_ARGUMENT_ALIASES", "sqlState" : "42701", "messageParameters" : { "aliases" : "`first`" @@ -772,7 +772,7 @@ EXECUTE IMMEDIATE 'SELECT id FROM tbl_view WHERE id = :p' USING p, 'p' -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "ALL_PARAMETERS_MUST_BE_NAMED", + "condition" : "ALL_PARAMETERS_MUST_BE_NAMED", "sqlState" : "07001", "messageParameters" : { "exprs" : "\"p\"" @@ -785,7 +785,7 @@ EXECUTE IMMEDIATE 'SELECT id, data.f1 FROM tbl_view WHERE id = 10' INTO res_id, -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "DUPLICATE_ASSIGNMENTS", + "condition" : "DUPLICATE_ASSIGNMENTS", "sqlState" : "42701", "messageParameters" : { "nameList" : "`res_id`" @@ -798,7 +798,7 @@ EXECUTE IMMEDIATE 'EXECUTE IMMEDIATE \'SELECT id FROM tbl_view WHERE id = ? USIN -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "NESTED_EXECUTE_IMMEDIATE", + "condition" : "NESTED_EXECUTE_IMMEDIATE", "sqlState" : "07501", "messageParameters" : { "sqlString" : "EXECUTE IMMEDIATE 'SELECT ID FROM TBL_VIEW WHERE ID = ? USING 10'" @@ -820,7 +820,7 @@ EXECUTE IMMEDIATE sql_string -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "NULL_QUERY_STRING_EXECUTE_IMMEDIATE", + "condition" : "NULL_QUERY_STRING_EXECUTE_IMMEDIATE", "sqlState" : "22004", "messageParameters" : { "varName" : "`sql_string`" diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/extract.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/extract.sql.out index 987941eee05c..abbe8638597c 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/extract.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/extract.sql.out @@ -430,7 +430,7 @@ select extract(not_supported from c) from t -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_EXTRACT_FIELD", + "condition" : "INVALID_EXTRACT_FIELD", "sqlState" : "42601", "messageParameters" : { "expr" : "\"c\"", @@ -451,7 +451,7 @@ select extract(not_supported from i) from t -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_EXTRACT_FIELD", + "condition" : "INVALID_EXTRACT_FIELD", "sqlState" : "42601", "messageParameters" : { "expr" : "\"i\"", @@ -472,7 +472,7 @@ select extract(not_supported from j) from t -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_EXTRACT_FIELD", + "condition" : "INVALID_EXTRACT_FIELD", "sqlState" : "42601", "messageParameters" : { "expr" : "\"j\"", @@ -911,7 +911,7 @@ select date_part('not_supported', c) from t -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_EXTRACT_FIELD", + "condition" : "INVALID_EXTRACT_FIELD", "sqlState" : "42601", "messageParameters" : { "expr" : "\"c\"", @@ -932,7 +932,7 @@ select date_part(c, c) from t -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "NON_FOLDABLE_ARGUMENT", + "condition" : "NON_FOLDABLE_ARGUMENT", "sqlState" : "42K08", "messageParameters" : { "funcName" : "`date_part`", @@ -965,7 +965,7 @@ select date_part(i, i) from t -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "NON_FOLDABLE_ARGUMENT", + "condition" : "NON_FOLDABLE_ARGUMENT", "sqlState" : "42K08", "messageParameters" : { "funcName" : "`date_part`", @@ -1234,7 +1234,7 @@ select extract(DAY from interval '2-1' YEAR TO MONTH) -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_EXTRACT_FIELD", + "condition" : "INVALID_EXTRACT_FIELD", "sqlState" : "42601", "messageParameters" : { "expr" : "\"INTERVAL '2-1' YEAR TO MONTH\"", @@ -1255,7 +1255,7 @@ select date_part('DAY', interval '2-1' YEAR TO MONTH) -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_EXTRACT_FIELD", + "condition" : "INVALID_EXTRACT_FIELD", "sqlState" : "42601", "messageParameters" : { "expr" : "\"INTERVAL '2-1' YEAR TO MONTH\"", @@ -1276,7 +1276,7 @@ select date_part('not_supported', interval '2-1' YEAR TO MONTH) -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_EXTRACT_FIELD", + "condition" : "INVALID_EXTRACT_FIELD", "sqlState" : "42601", "messageParameters" : { "expr" : "\"INTERVAL '2-1' YEAR TO MONTH\"", @@ -1388,7 +1388,7 @@ select extract(MONTH from interval '123 12:34:56.789123123' DAY TO SECOND) -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_EXTRACT_FIELD", + "condition" : "INVALID_EXTRACT_FIELD", "sqlState" : "42601", "messageParameters" : { "expr" : "\"INTERVAL '123 12:34:56.789123' DAY TO SECOND\"", @@ -1409,7 +1409,7 @@ select date_part('not_supported', interval '123 12:34:56.789123123' DAY TO SECON -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_EXTRACT_FIELD", + "condition" : "INVALID_EXTRACT_FIELD", "sqlState" : "42601", "messageParameters" : { "expr" : "\"INTERVAL '123 12:34:56.789123' DAY TO SECOND\"", diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/group-analytics.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/group-analytics.sql.out index 8b2e55f12628..f085e47c08ec 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/group-analytics.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/group-analytics.sql.out @@ -116,7 +116,7 @@ SELECT course, year, SUM(earnings) FROM courseSales GROUP BY ROLLUP(course, year -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0051", + "condition" : "_LEGACY_ERROR_TEMP_0051", "messageParameters" : { "element" : "ROLLUP" }, @@ -165,7 +165,7 @@ SELECT course, year, SUM(earnings) FROM courseSales GROUP BY CUBE(course, year, -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0051", + "condition" : "_LEGACY_ERROR_TEMP_0051", "messageParameters" : { "element" : "CUBE" }, @@ -332,7 +332,7 @@ SELECT course, year, GROUPING(course) FROM courseSales GROUP BY course, year -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNSUPPORTED_GROUPING_EXPRESSION", + "condition" : "UNSUPPORTED_GROUPING_EXPRESSION", "sqlState" : "42K0E", "queryContext" : [ { "objectType" : "", @@ -349,7 +349,7 @@ SELECT course, year, GROUPING_ID(course, year) FROM courseSales GROUP BY course, -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNSUPPORTED_GROUPING_EXPRESSION", + "condition" : "UNSUPPORTED_GROUPING_EXPRESSION", "sqlState" : "42K0E", "queryContext" : [ { "objectType" : "", @@ -399,7 +399,7 @@ SELECT course, year FROM courseSales GROUP BY course, year HAVING GROUPING(cours -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "UNSUPPORTED_GROUPING_EXPRESSION", + "condition" : "UNSUPPORTED_GROUPING_EXPRESSION", "sqlState" : "42K0E" } @@ -409,7 +409,7 @@ SELECT course, year FROM courseSales GROUP BY course, year HAVING GROUPING_ID(co -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "UNSUPPORTED_GROUPING_EXPRESSION", + "condition" : "UNSUPPORTED_GROUPING_EXPRESSION", "sqlState" : "42K0E" } @@ -469,7 +469,7 @@ SELECT course, year FROM courseSales GROUP BY course, year ORDER BY GROUPING(cou -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "UNSUPPORTED_GROUPING_EXPRESSION", + "condition" : "UNSUPPORTED_GROUPING_EXPRESSION", "sqlState" : "42K0E" } @@ -479,7 +479,7 @@ SELECT course, year FROM courseSales GROUP BY course, year ORDER BY GROUPING_ID( -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "UNSUPPORTED_GROUPING_EXPRESSION", + "condition" : "UNSUPPORTED_GROUPING_EXPRESSION", "sqlState" : "42K0E" } diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/group-by-all-duckdb.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/group-by-all-duckdb.sql.out index 5f1cbea70989..4bf359c02ef2 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/group-by-all-duckdb.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/group-by-all-duckdb.sql.out @@ -75,7 +75,7 @@ SELECT (g+i)%2 + SUM(i), SUM(i), SUM(g) FROM integers GROUP BY ALL ORDER BY 1 -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_ALL_IN_GROUP_BY", + "condition" : "UNRESOLVED_ALL_IN_GROUP_BY", "sqlState" : "42803", "queryContext" : [ { "objectType" : "", @@ -128,7 +128,7 @@ SELECT c0 FROM (SELECT 1 c0, 1 c1 UNION ALL SELECT 1, 2) t0 GROUP BY ALL HAVING -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`c1`", diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/group-by-all-mosha.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/group-by-all-mosha.sql.out index 34d22ad6c617..2f90f863aa76 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/group-by-all-mosha.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/group-by-all-mosha.sql.out @@ -74,7 +74,7 @@ SELECT i + SUM(f) FROM stuff GROUP BY ALL -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_ALL_IN_GROUP_BY", + "condition" : "UNRESOLVED_ALL_IN_GROUP_BY", "sqlState" : "42803", "queryContext" : [ { "objectType" : "", @@ -132,7 +132,7 @@ SELECT SUM(i) si FROM stuff GROUP BY ALL HAVING i > 2 -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`i`", @@ -153,7 +153,7 @@ SELECT SUM(i) si FROM stuff GROUP BY ALL ORDER BY i DESC -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`i`", diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/group-by-all.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/group-by-all.sql.out index c2c77db6c3b1..14c6376b09a5 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/group-by-all.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/group-by-all.sql.out @@ -226,7 +226,7 @@ select id + count(*) from data group by all -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_ALL_IN_GROUP_BY", + "condition" : "UNRESOLVED_ALL_IN_GROUP_BY", "sqlState" : "42803", "queryContext" : [ { "objectType" : "", @@ -243,7 +243,7 @@ select (id + id) / 2 + count(*) * 2 from data group by all -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_ALL_IN_GROUP_BY", + "condition" : "UNRESOLVED_ALL_IN_GROUP_BY", "sqlState" : "42803", "queryContext" : [ { "objectType" : "", @@ -317,7 +317,7 @@ select (select count(*) from data d1 where d1.country = d2.country) + count(id) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_ALL_IN_GROUP_BY", + "condition" : "UNRESOLVED_ALL_IN_GROUP_BY", "sqlState" : "42803", "queryContext" : [ { "objectType" : "", @@ -334,7 +334,7 @@ select non_exist from data group by all -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`non_exist`", diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/group-by-filter.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/group-by-filter.sql.out index dd36da772355..8f1cd1cca8a3 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/group-by-filter.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/group-by-filter.sql.out @@ -86,7 +86,7 @@ SELECT a, COUNT(b) FILTER (WHERE a >= 2) FROM testData -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "MISSING_GROUP_BY", + "condition" : "MISSING_GROUP_BY", "sqlState" : "42803", "queryContext" : [ { "objectType" : "", @@ -343,7 +343,7 @@ SELECT a, COUNT(b) FILTER (WHERE a != 2) FROM testData GROUP BY b -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "MISSING_AGGREGATION", + "condition" : "MISSING_AGGREGATION", "sqlState" : "42803", "messageParameters" : { "expression" : "\"a\"", @@ -773,7 +773,7 @@ SELECT a + 2, COUNT(b) FILTER (WHERE b IN (1, 2)) FROM testData GROUP BY a + 1 -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "MISSING_AGGREGATION", + "condition" : "MISSING_AGGREGATION", "sqlState" : "42803", "messageParameters" : { "expression" : "\"a\"", @@ -980,7 +980,7 @@ SELECT count(num1) FILTER (WHERE rand(int(num2)) > 1) FROM FilterExpressionTestD -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_AGGREGATE_FILTER.NON_DETERMINISTIC", + "condition" : "INVALID_AGGREGATE_FILTER.NON_DETERMINISTIC", "sqlState" : "42903", "messageParameters" : { "filterExpr" : "\"(rand(num2) > 1)\"" @@ -1000,7 +1000,7 @@ SELECT sum(num1) FILTER (WHERE str) FROM FilterExpressionTestData -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_AGGREGATE_FILTER.NOT_BOOLEAN", + "condition" : "INVALID_AGGREGATE_FILTER.NOT_BOOLEAN", "sqlState" : "42903", "messageParameters" : { "filterExpr" : "\"str\"" @@ -1020,7 +1020,7 @@ SELECT sum(num1) FILTER (WHERE max(num2) > 1) FROM FilterExpressionTestData -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_AGGREGATE_FILTER.CONTAINS_AGGREGATE", + "condition" : "INVALID_AGGREGATE_FILTER.CONTAINS_AGGREGATE", "sqlState" : "42903", "messageParameters" : { "aggExpr" : "\"max(num2)\"", @@ -1041,7 +1041,7 @@ SELECT sum(num1) FILTER (WHERE nth_value(num2, 2) OVER(ORDER BY num2) > 1) FROM -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_AGGREGATE_FILTER.CONTAINS_WINDOW_FUNCTION", + "condition" : "INVALID_AGGREGATE_FILTER.CONTAINS_WINDOW_FUNCTION", "sqlState" : "42903", "messageParameters" : { "filterExpr" : "\"(nth_value(num2, 2) OVER (ORDER BY num2 ASC NULLS FIRST RANGE BETWEEN UNBOUNDED PRECEDING AND CURRENT ROW) > 1)\"", diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/group-by-ordinal.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/group-by-ordinal.sql.out index 904b35559ced..79f6f24881bd 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/group-by-ordinal.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/group-by-ordinal.sql.out @@ -111,7 +111,7 @@ select a, b from data group by -1 -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "GROUP_BY_POS_OUT_OF_RANGE", + "condition" : "GROUP_BY_POS_OUT_OF_RANGE", "sqlState" : "42805", "messageParameters" : { "index" : "-1", @@ -132,7 +132,7 @@ select a, b from data group by 0 -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "GROUP_BY_POS_OUT_OF_RANGE", + "condition" : "GROUP_BY_POS_OUT_OF_RANGE", "sqlState" : "42805", "messageParameters" : { "index" : "0", @@ -153,7 +153,7 @@ select a, b from data group by 3 -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "GROUP_BY_POS_OUT_OF_RANGE", + "condition" : "GROUP_BY_POS_OUT_OF_RANGE", "sqlState" : "42805", "messageParameters" : { "index" : "3", @@ -174,7 +174,7 @@ select a, b, sum(b) from data group by 3 -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "GROUP_BY_POS_AGGREGATE", + "condition" : "GROUP_BY_POS_AGGREGATE", "sqlState" : "42903", "messageParameters" : { "aggExpr" : "sum(data.b) AS `sum(b)`", @@ -195,7 +195,7 @@ select a, b, sum(b) + 2 from data group by 3 -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "GROUP_BY_POS_AGGREGATE", + "condition" : "GROUP_BY_POS_AGGREGATE", "sqlState" : "42903", "messageParameters" : { "aggExpr" : "(sum(data.b) + CAST(2 AS BIGINT)) AS `(sum(b) + 2)`", @@ -224,7 +224,7 @@ select * from data group by a, b, 1 -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "STAR_GROUP_BY_POS", + "condition" : "STAR_GROUP_BY_POS", "sqlState" : "0A000" } @@ -392,7 +392,7 @@ select a, b, count(1) from data group by a, -1 -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "GROUP_BY_POS_OUT_OF_RANGE", + "condition" : "GROUP_BY_POS_OUT_OF_RANGE", "sqlState" : "42805", "messageParameters" : { "index" : "-1", @@ -413,7 +413,7 @@ select a, b, count(1) from data group by a, 3 -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "GROUP_BY_POS_AGGREGATE", + "condition" : "GROUP_BY_POS_AGGREGATE", "sqlState" : "42903", "messageParameters" : { "aggExpr" : "count(1) AS `count(1)`", @@ -434,7 +434,7 @@ select a, b, count(1) from data group by cube(-1, 2) -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "GROUP_BY_POS_OUT_OF_RANGE", + "condition" : "GROUP_BY_POS_OUT_OF_RANGE", "sqlState" : "42805", "messageParameters" : { "index" : "-1", @@ -455,7 +455,7 @@ select a, b, count(1) from data group by cube(1, 3) -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "GROUP_BY_POS_AGGREGATE", + "condition" : "GROUP_BY_POS_AGGREGATE", "sqlState" : "42903", "messageParameters" : { "aggExpr" : "count(1) AS `count(1)`", diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/group-by.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/group-by.sql.out index 304b382c7bbe..607b2401e853 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/group-by.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/group-by.sql.out @@ -17,7 +17,7 @@ SELECT a, COUNT(b) FROM testData -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "MISSING_GROUP_BY", + "condition" : "MISSING_GROUP_BY", "sqlState" : "42803", "queryContext" : [ { "objectType" : "", @@ -58,7 +58,7 @@ SELECT a, COUNT(b) FROM testData GROUP BY b -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "MISSING_AGGREGATION", + "condition" : "MISSING_AGGREGATION", "sqlState" : "42803", "messageParameters" : { "expression" : "\"a\"", @@ -147,7 +147,7 @@ SELECT a + 2, COUNT(b) FROM testData GROUP BY a + 1 -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "MISSING_AGGREGATION", + "condition" : "MISSING_AGGREGATION", "sqlState" : "42803", "messageParameters" : { "expression" : "\"a\"", @@ -232,7 +232,7 @@ SELECT a AS k, COUNT(non_existing) FROM testData GROUP BY k -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`non_existing`", @@ -253,7 +253,7 @@ SELECT COUNT(b) AS k FROM testData GROUP BY k -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "GROUP_BY_AGGREGATE", + "condition" : "GROUP_BY_AGGREGATE", "sqlState" : "42903", "messageParameters" : { "sqlExpr" : "count(testdata.b)" @@ -284,7 +284,7 @@ SELECT k AS a, COUNT(v) FROM testDataHasSameNameWithAlias GROUP BY a -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "MISSING_AGGREGATION", + "condition" : "MISSING_AGGREGATION", "sqlState" : "42803", "messageParameters" : { "expression" : "\"k\"", @@ -304,7 +304,7 @@ SELECT a AS k, COUNT(b) FROM testData GROUP BY k -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`k`", @@ -412,7 +412,7 @@ SELECT id FROM range(10) HAVING id > 0 -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "MISSING_GROUP_BY", + "condition" : "MISSING_GROUP_BY", "sqlState" : "42803", "queryContext" : [ { "objectType" : "", @@ -443,7 +443,7 @@ SELECT 1 FROM range(10) HAVING MAX(id) > 0 -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INVALID_WHERE_CONDITION", + "condition" : "INVALID_WHERE_CONDITION", "sqlState" : "42903", "messageParameters" : { "condition" : "\"(max(id) > 0)\"", @@ -630,7 +630,7 @@ SELECT every(1) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"1\"", @@ -654,7 +654,7 @@ SELECT some(1S) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"1\"", @@ -678,7 +678,7 @@ SELECT any(1L) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"1\"", @@ -709,7 +709,7 @@ SELECT bool_and(1.0) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"1.0\"", @@ -733,7 +733,7 @@ SELECT bool_or(1.0D) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"1.0\"", @@ -873,7 +873,7 @@ SELECT count(*) FROM test_agg WHERE count(*) > 1L -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INVALID_WHERE_CONDITION", + "condition" : "INVALID_WHERE_CONDITION", "sqlState" : "42903", "messageParameters" : { "condition" : "\"(count(1) > 1)\"", @@ -894,7 +894,7 @@ SELECT count(*) FROM test_agg WHERE count(*) + 1L > 1L -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INVALID_WHERE_CONDITION", + "condition" : "INVALID_WHERE_CONDITION", "sqlState" : "42903", "messageParameters" : { "condition" : "\"((count(1) + 1) > 1)\"", @@ -915,7 +915,7 @@ SELECT count(*) FROM test_agg WHERE k = 1 or k = 2 or count(*) + 1L > 1L or max( -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INVALID_WHERE_CONDITION", + "condition" : "INVALID_WHERE_CONDITION", "sqlState" : "42903", "messageParameters" : { "condition" : "\"(((k = 1) OR (k = 2)) OR (((count(1) + 1) > 1) OR (max(k) > 1)))\"", diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/grouping_set.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/grouping_set.sql.out index b73ee16c8bde..9c6283c8222a 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/grouping_set.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/grouping_set.sql.out @@ -161,7 +161,7 @@ SELECT a, b, c, count(d) FROM grouping GROUP BY WITH ROLLUP -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'ROLLUP'", @@ -175,7 +175,7 @@ SELECT a, b, c, count(d) FROM grouping GROUP BY WITH CUBE -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'CUBE'", @@ -189,7 +189,7 @@ SELECT c1 FROM (values (1,2), (3,2)) t(c1, c2) GROUP BY GROUPING SETS (()) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "MISSING_AGGREGATION", + "condition" : "MISSING_AGGREGATION", "sqlState" : "42803", "messageParameters" : { "expression" : "\"c1\"", diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/having.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/having.sql.out index 78cf1223da50..7592c51c7ecb 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/having.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/having.sql.out @@ -50,7 +50,7 @@ SELECT count(k) FROM hav GROUP BY v HAVING v = array(1) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"INT\"", diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/higher-order-functions.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/higher-order-functions.sql.out index c06d1e5534ae..0359deab7813 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/higher-order-functions.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/higher-order-functions.sql.out @@ -20,7 +20,7 @@ select upper(x -> x) as v -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_LAMBDA_FUNCTION_CALL.NON_HIGHER_ORDER_FUNCTION", + "condition" : "INVALID_LAMBDA_FUNCTION_CALL.NON_HIGHER_ORDER_FUNCTION", "sqlState" : "42K0D", "messageParameters" : { "class" : "org.apache.spark.sql.catalyst.expressions.Upper" @@ -40,7 +40,7 @@ select ceil(x -> x) as v -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_LAMBDA_FUNCTION_CALL.NON_HIGHER_ORDER_FUNCTION", + "condition" : "INVALID_LAMBDA_FUNCTION_CALL.NON_HIGHER_ORDER_FUNCTION", "sqlState" : "42K0D", "messageParameters" : { "class" : "org.apache.spark.sql.catalyst.expressions.CeilExpressionBuilder$" diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/hll.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/hll.sql.out index 23701fb092f7..e70ea53a085e 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/hll.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/hll.sql.out @@ -82,7 +82,7 @@ FROM VALUES (ARRAY(1, 2)), (ARRAY(3, 4)) tab(col) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"col\"", @@ -165,7 +165,7 @@ SELECT hll_union(1, 2) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"1\"", diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/identifier-clause.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/identifier-clause.sql.out index 87d537f94981..e79a549f8406 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/identifier-clause.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/identifier-clause.sql.out @@ -93,7 +93,7 @@ DELETE FROM IDENTIFIER('ta' || 'b') WHERE 1=0 -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "UNSUPPORTED_FEATURE.TABLE_OPERATION", + "condition" : "UNSUPPORTED_FEATURE.TABLE_OPERATION", "sqlState" : "0A000", "messageParameters" : { "operation" : "DELETE", @@ -107,7 +107,7 @@ UPDATE IDENTIFIER('ta' || 'b') SET c1 = 2 -- !query analysis org.apache.spark.SparkUnsupportedOperationException { - "errorClass" : "_LEGACY_ERROR_TEMP_2096", + "condition" : "_LEGACY_ERROR_TEMP_2096", "messageParameters" : { "ddl" : "UPDATE TABLE" } @@ -120,7 +120,7 @@ MERGE INTO IDENTIFIER('ta' || 'b') AS t USING IDENTIFIER('ta' || 'b') AS s ON s. -- !query analysis org.apache.spark.SparkUnsupportedOperationException { - "errorClass" : "_LEGACY_ERROR_TEMP_2096", + "condition" : "_LEGACY_ERROR_TEMP_2096", "messageParameters" : { "ddl" : "MERGE INTO TABLE" } @@ -252,7 +252,7 @@ REPLACE TABLE IDENTIFIER('identifier_clauses.' || 'tab')(c1 INT) USING CSV -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "UNSUPPORTED_FEATURE.TABLE_OPERATION", + "condition" : "UNSUPPORTED_FEATURE.TABLE_OPERATION", "sqlState" : "0A000", "messageParameters" : { "operation" : "REPLACE TABLE", @@ -369,7 +369,7 @@ REPAIR TABLE IDENTIFIER('ta' || 'b') -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "NOT_A_PARTITIONED_TABLE", + "condition" : "NOT_A_PARTITIONED_TABLE", "sqlState" : "42809", "messageParameters" : { "operation" : "MSCK REPAIR TABLE", @@ -613,7 +613,7 @@ SELECT IDENTIFIER('c 1') FROM VALUES(1) AS T(`c 1`) -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'1'", @@ -634,7 +634,7 @@ SELECT IDENTIFIER('') FROM VALUES(1) AS T(``) -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_EMPTY_STATEMENT", + "condition" : "PARSE_EMPTY_STATEMENT", "sqlState" : "42617", "queryContext" : [ { "objectType" : "", @@ -651,7 +651,7 @@ VALUES(IDENTIFIER(CAST(NULL AS STRING))) -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "NOT_A_CONSTANT_STRING.NULL", + "condition" : "NOT_A_CONSTANT_STRING.NULL", "sqlState" : "42601", "messageParameters" : { "expr" : "CAST(NULL AS STRING)", @@ -672,7 +672,7 @@ VALUES(IDENTIFIER(1)) -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "NOT_A_CONSTANT_STRING.WRONG_TYPE", + "condition" : "NOT_A_CONSTANT_STRING.WRONG_TYPE", "sqlState" : "42601", "messageParameters" : { "dataType" : "int", @@ -694,7 +694,7 @@ VALUES(IDENTIFIER(SUBSTR('HELLO', 1, RAND() + 1))) -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "NOT_A_CONSTANT_STRING.NOT_CONSTANT", + "condition" : "NOT_A_CONSTANT_STRING.NOT_CONSTANT", "sqlState" : "42601", "messageParameters" : { "expr" : "substr('HELLO', 1, CAST((rand() + CAST(1 AS DOUBLE)) AS INT))", @@ -715,7 +715,7 @@ SELECT `IDENTIFIER`('abs')(c1) FROM VALUES(-1) AS T(c1) -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "UNRESOLVED_ROUTINE", + "condition" : "UNRESOLVED_ROUTINE", "sqlState" : "42883", "messageParameters" : { "routineName" : "`IDENTIFIER`", @@ -736,7 +736,7 @@ CREATE TABLE IDENTIFIER(1)(c1 INT) USING csv -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "NOT_A_CONSTANT_STRING.WRONG_TYPE", + "condition" : "NOT_A_CONSTANT_STRING.WRONG_TYPE", "sqlState" : "42601", "messageParameters" : { "dataType" : "int", @@ -758,7 +758,7 @@ CREATE TABLE IDENTIFIER('a.b.c')(c1 INT) USING csv -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "REQUIRES_SINGLE_PART_NAMESPACE", + "condition" : "REQUIRES_SINGLE_PART_NAMESPACE", "sqlState" : "42K05", "messageParameters" : { "namespace" : "`a`.`b`", @@ -772,7 +772,7 @@ CREATE VIEW IDENTIFIER('a.b.c')(c1) AS VALUES(1) -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "REQUIRES_SINGLE_PART_NAMESPACE", + "condition" : "REQUIRES_SINGLE_PART_NAMESPACE", "sqlState" : "42K05", "messageParameters" : { "namespace" : "`a`.`b`", @@ -786,7 +786,7 @@ DROP TABLE IDENTIFIER('a.b.c') -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "REQUIRES_SINGLE_PART_NAMESPACE", + "condition" : "REQUIRES_SINGLE_PART_NAMESPACE", "sqlState" : "42K05", "messageParameters" : { "namespace" : "`a`.`b`", @@ -800,7 +800,7 @@ DROP VIEW IDENTIFIER('a.b.c') -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "REQUIRES_SINGLE_PART_NAMESPACE", + "condition" : "REQUIRES_SINGLE_PART_NAMESPACE", "sqlState" : "42K05", "messageParameters" : { "namespace" : "`a`.`b`", @@ -814,7 +814,7 @@ COMMENT ON TABLE IDENTIFIER('a.b.c.d') IS 'hello' -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "REQUIRES_SINGLE_PART_NAMESPACE", + "condition" : "REQUIRES_SINGLE_PART_NAMESPACE", "sqlState" : "42K05", "messageParameters" : { "namespace" : "`a`.`b`.`c`", @@ -828,7 +828,7 @@ VALUES(IDENTIFIER(1)()) -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "NOT_A_CONSTANT_STRING.WRONG_TYPE", + "condition" : "NOT_A_CONSTANT_STRING.WRONG_TYPE", "sqlState" : "42601", "messageParameters" : { "dataType" : "int", @@ -850,7 +850,7 @@ VALUES(IDENTIFIER('a.b.c.d')()) -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "IDENTIFIER_TOO_MANY_NAME_PARTS", + "condition" : "IDENTIFIER_TOO_MANY_NAME_PARTS", "sqlState" : "42601", "messageParameters" : { "identifier" : "`a`.`b`.`c`.`d`" @@ -870,7 +870,7 @@ CREATE TEMPORARY FUNCTION IDENTIFIER('default.my' || 'DoubleAvg') AS 'test.org.a -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_SQL_SYNTAX.CREATE_TEMP_FUNC_WITH_DATABASE", + "condition" : "INVALID_SQL_SYNTAX.CREATE_TEMP_FUNC_WITH_DATABASE", "sqlState" : "42000", "messageParameters" : { "database" : "`default`" @@ -890,7 +890,7 @@ DROP TEMPORARY FUNCTION IDENTIFIER('default.my' || 'DoubleAvg') -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_SQL_SYNTAX.MULTI_PART_NAME", + "condition" : "INVALID_SQL_SYNTAX.MULTI_PART_NAME", "sqlState" : "42000", "messageParameters" : { "name" : "`default`.`myDoubleAvg`", @@ -911,7 +911,7 @@ CREATE TEMPORARY VIEW IDENTIFIER('default.v')(c1) AS VALUES(1) -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "TEMP_VIEW_NAME_TOO_MANY_NAME_PARTS", + "condition" : "TEMP_VIEW_NAME_TOO_MANY_NAME_PARTS", "sqlState" : "428EK", "messageParameters" : { "actualName" : "`default`.`v`" @@ -1063,7 +1063,7 @@ SELECT row_number() OVER IDENTIFIER('x.win') FROM VALUES(1) AS T(c1) WINDOW win -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "''x.win''", @@ -1077,7 +1077,7 @@ SELECT T1.c1 FROM VALUES(1) AS T1(c1) JOIN VALUES(1) AS T2(c1) USING (IDENTIFIER -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'('", @@ -1091,7 +1091,7 @@ SELECT IDENTIFIER('t').c1 FROM VALUES(1) AS T(c1) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`t`", @@ -1112,7 +1112,7 @@ SELECT map('a', 1).IDENTIFIER('a') FROM VALUES(1) AS T(c1) -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "''a''", @@ -1126,7 +1126,7 @@ SELECT named_struct('a', 1).IDENTIFIER('a') FROM VALUES(1) AS T(c1) -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "''a''", @@ -1140,7 +1140,7 @@ SELECT * FROM s.IDENTIFIER('tab') -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_SQL_SYNTAX.INVALID_TABLE_VALUED_FUNC_NAME", + "condition" : "INVALID_SQL_SYNTAX.INVALID_TABLE_VALUED_FUNC_NAME", "sqlState" : "42000", "messageParameters" : { "funcName" : "`s`.`IDENTIFIER`" @@ -1160,7 +1160,7 @@ SELECT * FROM IDENTIFIER('s').IDENTIFIER('tab') -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'.'", @@ -1174,7 +1174,7 @@ SELECT * FROM IDENTIFIER('s').tab -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'.'", @@ -1188,7 +1188,7 @@ SELECT row_number() OVER IDENTIFIER('win') FROM VALUES(1) AS T(c1) WINDOW win AS -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "''win''", @@ -1202,7 +1202,7 @@ SELECT row_number() OVER win FROM VALUES(1) AS T(c1) WINDOW IDENTIFIER('win') AS -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'WINDOW'", @@ -1216,7 +1216,7 @@ WITH identifier('v')(identifier('c1')) AS (VALUES(1)) (SELECT c1 FROM v) -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "''v''", @@ -1230,7 +1230,7 @@ INSERT INTO tab(IDENTIFIER('c1')) VALUES(1) -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'('", @@ -1244,7 +1244,7 @@ CREATE OR REPLACE VIEW v(IDENTIFIER('c1')) AS VALUES(1) -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'('", @@ -1258,7 +1258,7 @@ CREATE TABLE tab(IDENTIFIER('c1') INT) USING CSV -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'('", diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/ilike-all.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/ilike-all.sql.out index cf9c2e12cc72..988dcc223285 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/ilike-all.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/ilike-all.sql.out @@ -199,7 +199,7 @@ SELECT company FROM ilike_any_table WHERE company ILIKE ALL () -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0064", + "condition" : "_LEGACY_ERROR_TEMP_0064", "messageParameters" : { "msg" : "Expected something between '(' and ')'." }, diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/ilike-any.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/ilike-any.sql.out index e20000f1463d..c738144040f8 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/ilike-any.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/ilike-any.sql.out @@ -199,7 +199,7 @@ SELECT company FROM ilike_any_table WHERE company ILIKE ANY () -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0064", + "condition" : "_LEGACY_ERROR_TEMP_0064", "messageParameters" : { "msg" : "Expected something between '(' and ')'." }, diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/inline-table.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/inline-table.sql.out index 78539effe188..6b3377324090 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/inline-table.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/inline-table.sql.out @@ -97,7 +97,7 @@ select * from values ("one", rand(5)), ("two", 3.0D) as data(a, b) -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_INLINE_TABLE.CANNOT_EVALUATE_EXPRESSION_IN_INLINE_TABLE", + "condition" : "INVALID_INLINE_TABLE.CANNOT_EVALUATE_EXPRESSION_IN_INLINE_TABLE", "sqlState" : "42000", "messageParameters" : { "expr" : "\"rand(5)\"" @@ -117,7 +117,7 @@ select * from values ("one", 2.0), ("two") as data(a, b) -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_INLINE_TABLE.NUM_COLUMNS_MISMATCH", + "condition" : "INVALID_INLINE_TABLE.NUM_COLUMNS_MISMATCH", "sqlState" : "42000", "messageParameters" : { "actualNumCols" : "1", @@ -139,7 +139,7 @@ select * from values ("one", array(0, 1)), ("two", struct(1, 2)) as data(a, b) -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_INLINE_TABLE.INCOMPATIBLE_TYPES_IN_INLINE_TABLE", + "condition" : "INVALID_INLINE_TABLE.INCOMPATIBLE_TYPES_IN_INLINE_TABLE", "sqlState" : "42000", "messageParameters" : { "colName" : "`b`" @@ -159,7 +159,7 @@ select * from values ("one"), ("two") as data(a, b) -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_INLINE_TABLE.NUM_COLUMNS_MISMATCH", + "condition" : "INVALID_INLINE_TABLE.NUM_COLUMNS_MISMATCH", "sqlState" : "42000", "messageParameters" : { "actualNumCols" : "1", @@ -181,7 +181,7 @@ select * from values ("one", random_not_exist_func(1)), ("two", 2) as data(a, b) -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "UNRESOLVED_ROUTINE", + "condition" : "UNRESOLVED_ROUTINE", "sqlState" : "42883", "messageParameters" : { "routineName" : "`random_not_exist_func`", @@ -202,7 +202,7 @@ select * from values ("one", count(1)), ("two", 2) as data(a, b) -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_INLINE_TABLE.CANNOT_EVALUATE_EXPRESSION_IN_INLINE_TABLE", + "condition" : "INVALID_INLINE_TABLE.CANNOT_EVALUATE_EXPRESSION_IN_INLINE_TABLE", "sqlState" : "42000", "messageParameters" : { "expr" : "\"count(1)\"" diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/intersect-all.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/intersect-all.sql.out index 69b4001ff348..b0a2e41ddd15 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/intersect-all.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/intersect-all.sql.out @@ -163,7 +163,7 @@ SELECT array(1), 2 -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -190,7 +190,7 @@ SELECT k, v FROM tab2 -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "NUM_COLUMNS_MISMATCH", + "condition" : "NUM_COLUMNS_MISMATCH", "sqlState" : "42826", "messageParameters" : { "firstNumColumns" : "1", diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/interval.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/interval.sql.out index c023e3b56f11..91d96cd61ee9 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/interval.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/interval.sql.out @@ -10,7 +10,7 @@ select interval 4 month 2 weeks 3 microseconds * 1.5 -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0029", + "condition" : "_LEGACY_ERROR_TEMP_0029", "messageParameters" : { "literal" : "interval 4 month 2 weeks 3 microseconds" }, @@ -161,7 +161,7 @@ select '2' / interval 2 second -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"STRING\"", @@ -183,7 +183,7 @@ select '2' / interval 2 year -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"STRING\"", @@ -261,7 +261,7 @@ select 2 / interval '2' year -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"INT\"", @@ -283,7 +283,7 @@ select 2 / interval '2' hour -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"INT\"", @@ -305,7 +305,7 @@ select null / interval '2' year -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"VOID\"", @@ -327,7 +327,7 @@ select null / interval '2' hour -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"VOID\"", @@ -349,7 +349,7 @@ select -interval '-1 month 1 day -1 second' -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0029", + "condition" : "_LEGACY_ERROR_TEMP_0029", "messageParameters" : { "literal" : "interval '-1 month 1 day -1 second'" }, @@ -382,7 +382,7 @@ select -interval -1 month 1 day -1 second -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0029", + "condition" : "_LEGACY_ERROR_TEMP_0029", "messageParameters" : { "literal" : "interval -1 month 1 day -1 second" }, @@ -415,7 +415,7 @@ select +interval '-1 month 1 day -1 second' -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0029", + "condition" : "_LEGACY_ERROR_TEMP_0029", "messageParameters" : { "literal" : "interval '-1 month 1 day -1 second'" }, @@ -448,7 +448,7 @@ select +interval -1 month 1 day -1 second -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0029", + "condition" : "_LEGACY_ERROR_TEMP_0029", "messageParameters" : { "literal" : "interval -1 month 1 day -1 second" }, @@ -817,7 +817,7 @@ select interval 1 year 2 month 3 week 4 day 5 hour 6 minute 7 seconds 8 millisec -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0029", + "condition" : "_LEGACY_ERROR_TEMP_0029", "messageParameters" : { "literal" : "interval 1 year 2 month 3 week 4 day 5 hour 6 minute 7 seconds 8 millisecond 9 microsecond" }, @@ -857,7 +857,7 @@ select interval '30' year '25' month '-100' day '40' hour '80' minute '299.88998 -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0029", + "condition" : "_LEGACY_ERROR_TEMP_0029", "messageParameters" : { "literal" : "interval '30' year '25' month '-100' day '40' hour '80' minute '299.889987299' second" }, @@ -981,7 +981,7 @@ select interval '20 15:40:32.99899999' day to hour -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_INTERVAL_FORMAT.UNMATCHED_FORMAT_STRING_WITH_NOTICE", + "condition" : "INVALID_INTERVAL_FORMAT.UNMATCHED_FORMAT_STRING_WITH_NOTICE", "sqlState" : "22006", "messageParameters" : { "input" : "20 15:40:32.99899999", @@ -1004,7 +1004,7 @@ select interval '20 15:40:32.99899999' day to minute -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_INTERVAL_FORMAT.UNMATCHED_FORMAT_STRING_WITH_NOTICE", + "condition" : "INVALID_INTERVAL_FORMAT.UNMATCHED_FORMAT_STRING_WITH_NOTICE", "sqlState" : "22006", "messageParameters" : { "input" : "20 15:40:32.99899999", @@ -1027,7 +1027,7 @@ select interval '15:40:32.99899999' hour to minute -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_INTERVAL_FORMAT.UNMATCHED_FORMAT_STRING_WITH_NOTICE", + "condition" : "INVALID_INTERVAL_FORMAT.UNMATCHED_FORMAT_STRING_WITH_NOTICE", "sqlState" : "22006", "messageParameters" : { "input" : "15:40:32.99899999", @@ -1050,7 +1050,7 @@ select interval '15:40.99899999' hour to second -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_INTERVAL_FORMAT.UNMATCHED_FORMAT_STRING_WITH_NOTICE", + "condition" : "INVALID_INTERVAL_FORMAT.UNMATCHED_FORMAT_STRING_WITH_NOTICE", "sqlState" : "22006", "messageParameters" : { "input" : "15:40.99899999", @@ -1073,7 +1073,7 @@ select interval '15:40' hour to second -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_INTERVAL_FORMAT.UNMATCHED_FORMAT_STRING_WITH_NOTICE", + "condition" : "INVALID_INTERVAL_FORMAT.UNMATCHED_FORMAT_STRING_WITH_NOTICE", "sqlState" : "22006", "messageParameters" : { "input" : "15:40", @@ -1096,7 +1096,7 @@ select interval '20 40:32.99899999' minute to second -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_INTERVAL_FORMAT.UNMATCHED_FORMAT_STRING_WITH_NOTICE", + "condition" : "INVALID_INTERVAL_FORMAT.UNMATCHED_FORMAT_STRING_WITH_NOTICE", "sqlState" : "22006", "messageParameters" : { "input" : "20 40:32.99899999", @@ -1119,7 +1119,7 @@ select interval 10 nanoseconds -- !query analysis org.apache.spark.SparkIllegalArgumentException { - "errorClass" : "INVALID_INTERVAL_FORMAT.INVALID_UNIT", + "condition" : "INVALID_INTERVAL_FORMAT.INVALID_UNIT", "sqlState" : "22006", "messageParameters" : { "input" : " 10 nanoseconds", @@ -1203,7 +1203,7 @@ select interval 1 fake_unit -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'fake_unit'", @@ -1217,7 +1217,7 @@ select interval 1 year to month -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0027", + "condition" : "_LEGACY_ERROR_TEMP_0027", "queryContext" : [ { "objectType" : "", "objectName" : "", @@ -1233,7 +1233,7 @@ select interval '1' year to second -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_INTERVAL_FORMAT.UNSUPPORTED_FROM_TO_EXPRESSION", + "condition" : "INVALID_INTERVAL_FORMAT.UNSUPPORTED_FROM_TO_EXPRESSION", "sqlState" : "22006", "messageParameters" : { "from" : "year", @@ -1255,7 +1255,7 @@ select interval '10-9' year to month '2-1' year to month -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0024", + "condition" : "_LEGACY_ERROR_TEMP_0024", "queryContext" : [ { "objectType" : "", "objectName" : "", @@ -1271,7 +1271,7 @@ select interval '10-9' year to month '12:11:10' hour to second -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0024", + "condition" : "_LEGACY_ERROR_TEMP_0024", "queryContext" : [ { "objectType" : "", "objectName" : "", @@ -1287,7 +1287,7 @@ select interval '1 15:11' day to minute '12:11:10' hour to second -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0024", + "condition" : "_LEGACY_ERROR_TEMP_0024", "queryContext" : [ { "objectType" : "", "objectName" : "", @@ -1303,7 +1303,7 @@ select interval 1 year '2-1' year to month -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0024", + "condition" : "_LEGACY_ERROR_TEMP_0024", "queryContext" : [ { "objectType" : "", "objectName" : "", @@ -1319,7 +1319,7 @@ select interval 1 year '12:11:10' hour to second -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0024", + "condition" : "_LEGACY_ERROR_TEMP_0024", "queryContext" : [ { "objectType" : "", "objectName" : "", @@ -1335,7 +1335,7 @@ select interval '10-9' year to month '1' year -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0024", + "condition" : "_LEGACY_ERROR_TEMP_0024", "queryContext" : [ { "objectType" : "", "objectName" : "", @@ -1351,7 +1351,7 @@ select interval '12:11:10' hour to second '1' year -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0024", + "condition" : "_LEGACY_ERROR_TEMP_0024", "queryContext" : [ { "objectType" : "", "objectName" : "", @@ -1367,7 +1367,7 @@ select interval (-30) day -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "UNRESOLVED_ROUTINE", + "condition" : "UNRESOLVED_ROUTINE", "sqlState" : "42883", "messageParameters" : { "routineName" : "`interval`", @@ -1388,7 +1388,7 @@ select interval (a + 1) day -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "UNRESOLVED_ROUTINE", + "condition" : "UNRESOLVED_ROUTINE", "sqlState" : "42883", "messageParameters" : { "routineName" : "`interval`", @@ -1409,7 +1409,7 @@ select interval 30 day day day -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'day'", @@ -1423,7 +1423,7 @@ select interval (-30) days -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "UNRESOLVED_ROUTINE", + "condition" : "UNRESOLVED_ROUTINE", "sqlState" : "42883", "messageParameters" : { "routineName" : "`interval`", @@ -1444,7 +1444,7 @@ select interval (a + 1) days -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "UNRESOLVED_ROUTINE", + "condition" : "UNRESOLVED_ROUTINE", "sqlState" : "42883", "messageParameters" : { "routineName" : "`interval`", @@ -1465,7 +1465,7 @@ select interval 30 days days days -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'days'", @@ -1486,7 +1486,7 @@ SELECT INTERVAL '178956970-8' YEAR TO MONTH -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_INTERVAL_FORMAT.INTERVAL_PARSING", + "condition" : "INVALID_INTERVAL_FORMAT.INTERVAL_PARSING", "sqlState" : "22006", "messageParameters" : { "input" : "178956970-8", @@ -1540,7 +1540,7 @@ select -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"INTERVAL YEAR\"", @@ -1577,7 +1577,7 @@ select interval '2' year + '3-3' -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"INTERVAL YEAR\"", @@ -1599,7 +1599,7 @@ select interval '2' year - '4' -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"INTERVAL YEAR\"", @@ -1643,7 +1643,7 @@ select interval '2' year + str from interval_view -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"INTERVAL YEAR\"", @@ -1665,7 +1665,7 @@ select interval '2' year - str from interval_view -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"INTERVAL YEAR\"", @@ -1709,7 +1709,7 @@ select interval '2-2' year to month + interval '3' day -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"INTERVAL '2-2' YEAR TO MONTH\"", @@ -1733,7 +1733,7 @@ select interval '3' day + interval '2-2' year to month -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"INTERVAL '2-2' YEAR TO MONTH\"", @@ -1757,7 +1757,7 @@ select interval '2-2' year to month - interval '3' day -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"INTERVAL '2-2' YEAR TO MONTH\"", @@ -1781,7 +1781,7 @@ select interval '3' day - interval '2-2' year to month -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"INTERVAL DAY\"", @@ -1803,7 +1803,7 @@ select 1 - interval '2' second -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"1\"", @@ -1827,7 +1827,7 @@ select 1 + interval '2' month -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"INT\"", @@ -1849,7 +1849,7 @@ select interval '2' second + 1 -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"1\"", @@ -1873,7 +1873,7 @@ select interval '2' month - 1 -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"INTERVAL MONTH\"", @@ -1937,7 +1937,7 @@ select interval '-\t2-2\t' year to month -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_INTERVAL_FORMAT.UNMATCHED_FORMAT_STRING", + "condition" : "INVALID_INTERVAL_FORMAT.UNMATCHED_FORMAT_STRING", "sqlState" : "22006", "messageParameters" : { "input" : "-\t2-2\t", @@ -1967,7 +1967,7 @@ select interval '\n-\t10\t 12:34:46.789\t' day to second -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_INTERVAL_FORMAT.UNMATCHED_FORMAT_STRING_WITH_NOTICE", + "condition" : "INVALID_INTERVAL_FORMAT.UNMATCHED_FORMAT_STRING_WITH_NOTICE", "sqlState" : "22006", "messageParameters" : { "input" : "\n-\t10\t 12:34:46.789\t", @@ -1990,7 +1990,7 @@ select interval '中文 interval 1 day' -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_TYPED_LITERAL", + "condition" : "INVALID_TYPED_LITERAL", "sqlState" : "42604", "messageParameters" : { "value" : "'中文 interval 1 day'", @@ -2011,7 +2011,7 @@ select interval 'interval中文 1 day' -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_TYPED_LITERAL", + "condition" : "INVALID_TYPED_LITERAL", "sqlState" : "42604", "messageParameters" : { "value" : "'interval中文 1 day'", @@ -2032,7 +2032,7 @@ select interval 'interval 1中文day' -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_TYPED_LITERAL", + "condition" : "INVALID_TYPED_LITERAL", "sqlState" : "42604", "messageParameters" : { "value" : "'interval 1中文day'", @@ -2141,7 +2141,7 @@ select interval '+' -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_TYPED_LITERAL", + "condition" : "INVALID_TYPED_LITERAL", "sqlState" : "42604", "messageParameters" : { "value" : "'+'", @@ -2162,7 +2162,7 @@ select interval '+.' -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_TYPED_LITERAL", + "condition" : "INVALID_TYPED_LITERAL", "sqlState" : "42604", "messageParameters" : { "value" : "'+.'", @@ -2183,7 +2183,7 @@ select interval '1' -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_TYPED_LITERAL", + "condition" : "INVALID_TYPED_LITERAL", "sqlState" : "42604", "messageParameters" : { "value" : "'1'", @@ -2204,7 +2204,7 @@ select interval '1.2' -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_TYPED_LITERAL", + "condition" : "INVALID_TYPED_LITERAL", "sqlState" : "42604", "messageParameters" : { "value" : "'1.2'", @@ -2225,7 +2225,7 @@ select interval '- 2' -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_TYPED_LITERAL", + "condition" : "INVALID_TYPED_LITERAL", "sqlState" : "42604", "messageParameters" : { "value" : "'- 2'", @@ -2246,7 +2246,7 @@ select interval '1 day -' -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_TYPED_LITERAL", + "condition" : "INVALID_TYPED_LITERAL", "sqlState" : "42604", "messageParameters" : { "value" : "'1 day -'", @@ -2267,7 +2267,7 @@ select interval '1 day 1' -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_TYPED_LITERAL", + "condition" : "INVALID_TYPED_LITERAL", "sqlState" : "42604", "messageParameters" : { "value" : "'1 day 1'", @@ -2288,7 +2288,7 @@ select interval '1 day 2' day -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0026", + "condition" : "_LEGACY_ERROR_TEMP_0026", "messageParameters" : { "value" : "1 day 2" }, @@ -2307,7 +2307,7 @@ select interval 'interval 1' day -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0026", + "condition" : "_LEGACY_ERROR_TEMP_0026", "messageParameters" : { "value" : "interval 1" }, @@ -2501,7 +2501,7 @@ SELECT INTERVAL '106751992 04' DAY TO HOUR -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0063", + "condition" : "_LEGACY_ERROR_TEMP_0063", "messageParameters" : { "msg" : "requirement failed: day 106751992 outside range [0, 106751991]" }, @@ -2520,7 +2520,7 @@ SELECT INTERVAL '-106751992 04' DAY TO HOUR -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0063", + "condition" : "_LEGACY_ERROR_TEMP_0063", "messageParameters" : { "msg" : "requirement failed: day 106751992 outside range [0, 106751991]" }, @@ -2539,7 +2539,7 @@ SELECT INTERVAL '2562047789:00' HOUR TO MINUTE -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0063", + "condition" : "_LEGACY_ERROR_TEMP_0063", "messageParameters" : { "msg" : "requirement failed: hour 2562047789 outside range [0, 2562047788]" }, @@ -2558,7 +2558,7 @@ SELECT INTERVAL '-2562047789:00' HOUR TO MINUTE -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0063", + "condition" : "_LEGACY_ERROR_TEMP_0063", "messageParameters" : { "msg" : "requirement failed: hour 2562047789 outside range [0, 2562047788]" }, @@ -2577,7 +2577,7 @@ SELECT INTERVAL '153722867281:54.775808' MINUTE TO SECOND -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0063", + "condition" : "_LEGACY_ERROR_TEMP_0063", "messageParameters" : { "msg" : "requirement failed: minute 153722867281 outside range [0, 153722867280]" }, @@ -2596,7 +2596,7 @@ SELECT INTERVAL '-153722867281:54.775808' MINUTE TO SECOND -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0063", + "condition" : "_LEGACY_ERROR_TEMP_0063", "messageParameters" : { "msg" : "requirement failed: minute 153722867281 outside range [0, 153722867280]" }, @@ -2727,7 +2727,7 @@ SELECT INTERVAL 1 MONTH > INTERVAL 20 DAYS -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"INTERVAL MONTH\"", @@ -2749,7 +2749,7 @@ SELECT INTERVAL '1' DAY < '1' -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"INTERVAL DAY\"", @@ -2771,7 +2771,7 @@ SELECT INTERVAL '1' DAY = '1' -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"INTERVAL DAY\"", @@ -2793,7 +2793,7 @@ SELECT INTERVAL '1' DAY > '1' -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"INTERVAL DAY\"", @@ -2815,7 +2815,7 @@ SELECT '1' < INTERVAL '1' DAY -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"STRING\"", @@ -2837,7 +2837,7 @@ SELECT '1' = INTERVAL '1' DAY -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"STRING\"", @@ -2859,7 +2859,7 @@ SELECT '1' > INTERVAL '1' DAY -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"STRING\"", @@ -2881,7 +2881,7 @@ SELECT INTERVAL '1' YEAR < '1' -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"INTERVAL YEAR\"", @@ -2903,7 +2903,7 @@ SELECT INTERVAL '1' YEAR = '1' -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"INTERVAL YEAR\"", @@ -2925,7 +2925,7 @@ SELECT INTERVAL '1' YEAR > '1' -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"INTERVAL YEAR\"", @@ -2947,7 +2947,7 @@ SELECT '1' < INTERVAL '1' YEAR -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"STRING\"", @@ -2969,7 +2969,7 @@ SELECT '1' = INTERVAL '1' YEAR -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"STRING\"", @@ -2991,7 +2991,7 @@ SELECT '1' > INTERVAL '1' YEAR -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"STRING\"", @@ -3027,7 +3027,7 @@ SELECT array(INTERVAL 1 MONTH, INTERVAL 20 DAYS) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "(\"INTERVAL MONTH\" or \"INTERVAL DAY\")", @@ -3063,7 +3063,7 @@ SELECT coalesce(INTERVAL 1 MONTH, INTERVAL 20 DAYS) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "(\"INTERVAL MONTH\" or \"INTERVAL DAY\")", @@ -3127,7 +3127,7 @@ SELECT div(INTERVAL '1' MONTH, INTERVAL '-1' DAY) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"INTERVAL MONTH\"", diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/join-lateral.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/join-lateral.sql.out index 390fcf5e3015..6dfbf13ce359 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/join-lateral.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/join-lateral.sql.out @@ -177,7 +177,7 @@ SELECT * FROM t1, LATERAL (SELECT t1.*, t2.* FROM t2, LATERAL (SELECT t1.*, t2.* -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "CANNOT_RESOLVE_STAR_EXPAND", + "condition" : "CANNOT_RESOLVE_STAR_EXPAND", "sqlState" : "42704", "messageParameters" : { "columns" : "`c1`, `c2`", @@ -240,7 +240,7 @@ SELECT * FROM t1 NATURAL JOIN LATERAL (SELECT c1 + c2 AS c2) -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INCOMPATIBLE_JOIN_TYPES", + "condition" : "INCOMPATIBLE_JOIN_TYPES", "sqlState" : "42613", "messageParameters" : { "joinType1" : "LATERAL", @@ -261,7 +261,7 @@ SELECT * FROM t1 JOIN LATERAL (SELECT c1 + c2 AS c2) USING (c2) -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "UNSUPPORTED_FEATURE.LATERAL_JOIN_USING", + "condition" : "UNSUPPORTED_FEATURE.LATERAL_JOIN_USING", "sqlState" : "0A000", "queryContext" : [ { "objectType" : "", @@ -446,7 +446,7 @@ SELECT * FROM t1 JOIN LATERAL (SELECT t1.c1 AS a, t2.c1 AS b) s JOIN t2 ON s.b = -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITHOUT_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITHOUT_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`t2`.`c1`" @@ -478,7 +478,7 @@ SELECT * FROM t1, LATERAL (SELECT c1 + c2 + rand(0) AS c3) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNSUPPORTED_SUBQUERY_EXPRESSION_CATEGORY.NON_DETERMINISTIC_LATERAL_SUBQUERIES", + "condition" : "UNSUPPORTED_SUBQUERY_EXPRESSION_CATEGORY.NON_DETERMINISTIC_LATERAL_SUBQUERIES", "sqlState" : "0A000", "messageParameters" : { "treeNode" : "LateralJoin lateral-subquery#x [c1#x && c2#x], Inner\n: +- SubqueryAlias __auto_generated_subquery_name\n: +- Project [(cast((outer(c1#x) + outer(c2#x)) as double) + rand(number)) AS c3#x]\n: +- OneRowRelation\n+- SubqueryAlias spark_catalog.default.t1\n +- View (`spark_catalog`.`default`.`t1`, [c1#x, c2#x])\n +- Project [cast(col1#x as int) AS c1#x, cast(col2#x as int) AS c2#x]\n +- LocalRelation [col1#x, col2#x]\n" @@ -498,7 +498,7 @@ SELECT * FROM t1, LATERAL (SELECT rand(0) FROM t2) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNSUPPORTED_SUBQUERY_EXPRESSION_CATEGORY.NON_DETERMINISTIC_LATERAL_SUBQUERIES", + "condition" : "UNSUPPORTED_SUBQUERY_EXPRESSION_CATEGORY.NON_DETERMINISTIC_LATERAL_SUBQUERIES", "sqlState" : "0A000", "messageParameters" : { "treeNode" : "LateralJoin lateral-subquery#x [], Inner\n: +- SubqueryAlias __auto_generated_subquery_name\n: +- Project [rand(number) AS rand(number)#x]\n: +- SubqueryAlias spark_catalog.default.t2\n: +- View (`spark_catalog`.`default`.`t2`, [c1#x, c2#x])\n: +- Project [cast(col1#x as int) AS c1#x, cast(col2#x as int) AS c2#x]\n: +- LocalRelation [col1#x, col2#x]\n+- SubqueryAlias spark_catalog.default.t1\n +- View (`spark_catalog`.`default`.`t1`, [c1#x, c2#x])\n +- Project [cast(col1#x as int) AS c1#x, cast(col2#x as int) AS c2#x]\n +- LocalRelation [col1#x, col2#x]\n" @@ -518,7 +518,7 @@ SELECT * FROM t1 JOIN LATERAL (SELECT * FROM t2) s ON t1.c1 + rand(0) = s.c1 -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNSUPPORTED_SUBQUERY_EXPRESSION_CATEGORY.LATERAL_JOIN_CONDITION_NON_DETERMINISTIC", + "condition" : "UNSUPPORTED_SUBQUERY_EXPRESSION_CATEGORY.LATERAL_JOIN_CONDITION_NON_DETERMINISTIC", "sqlState" : "0A000", "messageParameters" : { "condition" : "((CAST(spark_catalog.default.t1.c1 AS DOUBLE) + rand(0)) = CAST(s.c1 AS DOUBLE))" @@ -658,7 +658,7 @@ SELECT * FROM t1, LATERAL (SELECT * FROM t2, LATERAL (SELECT t1.c1 + t2.c1)) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITHOUT_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITHOUT_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`t1`.`c1`" @@ -678,7 +678,7 @@ SELECT * FROM t1, LATERAL (SELECT * FROM (SELECT c1), LATERAL (SELECT c2)) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITHOUT_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITHOUT_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`c2`" @@ -739,7 +739,7 @@ SELECT * FROM t1, LATERAL (SELECT c1, (SELECT SUM(c2) FROM t2 WHERE c1 = t1.c1)) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`t1`.`c1`", @@ -1985,7 +1985,7 @@ SELECT * FROM t1 JOIN LATERAL -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNSUPPORTED_SUBQUERY_EXPRESSION_CATEGORY.CORRELATED_REFERENCE", + "condition" : "UNSUPPORTED_SUBQUERY_EXPRESSION_CATEGORY.CORRELATED_REFERENCE", "sqlState" : "0A000", "messageParameters" : { "sqlExprs" : "\"(c1 - c1)\",\"(c1 - c1) AS `(c1 - outer(spark_catalog.default.t1.c1))`\"" @@ -2287,7 +2287,7 @@ SELECT * FROM t1 JOIN LATERAL (SELECT sum(c1) FROM -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNSUPPORTED_SUBQUERY_EXPRESSION_CATEGORY.ACCESSING_OUTER_QUERY_COLUMN_IS_NOT_ALLOWED", + "condition" : "UNSUPPORTED_SUBQUERY_EXPRESSION_CATEGORY.ACCESSING_OUTER_QUERY_COLUMN_IS_NOT_ALLOWED", "sqlState" : "0A000", "messageParameters" : { "treeNode" : "Filter (c1#x > outer(c2#x))\n+- SubqueryAlias spark_catalog.default.t4\n +- View (`spark_catalog`.`default`.`t4`, [c1#x, c2#x])\n +- Project [cast(col1#x as int) AS c1#x, cast(col2#x as int) AS c2#x]\n +- LocalRelation [col1#x, col2#x]\n" @@ -2844,7 +2844,7 @@ SELECT t.* FROM t1, LATERAL stack(c1, c2) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.NON_FOLDABLE_INPUT", + "condition" : "DATATYPE_MISMATCH.NON_FOLDABLE_INPUT", "sqlState" : "42K09", "messageParameters" : { "inputExpr" : "\"outer(spark_catalog.default.t1.c1)\"", diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/json-functions.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/json-functions.sql.out index 842b190c5a75..185728ddc46f 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/json-functions.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/json-functions.sql.out @@ -60,7 +60,7 @@ select to_json(named_struct('a', 1, 'b', 2), named_struct('mode', 'PERMISSIVE')) -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_OPTIONS.NON_MAP_FUNCTION", + "condition" : "INVALID_OPTIONS.NON_MAP_FUNCTION", "sqlState" : "42K06", "queryContext" : [ { "objectType" : "", @@ -77,7 +77,7 @@ select to_json(named_struct('a', 1, 'b', 2), map('mode', 1)) -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_OPTIONS.NON_STRING_TYPE", + "condition" : "INVALID_OPTIONS.NON_STRING_TYPE", "sqlState" : "42K06", "messageParameters" : { "mapType" : "\"MAP\"" @@ -97,7 +97,7 @@ select to_json() -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", + "condition" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", "sqlState" : "42605", "messageParameters" : { "actualNum" : "0", @@ -134,7 +134,7 @@ select from_json('{"a":1}', 1) -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_SCHEMA.NON_STRING_LITERAL", + "condition" : "INVALID_SCHEMA.NON_STRING_LITERAL", "sqlState" : "42K07", "messageParameters" : { "inputSchema" : "\"1\"" @@ -154,7 +154,7 @@ select from_json('{"a":1}', 'a InvalidType') -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'InvalidType'", @@ -175,7 +175,7 @@ select from_json('{"a":1}', 'a INT', named_struct('mode', 'PERMISSIVE')) -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_OPTIONS.NON_MAP_FUNCTION", + "condition" : "INVALID_OPTIONS.NON_MAP_FUNCTION", "sqlState" : "42K06", "queryContext" : [ { "objectType" : "", @@ -192,7 +192,7 @@ select from_json('{"a":1}', 'a INT', map('mode', 1)) -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_OPTIONS.NON_STRING_TYPE", + "condition" : "INVALID_OPTIONS.NON_STRING_TYPE", "sqlState" : "42K06", "messageParameters" : { "mapType" : "\"MAP\"" @@ -212,7 +212,7 @@ select from_json() -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", + "condition" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", "sqlState" : "42605", "messageParameters" : { "actualNum" : "0", @@ -430,7 +430,7 @@ select schema_of_json(null) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_NULL", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_NULL", "sqlState" : "42K09", "messageParameters" : { "exprName" : "json", @@ -459,7 +459,7 @@ SELECT schema_of_json(jsonField) FROM jsonTable -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.NON_FOLDABLE_INPUT", + "condition" : "DATATYPE_MISMATCH.NON_FOLDABLE_INPUT", "sqlState" : "42K09", "messageParameters" : { "inputExpr" : "\"jsonField\"", @@ -489,7 +489,7 @@ select json_array_length(2) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"2\"", @@ -513,7 +513,7 @@ select json_array_length() -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", + "condition" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", "sqlState" : "42605", "messageParameters" : { "actualNum" : "0", @@ -592,7 +592,7 @@ select json_object_keys() -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", + "condition" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", "sqlState" : "42605", "messageParameters" : { "actualNum" : "0", @@ -622,7 +622,7 @@ select json_object_keys(200) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"200\"", diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/like-all.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/like-all.sql.out index dab3e27be69a..618d1fddf0d4 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/like-all.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/like-all.sql.out @@ -199,7 +199,7 @@ SELECT company FROM like_all_table WHERE company LIKE ALL () -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0064", + "condition" : "_LEGACY_ERROR_TEMP_0064", "messageParameters" : { "msg" : "Expected something between '(' and ')'." }, diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/like-any.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/like-any.sql.out index e075fe6c0a69..7b905c2ebf30 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/like-any.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/like-any.sql.out @@ -199,7 +199,7 @@ SELECT company FROM like_any_table WHERE company LIKE ANY () -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0064", + "condition" : "_LEGACY_ERROR_TEMP_0064", "messageParameters" : { "msg" : "Expected something between '(' and ')'." }, diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/limit.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/limit.sql.out index e92dcfbc069a..999e1cf5bbcf 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/limit.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/limit.sql.out @@ -54,7 +54,7 @@ SELECT * FROM testdata LIMIT -1 -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INVALID_LIMIT_LIKE_EXPRESSION.IS_NEGATIVE", + "condition" : "INVALID_LIMIT_LIKE_EXPRESSION.IS_NEGATIVE", "sqlState" : "42K0E", "messageParameters" : { "expr" : "\"-1\"", @@ -76,7 +76,7 @@ SELECT * FROM testData TABLESAMPLE (-1 ROWS) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INVALID_LIMIT_LIKE_EXPRESSION.IS_NEGATIVE", + "condition" : "INVALID_LIMIT_LIKE_EXPRESSION.IS_NEGATIVE", "sqlState" : "42K0E", "messageParameters" : { "expr" : "\"-1\"", @@ -108,7 +108,7 @@ SELECT * FROM testdata LIMIT CAST(NULL AS INT) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INVALID_LIMIT_LIKE_EXPRESSION.IS_NULL", + "condition" : "INVALID_LIMIT_LIKE_EXPRESSION.IS_NULL", "sqlState" : "42K0E", "messageParameters" : { "expr" : "\"CAST(NULL AS INT)\"", @@ -129,7 +129,7 @@ SELECT * FROM testdata LIMIT key > 3 -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INVALID_LIMIT_LIKE_EXPRESSION.IS_UNFOLDABLE", + "condition" : "INVALID_LIMIT_LIKE_EXPRESSION.IS_UNFOLDABLE", "sqlState" : "42K0E", "messageParameters" : { "expr" : "\"(key > 3)\"", @@ -150,7 +150,7 @@ SELECT * FROM testdata LIMIT true -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INVALID_LIMIT_LIKE_EXPRESSION.DATA_TYPE", + "condition" : "INVALID_LIMIT_LIKE_EXPRESSION.DATA_TYPE", "sqlState" : "42K0E", "messageParameters" : { "dataType" : "\"BOOLEAN\"", @@ -165,7 +165,7 @@ SELECT * FROM testdata LIMIT 'a' -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INVALID_LIMIT_LIKE_EXPRESSION.DATA_TYPE", + "condition" : "INVALID_LIMIT_LIKE_EXPRESSION.DATA_TYPE", "sqlState" : "42K0E", "messageParameters" : { "dataType" : "\"STRING\"", diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/listagg-collations.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/listagg-collations.sql.out index ca471858a541..5bced5e897e2 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/listagg-collations.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/listagg-collations.sql.out @@ -76,7 +76,7 @@ SELECT listagg(DISTINCT c1 COLLATE utf8_lcase) WITHIN GROUP (ORDER BY c1 COLLATE -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INVALID_WITHIN_GROUP_EXPRESSION.MISMATCH_WITH_DISTINCT_INPUT", + "condition" : "INVALID_WITHIN_GROUP_EXPRESSION.MISMATCH_WITH_DISTINCT_INPUT", "sqlState" : "42K0K", "messageParameters" : { "funcArg" : "\"collate(c1, utf8_lcase)\"", diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/listagg.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/listagg.sql.out index 71eb3f8ca76b..9ad94bce3a2b 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/listagg.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/listagg.sql.out @@ -265,7 +265,7 @@ SELECT listagg(c1) FROM (VALUES (ARRAY('a', 'b'))) AS t(c1) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"c1\"", @@ -289,7 +289,7 @@ SELECT listagg(c1, ', ') FROM (VALUES (X'DEAD'), (X'BEEF')) AS t(c1) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "(\"BINARY\" or \"STRING\")", @@ -311,7 +311,7 @@ SELECT listagg(b, a) FROM df GROUP BY a -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.NON_FOLDABLE_INPUT", + "condition" : "DATATYPE_MISMATCH.NON_FOLDABLE_INPUT", "sqlState" : "42K09", "messageParameters" : { "inputExpr" : "\"a\"", @@ -350,7 +350,7 @@ SELECT listagg(a) WITHIN GROUP (ORDER BY a) OVER (ORDER BY a) FROM df -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INVALID_WINDOW_SPEC_FOR_AGGREGATION_FUNC", + "condition" : "INVALID_WINDOW_SPEC_FOR_AGGREGATION_FUNC", "sqlState" : "42601", "messageParameters" : { "aggFunc" : "\"listagg(a, NULL, a ASC NULLS FIRST)\"" @@ -370,7 +370,7 @@ SELECT string_agg(a) WITHIN GROUP (ORDER BY a) OVER (ORDER BY a) FROM df -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INVALID_WINDOW_SPEC_FOR_AGGREGATION_FUNC", + "condition" : "INVALID_WINDOW_SPEC_FOR_AGGREGATION_FUNC", "sqlState" : "42601", "messageParameters" : { "aggFunc" : "\"listagg(a, NULL, a ASC NULLS FIRST)\"" @@ -390,7 +390,7 @@ SELECT listagg(DISTINCT a) OVER (ORDER BY a) FROM df -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DISTINCT_WINDOW_FUNCTION_UNSUPPORTED", + "condition" : "DISTINCT_WINDOW_FUNCTION_UNSUPPORTED", "sqlState" : "0A000", "messageParameters" : { "windowExpr" : "\"listagg(DISTINCT a, NULL) OVER (ORDER BY a ASC NULLS FIRST RANGE BETWEEN UNBOUNDED PRECEDING AND CURRENT ROW)\"" @@ -410,7 +410,7 @@ SELECT listagg(DISTINCT a) WITHIN GROUP (ORDER BY b) FROM df -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INVALID_WITHIN_GROUP_EXPRESSION.MISMATCH_WITH_DISTINCT_INPUT", + "condition" : "INVALID_WITHIN_GROUP_EXPRESSION.MISMATCH_WITH_DISTINCT_INPUT", "sqlState" : "42K0K", "messageParameters" : { "funcArg" : "\"a\"", @@ -425,7 +425,7 @@ SELECT listagg(DISTINCT a) WITHIN GROUP (ORDER BY a, b) FROM df -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INVALID_WITHIN_GROUP_EXPRESSION.MISMATCH_WITH_DISTINCT_INPUT", + "condition" : "INVALID_WITHIN_GROUP_EXPRESSION.MISMATCH_WITH_DISTINCT_INPUT", "sqlState" : "42K0K", "messageParameters" : { "funcArg" : "\"a\"", diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/literals.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/literals.sql.out index 570cfb73444e..adaefb1c41ad 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/literals.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/literals.sql.out @@ -32,7 +32,7 @@ select 128Y -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_NUMERIC_LITERAL_RANGE", + "condition" : "INVALID_NUMERIC_LITERAL_RANGE", "sqlState" : "22003", "messageParameters" : { "maxValue" : "127", @@ -69,7 +69,7 @@ select 32768S -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_NUMERIC_LITERAL_RANGE", + "condition" : "INVALID_NUMERIC_LITERAL_RANGE", "sqlState" : "22003", "messageParameters" : { "maxValue" : "32767", @@ -106,7 +106,7 @@ select 9223372036854775808L -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_NUMERIC_LITERAL_RANGE", + "condition" : "INVALID_NUMERIC_LITERAL_RANGE", "sqlState" : "22003", "messageParameters" : { "maxValue" : "9223372036854775807", @@ -157,7 +157,7 @@ select 1234567890123456789012345678901234567890 -- !query analysis org.apache.spark.SparkArithmeticException { - "errorClass" : "DECIMAL_PRECISION_EXCEEDS_MAX_PRECISION", + "condition" : "DECIMAL_PRECISION_EXCEEDS_MAX_PRECISION", "sqlState" : "22003", "messageParameters" : { "maxPrecision" : "38", @@ -171,7 +171,7 @@ select 1234567890123456789012345678901234567890.0 -- !query analysis org.apache.spark.SparkArithmeticException { - "errorClass" : "DECIMAL_PRECISION_EXCEEDS_MAX_PRECISION", + "condition" : "DECIMAL_PRECISION_EXCEEDS_MAX_PRECISION", "sqlState" : "22003", "messageParameters" : { "maxPrecision" : "38", @@ -199,7 +199,7 @@ select -3.4028235E39f -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_NUMERIC_LITERAL_RANGE", + "condition" : "INVALID_NUMERIC_LITERAL_RANGE", "sqlState" : "22003", "messageParameters" : { "maxValue" : "3.4028234663852886E+38", @@ -236,7 +236,7 @@ select .e3 -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'.'", @@ -250,7 +250,7 @@ select 1E309, -1E309 -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_NUMERIC_LITERAL_RANGE", + "condition" : "INVALID_NUMERIC_LITERAL_RANGE", "sqlState" : "22003", "messageParameters" : { "maxValue" : "1.7976931348623157E+308", @@ -351,7 +351,7 @@ select date 'mar 11 2016' -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_TYPED_LITERAL", + "condition" : "INVALID_TYPED_LITERAL", "sqlState" : "42604", "messageParameters" : { "value" : "'mar 11 2016'", @@ -378,7 +378,7 @@ select timestamp '2016-33-11 20:54:00.000' -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_TYPED_LITERAL", + "condition" : "INVALID_TYPED_LITERAL", "sqlState" : "42604", "messageParameters" : { "value" : "'2016-33-11 20:54:00.000'", @@ -399,7 +399,7 @@ select GEO '(10,-6)' -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "UNSUPPORTED_TYPED_LITERAL", + "condition" : "UNSUPPORTED_TYPED_LITERAL", "sqlState" : "0A000", "messageParameters" : { "supportedTypes" : "\"DATE\", \"TIMESTAMP_NTZ\", \"TIMESTAMP_LTZ\", \"TIMESTAMP\", \"INTERVAL\", \"X\"", @@ -427,7 +427,7 @@ select 1.20E-38BD -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "DECIMAL_PRECISION_EXCEEDS_MAX_PRECISION", + "condition" : "DECIMAL_PRECISION_EXCEEDS_MAX_PRECISION", "sqlState" : "22003", "messageParameters" : { "maxPrecision" : "38", @@ -455,7 +455,7 @@ select X'XuZ' -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_TYPED_LITERAL", + "condition" : "INVALID_TYPED_LITERAL", "sqlState" : "42604", "messageParameters" : { "value" : "'XuZ'", @@ -483,7 +483,7 @@ select +date '1999-01-01' -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"DATE '1999-01-01'\"", @@ -507,7 +507,7 @@ select +timestamp '1999-01-01' -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"TIMESTAMP '1999-01-01 00:00:00'\"", @@ -538,7 +538,7 @@ select +map(1, 2) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"map(1, 2)\"", @@ -562,7 +562,7 @@ select +array(1,2) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"array(1, 2)\"", @@ -586,7 +586,7 @@ select +named_struct('a', 1, 'b', 'spark') -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"named_struct(a, 1, b, spark)\"", @@ -610,7 +610,7 @@ select +X'1' -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"X'01'\"", @@ -634,7 +634,7 @@ select -date '1999-01-01' -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"DATE '1999-01-01'\"", @@ -658,7 +658,7 @@ select -timestamp '1999-01-01' -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"TIMESTAMP '1999-01-01 00:00:00'\"", @@ -682,7 +682,7 @@ select -x'2379ACFe' -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"X'2379ACFE'\"", diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/map.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/map.sql.out index 177f73608fba..ae3d78dace4b 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/map.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/map.sql.out @@ -60,7 +60,7 @@ select map_contains_key(map('1', 'a', '2', 'b'), 1) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.MAP_FUNCTION_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.MAP_FUNCTION_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "\"MAP\"", @@ -84,7 +84,7 @@ select map_contains_key(map(1, 'a', 2, 'b'), '1') -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.MAP_FUNCTION_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.MAP_FUNCTION_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "\"MAP\"", diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/mask-functions.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/mask-functions.sql.out index dba6a8293c9f..3d562fadf247 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/mask-functions.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/mask-functions.sql.out @@ -39,7 +39,7 @@ SELECT mask('AbCD123-@$#', 'Qa', 'qa', 'da', 'oa') -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.INPUT_SIZE_NOT_ONE", + "condition" : "DATATYPE_MISMATCH.INPUT_SIZE_NOT_ONE", "sqlState" : "42K09", "messageParameters" : { "exprName" : "upperChar", @@ -405,7 +405,7 @@ SELECT mask(c1, replaceArg) from values('abcd-EFGH-8765-4321', 'a') as t(c1, rep -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.NON_FOLDABLE_INPUT", + "condition" : "DATATYPE_MISMATCH.NON_FOLDABLE_INPUT", "sqlState" : "42K09", "messageParameters" : { "inputExpr" : "\"replaceArg\"", @@ -428,7 +428,7 @@ SELECT mask(c1, replaceArg) from values('abcd-EFGH-8765-4321', 'ABC') as t(c1, r -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.NON_FOLDABLE_INPUT", + "condition" : "DATATYPE_MISMATCH.NON_FOLDABLE_INPUT", "sqlState" : "42K09", "messageParameters" : { "inputExpr" : "\"replaceArg\"", @@ -451,7 +451,7 @@ SELECT mask(c1, replaceArg) from values('abcd-EFGH-8765-4321', 123) as t(c1, rep -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"replaceArg\"", @@ -475,7 +475,7 @@ SELECT mask('abcd-EFGH-8765-4321', 'A', 'w', '') -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.INPUT_SIZE_NOT_ONE", + "condition" : "DATATYPE_MISMATCH.INPUT_SIZE_NOT_ONE", "sqlState" : "42K09", "messageParameters" : { "exprName" : "digitChar", @@ -496,7 +496,7 @@ SELECT mask('abcd-EFGH-8765-4321', 'A', 'abc') -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.INPUT_SIZE_NOT_ONE", + "condition" : "DATATYPE_MISMATCH.INPUT_SIZE_NOT_ONE", "sqlState" : "42K09", "messageParameters" : { "exprName" : "lowerChar", diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/misc-functions.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/misc-functions.sql.out index e30bdf12f4a3..ca80c417fe9e 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/misc-functions.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/misc-functions.sql.out @@ -124,7 +124,7 @@ SELECT raise_error('error message', Map()) -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", + "condition" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", "sqlState" : "42605", "messageParameters" : { "actualNum" : "2", @@ -147,7 +147,7 @@ SELECT raise_error('error message', 'some args') -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", + "condition" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", "sqlState" : "42605", "messageParameters" : { "actualNum" : "2", @@ -170,7 +170,7 @@ SELECT raise_error() -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", + "condition" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", "sqlState" : "42605", "messageParameters" : { "actualNum" : "0", diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/mode.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/mode.sql.out index 8028c344140f..95c2db670a87 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/mode.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/mode.sql.out @@ -74,7 +74,7 @@ SELECT department, mode(DISTINCT salary) FROM basic_pays GROUP BY department ORD -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_WITHIN_GROUP_EXPRESSION.DISTINCT_UNSUPPORTED", + "condition" : "INVALID_WITHIN_GROUP_EXPRESSION.DISTINCT_UNSUPPORTED", "sqlState" : "42K0K", "messageParameters" : { "funcName" : "`mode`" @@ -118,7 +118,7 @@ SELECT mode(col, 'true') FROM VALUES (-10), (0), (10) AS tab(col) -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "UNEXPECTED_INPUT_TYPE", + "condition" : "UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "functionName" : "`mode`", @@ -142,7 +142,7 @@ SELECT mode(col, null) FROM VALUES (-10), (0), (10) AS tab(col) -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_NULL", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_NULL", "sqlState" : "42K09", "messageParameters" : { "exprName" : "`deterministic`", @@ -163,7 +163,7 @@ SELECT mode(col, b) FROM VALUES (-10, false), (0, false), (10, false) AS tab(col -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "NON_FOLDABLE_ARGUMENT", + "condition" : "NON_FOLDABLE_ARGUMENT", "sqlState" : "42K08", "messageParameters" : { "funcName" : "`mode`", @@ -379,7 +379,7 @@ FROM basic_pays -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_WITHIN_GROUP_EXPRESSION.DISTINCT_UNSUPPORTED", + "condition" : "INVALID_WITHIN_GROUP_EXPRESSION.DISTINCT_UNSUPPORTED", "sqlState" : "42K0K", "messageParameters" : { "funcName" : "`mode`" @@ -401,7 +401,7 @@ FROM basic_pays -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_WITHIN_GROUP_EXPRESSION.WITHIN_GROUP_MISSING", + "condition" : "INVALID_WITHIN_GROUP_EXPRESSION.WITHIN_GROUP_MISSING", "sqlState" : "42K0K", "messageParameters" : { "funcName" : "`mode`" @@ -423,7 +423,7 @@ FROM basic_pays -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_WITHIN_GROUP_EXPRESSION.WRONG_NUM_ORDERINGS", + "condition" : "INVALID_WITHIN_GROUP_EXPRESSION.WRONG_NUM_ORDERINGS", "sqlState" : "42K0K", "messageParameters" : { "actualNum" : "1", diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/named-function-arguments.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/named-function-arguments.sql.out index 2315a5f0678a..bafc3a87df82 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/named-function-arguments.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/named-function-arguments.sql.out @@ -252,7 +252,7 @@ SELECT * FROM explode(collection => explode(array(1))) -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "UNSUPPORTED_GENERATOR.NESTED_IN_EXPRESSIONS", + "condition" : "UNSUPPORTED_GENERATOR.NESTED_IN_EXPRESSIONS", "sqlState" : "42K0E", "messageParameters" : { "expression" : "\"explode(explode(array(1)))\"" @@ -272,7 +272,7 @@ SELECT * FROM explode(collection => explode(collection => array(1))) -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "UNSUPPORTED_GENERATOR.NESTED_IN_EXPRESSIONS", + "condition" : "UNSUPPORTED_GENERATOR.NESTED_IN_EXPRESSIONS", "sqlState" : "42K0E", "messageParameters" : { "expression" : "\"explode(explode(array(1)))\"" @@ -300,7 +300,7 @@ SELECT * FROM explode(collection => TABLE(v)) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"functiontablesubqueryargumentexpression()\"", @@ -324,7 +324,7 @@ SELECT mask(lowerChar => 'q', 'AbCD123-@$#', upperChar => 'Q', otherChar => 'o', -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "UNEXPECTED_POSITIONAL_ARGUMENT", + "condition" : "UNEXPECTED_POSITIONAL_ARGUMENT", "sqlState" : "4274K", "messageParameters" : { "parameterName" : "`lowerChar`", @@ -345,7 +345,7 @@ SELECT mask('AbCD123-@$#', lowerChar => 'q', upperChar => 'Q', otherChar => 'o', -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "DUPLICATE_ROUTINE_PARAMETER_ASSIGNMENT.DOUBLE_NAMED_ARGUMENT_REFERENCE", + "condition" : "DUPLICATE_ROUTINE_PARAMETER_ASSIGNMENT.DOUBLE_NAMED_ARGUMENT_REFERENCE", "sqlState" : "4274K", "messageParameters" : { "parameterName" : "`digitChar`", @@ -366,7 +366,7 @@ SELECT mask('AbCD123-@$#', lowerChar => 'q', upperChar => 'Q', otherChar => 'o', -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "DUPLICATE_ROUTINE_PARAMETER_ASSIGNMENT.BOTH_POSITIONAL_AND_NAMED", + "condition" : "DUPLICATE_ROUTINE_PARAMETER_ASSIGNMENT.BOTH_POSITIONAL_AND_NAMED", "sqlState" : "4274K", "messageParameters" : { "parameterName" : "`str`", @@ -387,7 +387,7 @@ SELECT mask(lowerChar => 'q', upperChar => 'Q', otherChar => 'o', digitChar => ' -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "REQUIRED_PARAMETER_NOT_FOUND", + "condition" : "REQUIRED_PARAMETER_NOT_FOUND", "sqlState" : "4274K", "messageParameters" : { "index" : "0", @@ -409,7 +409,7 @@ SELECT mask('AbCD123-@$#', lowerChar => 'q', upperChar => 'Q', otherChar => 'o', -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "UNRECOGNIZED_PARAMETER_NAME", + "condition" : "UNRECOGNIZED_PARAMETER_NAME", "sqlState" : "4274K", "messageParameters" : { "argumentName" : "`cellular`", @@ -431,7 +431,7 @@ SELECT encode(str => 'a', charset => 'utf-8') -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "NAMED_PARAMETERS_NOT_SUPPORTED", + "condition" : "NAMED_PARAMETERS_NOT_SUPPORTED", "sqlState" : "4274K", "messageParameters" : { "functionName" : "`encode`" @@ -451,7 +451,7 @@ SELECT mask('AbCD123-@$#', 'Q', 'q', 'd', 'o', 'k') -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", + "condition" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", "sqlState" : "42605", "messageParameters" : { "actualNum" : "6", diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/natural-join.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/natural-join.sql.out index 857c574af3d2..34bdfdb7f8e5 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/natural-join.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/natural-join.sql.out @@ -490,7 +490,7 @@ SELECT nt2.k FROM (SELECT * FROM nt1 natural join nt2) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`nt2`.`k`", diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/nonansi/array.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/nonansi/array.sql.out index 4db56d6c7056..f2fe445dfaa8 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/nonansi/array.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/nonansi/array.sql.out @@ -172,7 +172,7 @@ select sort_array(array('b', 'd'), '1') -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"1\"", @@ -338,7 +338,7 @@ select array_size(map('a', 1, 'b', 2)) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"map(a, 1, b, 2)\"", @@ -460,7 +460,7 @@ select array_insert(array(1, 2, 3), 3, "4") -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.ARRAY_FUNCTION_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.ARRAY_FUNCTION_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "\"ARRAY\"", @@ -552,7 +552,7 @@ select array_compact(id) from values (1) as t(id) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"id\"", diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/nonansi/cast.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/nonansi/cast.sql.out index e0687b564d3d..368d54a7f4e2 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/nonansi/cast.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/nonansi/cast.sql.out @@ -296,7 +296,7 @@ SELECT CAST(interval 3 month 1 hour AS string) -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0029", + "condition" : "_LEGACY_ERROR_TEMP_0029", "messageParameters" : { "literal" : "interval 3 month 1 hour" }, @@ -913,7 +913,7 @@ SELECT '2147483648' :: BINT -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "UNSUPPORTED_DATATYPE", + "condition" : "UNSUPPORTED_DATATYPE", "sqlState" : "0A000", "messageParameters" : { "typeName" : "\"BINT\"" @@ -933,7 +933,7 @@ SELECT '2147483648' :: SELECT -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "UNSUPPORTED_DATATYPE", + "condition" : "UNSUPPORTED_DATATYPE", "sqlState" : "0A000", "messageParameters" : { "typeName" : "\"SELECT\"" @@ -953,7 +953,7 @@ SELECT FALSE IS NOT NULL :: string -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'::'", diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/nonansi/date.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/nonansi/date.sql.out index 88c7d7b4e7d7..acdf55fdc387 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/nonansi/date.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/nonansi/date.sql.out @@ -18,7 +18,7 @@ select date '2020-01-01中文' -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_TYPED_LITERAL", + "condition" : "INVALID_TYPED_LITERAL", "sqlState" : "42604", "messageParameters" : { "value" : "'2020-01-01中文'", @@ -60,7 +60,7 @@ select date'015' -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_TYPED_LITERAL", + "condition" : "INVALID_TYPED_LITERAL", "sqlState" : "42604", "messageParameters" : { "value" : "'015'", @@ -81,7 +81,7 @@ select date'2021-4294967297-11' -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_TYPED_LITERAL", + "condition" : "INVALID_TYPED_LITERAL", "sqlState" : "42604", "messageParameters" : { "value" : "'2021-4294967297-11'", @@ -114,7 +114,7 @@ select curdate(1) -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", + "condition" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", "sqlState" : "42605", "messageParameters" : { "actualNum" : "1", @@ -285,7 +285,7 @@ select date_add('2011-11-11', 1L) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"1\"", @@ -309,7 +309,7 @@ select date_add('2011-11-11', 1.0) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"1.0\"", @@ -333,7 +333,7 @@ select date_add('2011-11-11', 1E1) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"10.0\"", @@ -364,7 +364,7 @@ select date_add('2011-11-11', '1.2') -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "SECOND_FUNCTION_ARGUMENT_NOT_INTEGER", + "condition" : "SECOND_FUNCTION_ARGUMENT_NOT_INTEGER", "sqlState" : "22023", "messageParameters" : { "functionName" : "date_add" @@ -429,7 +429,7 @@ select date_sub('2011-11-11', 1L) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"1\"", @@ -453,7 +453,7 @@ select date_sub('2011-11-11', 1.0) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"1.0\"", @@ -477,7 +477,7 @@ select date_sub('2011-11-11', 1E1) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"10.0\"", @@ -507,7 +507,7 @@ select date_sub(date'2011-11-11', '1.2') -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "SECOND_FUNCTION_ARGUMENT_NOT_INTEGER", + "condition" : "SECOND_FUNCTION_ARGUMENT_NOT_INTEGER", "sqlState" : "22023", "messageParameters" : { "functionName" : "date_sub" @@ -545,7 +545,7 @@ select date_add('2011-11-11', int_str) from date_view -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"int_str\"", @@ -569,7 +569,7 @@ select date_sub('2011-11-11', int_str) from date_view -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"int_str\"", @@ -615,7 +615,7 @@ select date '2011-11-11' + 1E1 -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"10.0\"", @@ -663,7 +663,7 @@ select date '2001-10-01' - '2001-09-28' -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"2001-09-28\"", @@ -711,7 +711,7 @@ select date '2001-09-28' - date_str from date_view -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"date_str\"", @@ -735,7 +735,7 @@ select date'2011-11-11' + '1' -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"1\"", @@ -759,7 +759,7 @@ select '1' + date'2011-11-11' -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"1\"", @@ -887,7 +887,7 @@ select dateadd('MICROSECOND', 1001, timestamp'2022-02-25 01:02:03.123') -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", + "condition" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", "sqlState" : "22023", "messageParameters" : { "functionName" : "`dateadd`", @@ -909,7 +909,7 @@ select date_add('QUARTER', 5, date'2022-02-25') -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", + "condition" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", "sqlState" : "22023", "messageParameters" : { "functionName" : "`date_add`", @@ -991,7 +991,7 @@ select date_diff('MILLISECOND', timestamp'2022-02-25 01:02:03.456', timestamp'20 -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", + "condition" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", "sqlState" : "22023", "messageParameters" : { "functionName" : "`date_diff`", @@ -1013,7 +1013,7 @@ select datediff('YEAR', date'2022-02-25', date'2023-02-25') -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", + "condition" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", "sqlState" : "22023", "messageParameters" : { "functionName" : "`datediff`", diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/nonansi/double-quoted-identifiers.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/nonansi/double-quoted-identifiers.sql.out index a02bf525f947..f2ee4fb48e63 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/nonansi/double-quoted-identifiers.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/nonansi/double-quoted-identifiers.sql.out @@ -4,7 +4,7 @@ SELECT 1 FROM "not_exist" -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'\"not_exist\"'", @@ -18,7 +18,7 @@ USE SCHEMA "not_exist" -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'\"not_exist\"'", @@ -32,7 +32,7 @@ ALTER TABLE "not_exist" ADD COLUMN not_exist int -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'\"not_exist\"'", @@ -46,7 +46,7 @@ ALTER TABLE not_exist ADD COLUMN "not_exist" int -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'\"not_exist\"'", @@ -60,7 +60,7 @@ SELECT 1 AS "not_exist" FROM not_exist -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'\"not_exist\"'", @@ -74,7 +74,7 @@ SELECT 1 FROM not_exist AS X("hello") -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'\"hello\"'", @@ -88,7 +88,7 @@ SELECT "not_exist"() -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'\"not_exist\"'", @@ -102,7 +102,7 @@ SELECT "not_exist".not_exist() -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'\"not_exist\"'", @@ -116,7 +116,7 @@ SELECT 1 FROM `hello` -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "TABLE_OR_VIEW_NOT_FOUND", + "condition" : "TABLE_OR_VIEW_NOT_FOUND", "sqlState" : "42P01", "messageParameters" : { "relationName" : "`hello`" @@ -136,7 +136,7 @@ USE SCHEMA `not_exist` -- !query analysis org.apache.spark.sql.catalyst.analysis.NoSuchNamespaceException { - "errorClass" : "SCHEMA_NOT_FOUND", + "condition" : "SCHEMA_NOT_FOUND", "sqlState" : "42704", "messageParameters" : { "schemaName" : "`spark_catalog`.`not_exist`" @@ -149,7 +149,7 @@ ALTER TABLE `not_exist` ADD COLUMN not_exist int -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "TABLE_OR_VIEW_NOT_FOUND", + "condition" : "TABLE_OR_VIEW_NOT_FOUND", "sqlState" : "42P01", "messageParameters" : { "relationName" : "`not_exist`" @@ -169,7 +169,7 @@ ALTER TABLE not_exist ADD COLUMN `not_exist` int -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "TABLE_OR_VIEW_NOT_FOUND", + "condition" : "TABLE_OR_VIEW_NOT_FOUND", "sqlState" : "42P01", "messageParameters" : { "relationName" : "`not_exist`" @@ -189,7 +189,7 @@ SELECT 1 AS `not_exist` FROM `not_exist` -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "TABLE_OR_VIEW_NOT_FOUND", + "condition" : "TABLE_OR_VIEW_NOT_FOUND", "sqlState" : "42P01", "messageParameters" : { "relationName" : "`not_exist`" @@ -209,7 +209,7 @@ SELECT 1 FROM not_exist AS X(`hello`) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "TABLE_OR_VIEW_NOT_FOUND", + "condition" : "TABLE_OR_VIEW_NOT_FOUND", "sqlState" : "42P01", "messageParameters" : { "relationName" : "`not_exist`" @@ -229,7 +229,7 @@ SELECT `not_exist`() -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "UNRESOLVED_ROUTINE", + "condition" : "UNRESOLVED_ROUTINE", "sqlState" : "42883", "messageParameters" : { "routineName" : "`not_exist`", @@ -250,7 +250,7 @@ SELECT `not_exist`.not_exist() -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "UNRESOLVED_ROUTINE", + "condition" : "UNRESOLVED_ROUTINE", "sqlState" : "42883", "messageParameters" : { "routineName" : "`not_exist`.`not_exist`", @@ -327,7 +327,7 @@ CREATE SCHEMA "myschema" -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'\"myschema\"'", @@ -342,7 +342,7 @@ CREATE TEMPORARY VIEW "myview"("c1") AS -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'\"myview\"'", @@ -356,7 +356,7 @@ SELECT "a1" AS "a2" FROM "myview" AS "atab"("a1") -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'\"a2\"'", @@ -370,7 +370,7 @@ DROP TABLE "myview" -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'\"myview\"'", @@ -384,7 +384,7 @@ DROP SCHEMA "myschema" -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'\"myschema\"'", diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/nonansi/higher-order-functions.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/nonansi/higher-order-functions.sql.out index 1281b19eb2f8..81608692a048 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/nonansi/higher-order-functions.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/nonansi/higher-order-functions.sql.out @@ -20,7 +20,7 @@ select upper(x -> x) as v -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_LAMBDA_FUNCTION_CALL.NON_HIGHER_ORDER_FUNCTION", + "condition" : "INVALID_LAMBDA_FUNCTION_CALL.NON_HIGHER_ORDER_FUNCTION", "sqlState" : "42K0D", "messageParameters" : { "class" : "org.apache.spark.sql.catalyst.expressions.Upper" @@ -40,7 +40,7 @@ select ceil(x -> x) as v -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_LAMBDA_FUNCTION_CALL.NON_HIGHER_ORDER_FUNCTION", + "condition" : "INVALID_LAMBDA_FUNCTION_CALL.NON_HIGHER_ORDER_FUNCTION", "sqlState" : "42K0D", "messageParameters" : { "class" : "org.apache.spark.sql.catalyst.expressions.CeilExpressionBuilder$" diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/nonansi/interval.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/nonansi/interval.sql.out index c0196bbe118e..36b0b8ec51b9 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/nonansi/interval.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/nonansi/interval.sql.out @@ -10,7 +10,7 @@ select interval 4 month 2 weeks 3 microseconds * 1.5 -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0029", + "condition" : "_LEGACY_ERROR_TEMP_0029", "messageParameters" : { "literal" : "interval 4 month 2 weeks 3 microseconds" }, @@ -161,7 +161,7 @@ select '2' / interval 2 second -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"STRING\"", @@ -183,7 +183,7 @@ select '2' / interval 2 year -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"STRING\"", @@ -261,7 +261,7 @@ select 2 / interval '2' year -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"INT\"", @@ -283,7 +283,7 @@ select 2 / interval '2' hour -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"INT\"", @@ -305,7 +305,7 @@ select null / interval '2' year -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"VOID\"", @@ -327,7 +327,7 @@ select null / interval '2' hour -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"VOID\"", @@ -349,7 +349,7 @@ select -interval '-1 month 1 day -1 second' -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0029", + "condition" : "_LEGACY_ERROR_TEMP_0029", "messageParameters" : { "literal" : "interval '-1 month 1 day -1 second'" }, @@ -382,7 +382,7 @@ select -interval -1 month 1 day -1 second -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0029", + "condition" : "_LEGACY_ERROR_TEMP_0029", "messageParameters" : { "literal" : "interval -1 month 1 day -1 second" }, @@ -415,7 +415,7 @@ select +interval '-1 month 1 day -1 second' -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0029", + "condition" : "_LEGACY_ERROR_TEMP_0029", "messageParameters" : { "literal" : "interval '-1 month 1 day -1 second'" }, @@ -448,7 +448,7 @@ select +interval -1 month 1 day -1 second -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0029", + "condition" : "_LEGACY_ERROR_TEMP_0029", "messageParameters" : { "literal" : "interval -1 month 1 day -1 second" }, @@ -817,7 +817,7 @@ select interval 1 year 2 month 3 week 4 day 5 hour 6 minute 7 seconds 8 millisec -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0029", + "condition" : "_LEGACY_ERROR_TEMP_0029", "messageParameters" : { "literal" : "interval 1 year 2 month 3 week 4 day 5 hour 6 minute 7 seconds 8 millisecond 9 microsecond" }, @@ -857,7 +857,7 @@ select interval '30' year '25' month '-100' day '40' hour '80' minute '299.88998 -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0029", + "condition" : "_LEGACY_ERROR_TEMP_0029", "messageParameters" : { "literal" : "interval '30' year '25' month '-100' day '40' hour '80' minute '299.889987299' second" }, @@ -981,7 +981,7 @@ select interval '20 15:40:32.99899999' day to hour -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_INTERVAL_FORMAT.UNMATCHED_FORMAT_STRING_WITH_NOTICE", + "condition" : "INVALID_INTERVAL_FORMAT.UNMATCHED_FORMAT_STRING_WITH_NOTICE", "sqlState" : "22006", "messageParameters" : { "input" : "20 15:40:32.99899999", @@ -1004,7 +1004,7 @@ select interval '20 15:40:32.99899999' day to minute -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_INTERVAL_FORMAT.UNMATCHED_FORMAT_STRING_WITH_NOTICE", + "condition" : "INVALID_INTERVAL_FORMAT.UNMATCHED_FORMAT_STRING_WITH_NOTICE", "sqlState" : "22006", "messageParameters" : { "input" : "20 15:40:32.99899999", @@ -1027,7 +1027,7 @@ select interval '15:40:32.99899999' hour to minute -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_INTERVAL_FORMAT.UNMATCHED_FORMAT_STRING_WITH_NOTICE", + "condition" : "INVALID_INTERVAL_FORMAT.UNMATCHED_FORMAT_STRING_WITH_NOTICE", "sqlState" : "22006", "messageParameters" : { "input" : "15:40:32.99899999", @@ -1050,7 +1050,7 @@ select interval '15:40.99899999' hour to second -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_INTERVAL_FORMAT.UNMATCHED_FORMAT_STRING_WITH_NOTICE", + "condition" : "INVALID_INTERVAL_FORMAT.UNMATCHED_FORMAT_STRING_WITH_NOTICE", "sqlState" : "22006", "messageParameters" : { "input" : "15:40.99899999", @@ -1073,7 +1073,7 @@ select interval '15:40' hour to second -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_INTERVAL_FORMAT.UNMATCHED_FORMAT_STRING_WITH_NOTICE", + "condition" : "INVALID_INTERVAL_FORMAT.UNMATCHED_FORMAT_STRING_WITH_NOTICE", "sqlState" : "22006", "messageParameters" : { "input" : "15:40", @@ -1096,7 +1096,7 @@ select interval '20 40:32.99899999' minute to second -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_INTERVAL_FORMAT.UNMATCHED_FORMAT_STRING_WITH_NOTICE", + "condition" : "INVALID_INTERVAL_FORMAT.UNMATCHED_FORMAT_STRING_WITH_NOTICE", "sqlState" : "22006", "messageParameters" : { "input" : "20 40:32.99899999", @@ -1119,7 +1119,7 @@ select interval 10 nanoseconds -- !query analysis org.apache.spark.SparkIllegalArgumentException { - "errorClass" : "INVALID_INTERVAL_FORMAT.INVALID_UNIT", + "condition" : "INVALID_INTERVAL_FORMAT.INVALID_UNIT", "sqlState" : "22006", "messageParameters" : { "input" : " 10 nanoseconds", @@ -1203,7 +1203,7 @@ select interval 1 fake_unit -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'fake_unit'", @@ -1217,7 +1217,7 @@ select interval 1 year to month -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0027", + "condition" : "_LEGACY_ERROR_TEMP_0027", "queryContext" : [ { "objectType" : "", "objectName" : "", @@ -1233,7 +1233,7 @@ select interval '1' year to second -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_INTERVAL_FORMAT.UNSUPPORTED_FROM_TO_EXPRESSION", + "condition" : "INVALID_INTERVAL_FORMAT.UNSUPPORTED_FROM_TO_EXPRESSION", "sqlState" : "22006", "messageParameters" : { "from" : "year", @@ -1255,7 +1255,7 @@ select interval '10-9' year to month '2-1' year to month -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0024", + "condition" : "_LEGACY_ERROR_TEMP_0024", "queryContext" : [ { "objectType" : "", "objectName" : "", @@ -1271,7 +1271,7 @@ select interval '10-9' year to month '12:11:10' hour to second -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0024", + "condition" : "_LEGACY_ERROR_TEMP_0024", "queryContext" : [ { "objectType" : "", "objectName" : "", @@ -1287,7 +1287,7 @@ select interval '1 15:11' day to minute '12:11:10' hour to second -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0024", + "condition" : "_LEGACY_ERROR_TEMP_0024", "queryContext" : [ { "objectType" : "", "objectName" : "", @@ -1303,7 +1303,7 @@ select interval 1 year '2-1' year to month -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0024", + "condition" : "_LEGACY_ERROR_TEMP_0024", "queryContext" : [ { "objectType" : "", "objectName" : "", @@ -1319,7 +1319,7 @@ select interval 1 year '12:11:10' hour to second -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0024", + "condition" : "_LEGACY_ERROR_TEMP_0024", "queryContext" : [ { "objectType" : "", "objectName" : "", @@ -1335,7 +1335,7 @@ select interval '10-9' year to month '1' year -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0024", + "condition" : "_LEGACY_ERROR_TEMP_0024", "queryContext" : [ { "objectType" : "", "objectName" : "", @@ -1351,7 +1351,7 @@ select interval '12:11:10' hour to second '1' year -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0024", + "condition" : "_LEGACY_ERROR_TEMP_0024", "queryContext" : [ { "objectType" : "", "objectName" : "", @@ -1367,7 +1367,7 @@ select interval (-30) day -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "UNRESOLVED_ROUTINE", + "condition" : "UNRESOLVED_ROUTINE", "sqlState" : "42883", "messageParameters" : { "routineName" : "`interval`", @@ -1388,7 +1388,7 @@ select interval (a + 1) day -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "UNRESOLVED_ROUTINE", + "condition" : "UNRESOLVED_ROUTINE", "sqlState" : "42883", "messageParameters" : { "routineName" : "`interval`", @@ -1409,7 +1409,7 @@ select interval 30 day day day -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'day'", @@ -1423,7 +1423,7 @@ select interval (-30) days -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "UNRESOLVED_ROUTINE", + "condition" : "UNRESOLVED_ROUTINE", "sqlState" : "42883", "messageParameters" : { "routineName" : "`interval`", @@ -1444,7 +1444,7 @@ select interval (a + 1) days -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "UNRESOLVED_ROUTINE", + "condition" : "UNRESOLVED_ROUTINE", "sqlState" : "42883", "messageParameters" : { "routineName" : "`interval`", @@ -1465,7 +1465,7 @@ select interval 30 days days days -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'days'", @@ -1486,7 +1486,7 @@ SELECT INTERVAL '178956970-8' YEAR TO MONTH -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_INTERVAL_FORMAT.INTERVAL_PARSING", + "condition" : "INVALID_INTERVAL_FORMAT.INTERVAL_PARSING", "sqlState" : "22006", "messageParameters" : { "input" : "178956970-8", @@ -1540,7 +1540,7 @@ select -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"INTERVAL YEAR\"", @@ -1577,7 +1577,7 @@ select interval '2' year + '3-3' -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"INTERVAL YEAR\"", @@ -1599,7 +1599,7 @@ select interval '2' year - '4' -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"INTERVAL YEAR\"", @@ -1643,7 +1643,7 @@ select interval '2' year + str from interval_view -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"INTERVAL YEAR\"", @@ -1665,7 +1665,7 @@ select interval '2' year - str from interval_view -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"INTERVAL YEAR\"", @@ -1709,7 +1709,7 @@ select interval '2-2' year to month + interval '3' day -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"INTERVAL '2-2' YEAR TO MONTH\"", @@ -1733,7 +1733,7 @@ select interval '3' day + interval '2-2' year to month -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"INTERVAL '2-2' YEAR TO MONTH\"", @@ -1757,7 +1757,7 @@ select interval '2-2' year to month - interval '3' day -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"INTERVAL '2-2' YEAR TO MONTH\"", @@ -1781,7 +1781,7 @@ select interval '3' day - interval '2-2' year to month -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"INTERVAL DAY\"", @@ -1803,7 +1803,7 @@ select 1 - interval '2' second -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"1\"", @@ -1827,7 +1827,7 @@ select 1 + interval '2' month -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"INT\"", @@ -1849,7 +1849,7 @@ select interval '2' second + 1 -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"1\"", @@ -1873,7 +1873,7 @@ select interval '2' month - 1 -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"INTERVAL MONTH\"", @@ -1937,7 +1937,7 @@ select interval '-\t2-2\t' year to month -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_INTERVAL_FORMAT.UNMATCHED_FORMAT_STRING", + "condition" : "INVALID_INTERVAL_FORMAT.UNMATCHED_FORMAT_STRING", "sqlState" : "22006", "messageParameters" : { "input" : "-\t2-2\t", @@ -1967,7 +1967,7 @@ select interval '\n-\t10\t 12:34:46.789\t' day to second -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_INTERVAL_FORMAT.UNMATCHED_FORMAT_STRING_WITH_NOTICE", + "condition" : "INVALID_INTERVAL_FORMAT.UNMATCHED_FORMAT_STRING_WITH_NOTICE", "sqlState" : "22006", "messageParameters" : { "input" : "\n-\t10\t 12:34:46.789\t", @@ -1990,7 +1990,7 @@ select interval '中文 interval 1 day' -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_TYPED_LITERAL", + "condition" : "INVALID_TYPED_LITERAL", "sqlState" : "42604", "messageParameters" : { "value" : "'中文 interval 1 day'", @@ -2011,7 +2011,7 @@ select interval 'interval中文 1 day' -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_TYPED_LITERAL", + "condition" : "INVALID_TYPED_LITERAL", "sqlState" : "42604", "messageParameters" : { "value" : "'interval中文 1 day'", @@ -2032,7 +2032,7 @@ select interval 'interval 1中文day' -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_TYPED_LITERAL", + "condition" : "INVALID_TYPED_LITERAL", "sqlState" : "42604", "messageParameters" : { "value" : "'interval 1中文day'", @@ -2141,7 +2141,7 @@ select interval '+' -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_TYPED_LITERAL", + "condition" : "INVALID_TYPED_LITERAL", "sqlState" : "42604", "messageParameters" : { "value" : "'+'", @@ -2162,7 +2162,7 @@ select interval '+.' -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_TYPED_LITERAL", + "condition" : "INVALID_TYPED_LITERAL", "sqlState" : "42604", "messageParameters" : { "value" : "'+.'", @@ -2183,7 +2183,7 @@ select interval '1' -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_TYPED_LITERAL", + "condition" : "INVALID_TYPED_LITERAL", "sqlState" : "42604", "messageParameters" : { "value" : "'1'", @@ -2204,7 +2204,7 @@ select interval '1.2' -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_TYPED_LITERAL", + "condition" : "INVALID_TYPED_LITERAL", "sqlState" : "42604", "messageParameters" : { "value" : "'1.2'", @@ -2225,7 +2225,7 @@ select interval '- 2' -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_TYPED_LITERAL", + "condition" : "INVALID_TYPED_LITERAL", "sqlState" : "42604", "messageParameters" : { "value" : "'- 2'", @@ -2246,7 +2246,7 @@ select interval '1 day -' -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_TYPED_LITERAL", + "condition" : "INVALID_TYPED_LITERAL", "sqlState" : "42604", "messageParameters" : { "value" : "'1 day -'", @@ -2267,7 +2267,7 @@ select interval '1 day 1' -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_TYPED_LITERAL", + "condition" : "INVALID_TYPED_LITERAL", "sqlState" : "42604", "messageParameters" : { "value" : "'1 day 1'", @@ -2288,7 +2288,7 @@ select interval '1 day 2' day -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0026", + "condition" : "_LEGACY_ERROR_TEMP_0026", "messageParameters" : { "value" : "1 day 2" }, @@ -2307,7 +2307,7 @@ select interval 'interval 1' day -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0026", + "condition" : "_LEGACY_ERROR_TEMP_0026", "messageParameters" : { "value" : "interval 1" }, @@ -2501,7 +2501,7 @@ SELECT INTERVAL '106751992 04' DAY TO HOUR -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0063", + "condition" : "_LEGACY_ERROR_TEMP_0063", "messageParameters" : { "msg" : "requirement failed: day 106751992 outside range [0, 106751991]" }, @@ -2520,7 +2520,7 @@ SELECT INTERVAL '-106751992 04' DAY TO HOUR -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0063", + "condition" : "_LEGACY_ERROR_TEMP_0063", "messageParameters" : { "msg" : "requirement failed: day 106751992 outside range [0, 106751991]" }, @@ -2539,7 +2539,7 @@ SELECT INTERVAL '2562047789:00' HOUR TO MINUTE -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0063", + "condition" : "_LEGACY_ERROR_TEMP_0063", "messageParameters" : { "msg" : "requirement failed: hour 2562047789 outside range [0, 2562047788]" }, @@ -2558,7 +2558,7 @@ SELECT INTERVAL '-2562047789:00' HOUR TO MINUTE -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0063", + "condition" : "_LEGACY_ERROR_TEMP_0063", "messageParameters" : { "msg" : "requirement failed: hour 2562047789 outside range [0, 2562047788]" }, @@ -2577,7 +2577,7 @@ SELECT INTERVAL '153722867281:54.775808' MINUTE TO SECOND -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0063", + "condition" : "_LEGACY_ERROR_TEMP_0063", "messageParameters" : { "msg" : "requirement failed: minute 153722867281 outside range [0, 153722867280]" }, @@ -2596,7 +2596,7 @@ SELECT INTERVAL '-153722867281:54.775808' MINUTE TO SECOND -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0063", + "condition" : "_LEGACY_ERROR_TEMP_0063", "messageParameters" : { "msg" : "requirement failed: minute 153722867281 outside range [0, 153722867280]" }, @@ -2727,7 +2727,7 @@ SELECT INTERVAL 1 MONTH > INTERVAL 20 DAYS -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"INTERVAL MONTH\"", @@ -2749,7 +2749,7 @@ SELECT INTERVAL '1' DAY < '1' -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"INTERVAL DAY\"", @@ -2771,7 +2771,7 @@ SELECT INTERVAL '1' DAY = '1' -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"INTERVAL DAY\"", @@ -2793,7 +2793,7 @@ SELECT INTERVAL '1' DAY > '1' -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"INTERVAL DAY\"", @@ -2815,7 +2815,7 @@ SELECT '1' < INTERVAL '1' DAY -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"STRING\"", @@ -2837,7 +2837,7 @@ SELECT '1' = INTERVAL '1' DAY -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"STRING\"", @@ -2859,7 +2859,7 @@ SELECT '1' > INTERVAL '1' DAY -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"STRING\"", @@ -2881,7 +2881,7 @@ SELECT INTERVAL '1' YEAR < '1' -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"INTERVAL YEAR\"", @@ -2903,7 +2903,7 @@ SELECT INTERVAL '1' YEAR = '1' -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"INTERVAL YEAR\"", @@ -2925,7 +2925,7 @@ SELECT INTERVAL '1' YEAR > '1' -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"INTERVAL YEAR\"", @@ -2947,7 +2947,7 @@ SELECT '1' < INTERVAL '1' YEAR -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"STRING\"", @@ -2969,7 +2969,7 @@ SELECT '1' = INTERVAL '1' YEAR -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"STRING\"", @@ -2991,7 +2991,7 @@ SELECT '1' > INTERVAL '1' YEAR -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"STRING\"", @@ -3027,7 +3027,7 @@ SELECT array(INTERVAL 1 MONTH, INTERVAL 20 DAYS) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "(\"INTERVAL MONTH\" or \"INTERVAL DAY\")", @@ -3063,7 +3063,7 @@ SELECT coalesce(INTERVAL 1 MONTH, INTERVAL 20 DAYS) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "(\"INTERVAL MONTH\" or \"INTERVAL DAY\")", @@ -3127,7 +3127,7 @@ SELECT div(INTERVAL '1' MONTH, INTERVAL '-1' DAY) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"INTERVAL MONTH\"", diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/nonansi/literals.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/nonansi/literals.sql.out index 570cfb73444e..adaefb1c41ad 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/nonansi/literals.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/nonansi/literals.sql.out @@ -32,7 +32,7 @@ select 128Y -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_NUMERIC_LITERAL_RANGE", + "condition" : "INVALID_NUMERIC_LITERAL_RANGE", "sqlState" : "22003", "messageParameters" : { "maxValue" : "127", @@ -69,7 +69,7 @@ select 32768S -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_NUMERIC_LITERAL_RANGE", + "condition" : "INVALID_NUMERIC_LITERAL_RANGE", "sqlState" : "22003", "messageParameters" : { "maxValue" : "32767", @@ -106,7 +106,7 @@ select 9223372036854775808L -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_NUMERIC_LITERAL_RANGE", + "condition" : "INVALID_NUMERIC_LITERAL_RANGE", "sqlState" : "22003", "messageParameters" : { "maxValue" : "9223372036854775807", @@ -157,7 +157,7 @@ select 1234567890123456789012345678901234567890 -- !query analysis org.apache.spark.SparkArithmeticException { - "errorClass" : "DECIMAL_PRECISION_EXCEEDS_MAX_PRECISION", + "condition" : "DECIMAL_PRECISION_EXCEEDS_MAX_PRECISION", "sqlState" : "22003", "messageParameters" : { "maxPrecision" : "38", @@ -171,7 +171,7 @@ select 1234567890123456789012345678901234567890.0 -- !query analysis org.apache.spark.SparkArithmeticException { - "errorClass" : "DECIMAL_PRECISION_EXCEEDS_MAX_PRECISION", + "condition" : "DECIMAL_PRECISION_EXCEEDS_MAX_PRECISION", "sqlState" : "22003", "messageParameters" : { "maxPrecision" : "38", @@ -199,7 +199,7 @@ select -3.4028235E39f -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_NUMERIC_LITERAL_RANGE", + "condition" : "INVALID_NUMERIC_LITERAL_RANGE", "sqlState" : "22003", "messageParameters" : { "maxValue" : "3.4028234663852886E+38", @@ -236,7 +236,7 @@ select .e3 -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'.'", @@ -250,7 +250,7 @@ select 1E309, -1E309 -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_NUMERIC_LITERAL_RANGE", + "condition" : "INVALID_NUMERIC_LITERAL_RANGE", "sqlState" : "22003", "messageParameters" : { "maxValue" : "1.7976931348623157E+308", @@ -351,7 +351,7 @@ select date 'mar 11 2016' -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_TYPED_LITERAL", + "condition" : "INVALID_TYPED_LITERAL", "sqlState" : "42604", "messageParameters" : { "value" : "'mar 11 2016'", @@ -378,7 +378,7 @@ select timestamp '2016-33-11 20:54:00.000' -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_TYPED_LITERAL", + "condition" : "INVALID_TYPED_LITERAL", "sqlState" : "42604", "messageParameters" : { "value" : "'2016-33-11 20:54:00.000'", @@ -399,7 +399,7 @@ select GEO '(10,-6)' -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "UNSUPPORTED_TYPED_LITERAL", + "condition" : "UNSUPPORTED_TYPED_LITERAL", "sqlState" : "0A000", "messageParameters" : { "supportedTypes" : "\"DATE\", \"TIMESTAMP_NTZ\", \"TIMESTAMP_LTZ\", \"TIMESTAMP\", \"INTERVAL\", \"X\"", @@ -427,7 +427,7 @@ select 1.20E-38BD -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "DECIMAL_PRECISION_EXCEEDS_MAX_PRECISION", + "condition" : "DECIMAL_PRECISION_EXCEEDS_MAX_PRECISION", "sqlState" : "22003", "messageParameters" : { "maxPrecision" : "38", @@ -455,7 +455,7 @@ select X'XuZ' -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_TYPED_LITERAL", + "condition" : "INVALID_TYPED_LITERAL", "sqlState" : "42604", "messageParameters" : { "value" : "'XuZ'", @@ -483,7 +483,7 @@ select +date '1999-01-01' -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"DATE '1999-01-01'\"", @@ -507,7 +507,7 @@ select +timestamp '1999-01-01' -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"TIMESTAMP '1999-01-01 00:00:00'\"", @@ -538,7 +538,7 @@ select +map(1, 2) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"map(1, 2)\"", @@ -562,7 +562,7 @@ select +array(1,2) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"array(1, 2)\"", @@ -586,7 +586,7 @@ select +named_struct('a', 1, 'b', 'spark') -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"named_struct(a, 1, b, spark)\"", @@ -610,7 +610,7 @@ select +X'1' -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"X'01'\"", @@ -634,7 +634,7 @@ select -date '1999-01-01' -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"DATE '1999-01-01'\"", @@ -658,7 +658,7 @@ select -timestamp '1999-01-01' -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"TIMESTAMP '1999-01-01 00:00:00'\"", @@ -682,7 +682,7 @@ select -x'2379ACFe' -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"X'2379ACFE'\"", diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/nonansi/map.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/nonansi/map.sql.out index cd8f0e043b9a..827ed30a4b0f 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/nonansi/map.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/nonansi/map.sql.out @@ -60,7 +60,7 @@ select map_contains_key(map('1', 'a', '2', 'b'), 1) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.MAP_FUNCTION_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.MAP_FUNCTION_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "\"MAP\"", @@ -84,7 +84,7 @@ select map_contains_key(map(1, 'a', 2, 'b'), '1') -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.MAP_FUNCTION_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.MAP_FUNCTION_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "\"MAP\"", diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/nonansi/string-functions.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/nonansi/string-functions.sql.out index 98664dedf820..02c7fb9292f4 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/nonansi/string-functions.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/nonansi/string-functions.sql.out @@ -4,7 +4,7 @@ select concat_ws() -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", + "condition" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", "sqlState" : "42605", "messageParameters" : { "actualNum" : "0", @@ -20,7 +20,7 @@ select format_string() -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", + "condition" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", "sqlState" : "42605", "messageParameters" : { "actualNum" : "0", @@ -779,7 +779,7 @@ select decode() -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", + "condition" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", "sqlState" : "42605", "messageParameters" : { "actualNum" : "0", @@ -802,7 +802,7 @@ select decode(encode('abc', 'utf-8')) -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", + "condition" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", "sqlState" : "42605", "messageParameters" : { "actualNum" : "1", @@ -1508,7 +1508,7 @@ select to_binary('abc', 1) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.INVALID_ARG_VALUE", + "condition" : "DATATYPE_MISMATCH.INVALID_ARG_VALUE", "sqlState" : "42K09", "messageParameters" : { "inputName" : "fmt", @@ -1532,7 +1532,7 @@ select to_binary('abc', 'invalidFormat') -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.INVALID_ARG_VALUE", + "condition" : "DATATYPE_MISMATCH.INVALID_ARG_VALUE", "sqlState" : "42K09", "messageParameters" : { "inputName" : "fmt", @@ -1564,7 +1564,7 @@ SELECT to_binary('abc', fmtField) FROM fmtTable -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.NON_FOLDABLE_INPUT", + "condition" : "DATATYPE_MISMATCH.NON_FOLDABLE_INPUT", "sqlState" : "42K09", "messageParameters" : { "inputExpr" : "\"fmtField\"", diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/nonansi/timestamp.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/nonansi/timestamp.sql.out index dcfd783b648f..764e0f2a61ae 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/nonansi/timestamp.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/nonansi/timestamp.sql.out @@ -10,7 +10,7 @@ select timestamp '2019-01-01中文' -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_TYPED_LITERAL", + "condition" : "INVALID_TYPED_LITERAL", "sqlState" : "42604", "messageParameters" : { "value" : "'2019-01-01中文'", @@ -31,7 +31,7 @@ select timestamp'4294967297' -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_TYPED_LITERAL", + "condition" : "INVALID_TYPED_LITERAL", "sqlState" : "42604", "messageParameters" : { "value" : "'4294967297'", @@ -52,7 +52,7 @@ select timestamp'2021-01-01T12:30:4294967297.123456' -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_TYPED_LITERAL", + "condition" : "INVALID_TYPED_LITERAL", "sqlState" : "42604", "messageParameters" : { "value" : "'2021-01-01T12:30:4294967297.123456'", @@ -553,7 +553,7 @@ select timestamp'2011-11-11 11:11:11' - '2011-11-11 11:11:10' -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"2011-11-11 11:11:10\"", @@ -577,7 +577,7 @@ select '2011-11-11 11:11:11' - timestamp'2011-11-11 11:11:10' -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"2011-11-11 11:11:11\"", @@ -621,7 +621,7 @@ select str - timestamp'2011-11-11 11:11:11' from ts_view -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"str\"", @@ -645,7 +645,7 @@ select timestamp'2011-11-11 11:11:11' - str from ts_view -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"str\"", @@ -669,7 +669,7 @@ select timestamp'2011-11-11 11:11:11' + '1' -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP\"", @@ -691,7 +691,7 @@ select '1' + timestamp'2011-11-11 11:11:11' -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DOUBLE\"", @@ -713,7 +713,7 @@ select timestamp'2011-11-11 11:11:11' + null -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP\"", @@ -735,7 +735,7 @@ select null + timestamp'2011-11-11 11:11:11' -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"VOID\"", @@ -842,7 +842,7 @@ select timestampadd('MONTH', -1, timestamp'2022-02-14 01:02:03') -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", + "condition" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", "sqlState" : "22023", "messageParameters" : { "functionName" : "`timestampadd`", @@ -864,7 +864,7 @@ select timestampadd('SECOND', -1, date'2022-02-15') -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", + "condition" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", "sqlState" : "22023", "messageParameters" : { "functionName" : "`timestampadd`", @@ -910,7 +910,7 @@ select timestampdiff('MINUTE', timestamp'2022-02-14 01:02:03', timestamp'2022-02 -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", + "condition" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", "sqlState" : "22023", "messageParameters" : { "functionName" : "`timestampdiff`", @@ -932,7 +932,7 @@ select timestampdiff('YEAR', date'2022-02-15', date'2023-02-15') -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", + "condition" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", "sqlState" : "22023", "messageParameters" : { "functionName" : "`timestampdiff`", @@ -978,7 +978,7 @@ select timediff('MINUTE', timestamp'2023-02-14 01:02:03', timestamp'2023-02-14 0 -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", + "condition" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", "sqlState" : "22023", "messageParameters" : { "functionName" : "`timediff`", @@ -1000,7 +1000,7 @@ select timediff('YEAR', date'2020-02-15', date'2023-02-15') -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", + "condition" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", "sqlState" : "22023", "messageParameters" : { "functionName" : "`timediff`", diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/nonansi/try_arithmetic.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/nonansi/try_arithmetic.sql.out index caf997f6ccbb..a296c9f94cc8 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/nonansi/try_arithmetic.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/nonansi/try_arithmetic.sql.out @@ -150,7 +150,7 @@ SELECT try_add(interval 2 year, interval 2 second) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"INTERVAL '2' YEAR\"", diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/order-by-all.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/order-by-all.sql.out index b1447307d854..10c7da465402 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/order-by-all.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/order-by-all.sql.out @@ -234,7 +234,7 @@ from values('Lisa', 'Sales', 10000, 35) as T(name, dept, salary, age) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`all`", diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/order-by-ordinal.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/order-by-ordinal.sql.out index f1f189517dea..96569abb8b34 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/order-by-ordinal.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/order-by-ordinal.sql.out @@ -79,7 +79,7 @@ select * from data order by 0 -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "ORDER_BY_POS_OUT_OF_RANGE", + "condition" : "ORDER_BY_POS_OUT_OF_RANGE", "sqlState" : "42805", "messageParameters" : { "index" : "0", @@ -100,7 +100,7 @@ select * from data order by -1 -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "ORDER_BY_POS_OUT_OF_RANGE", + "condition" : "ORDER_BY_POS_OUT_OF_RANGE", "sqlState" : "42805", "messageParameters" : { "index" : "-1", @@ -121,7 +121,7 @@ select * from data order by 3 -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "ORDER_BY_POS_OUT_OF_RANGE", + "condition" : "ORDER_BY_POS_OUT_OF_RANGE", "sqlState" : "42805", "messageParameters" : { "index" : "3", diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/percentiles.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/percentiles.sql.out index eb8102afa47e..3088e93ead21 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/percentiles.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/percentiles.sql.out @@ -156,7 +156,7 @@ FROM aggr -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_SQL_SYNTAX.FUNCTION_WITH_UNSUPPORTED_SYNTAX", + "condition" : "INVALID_SQL_SYNTAX.FUNCTION_WITH_UNSUPPORTED_SYNTAX", "sqlState" : "42000", "messageParameters" : { "prettyName" : "`round`", @@ -179,7 +179,7 @@ FROM aggr -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_SQL_SYNTAX.FUNCTION_WITH_UNSUPPORTED_SYNTAX", + "condition" : "INVALID_SQL_SYNTAX.FUNCTION_WITH_UNSUPPORTED_SYNTAX", "sqlState" : "42000", "messageParameters" : { "prettyName" : "`round`", @@ -202,7 +202,7 @@ FROM aggr -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_SQL_SYNTAX.FUNCTION_WITH_UNSUPPORTED_SYNTAX", + "condition" : "INVALID_SQL_SYNTAX.FUNCTION_WITH_UNSUPPORTED_SYNTAX", "sqlState" : "42000", "messageParameters" : { "prettyName" : "`percentile`", @@ -225,7 +225,7 @@ FROM aggr -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_SQL_SYNTAX.FUNCTION_WITH_UNSUPPORTED_SYNTAX", + "condition" : "INVALID_SQL_SYNTAX.FUNCTION_WITH_UNSUPPORTED_SYNTAX", "sqlState" : "42000", "messageParameters" : { "prettyName" : "`percentile`", @@ -248,7 +248,7 @@ FROM aggr -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_WITHIN_GROUP_EXPRESSION.DISTINCT_UNSUPPORTED", + "condition" : "INVALID_WITHIN_GROUP_EXPRESSION.DISTINCT_UNSUPPORTED", "sqlState" : "42K0K", "messageParameters" : { "funcName" : "`percentile_cont`" @@ -270,7 +270,7 @@ FROM aggr -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_WITHIN_GROUP_EXPRESSION.DISTINCT_UNSUPPORTED", + "condition" : "INVALID_WITHIN_GROUP_EXPRESSION.DISTINCT_UNSUPPORTED", "sqlState" : "42K0K", "messageParameters" : { "funcName" : "`percentile_cont`" @@ -292,7 +292,7 @@ FROM aggr -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", + "condition" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", "sqlState" : "42605", "messageParameters" : { "actualNum" : "0", @@ -317,7 +317,7 @@ FROM aggr -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", + "condition" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", "sqlState" : "42605", "messageParameters" : { "actualNum" : "0", @@ -342,7 +342,7 @@ FROM aggr -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_WITHIN_GROUP_EXPRESSION.WITHIN_GROUP_MISSING", + "condition" : "INVALID_WITHIN_GROUP_EXPRESSION.WITHIN_GROUP_MISSING", "sqlState" : "42K0K", "messageParameters" : { "funcName" : "`percentile_cont`" @@ -364,7 +364,7 @@ FROM aggr -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_WITHIN_GROUP_EXPRESSION.WITHIN_GROUP_MISSING", + "condition" : "INVALID_WITHIN_GROUP_EXPRESSION.WITHIN_GROUP_MISSING", "sqlState" : "42K0K", "messageParameters" : { "funcName" : "`percentile_cont`" @@ -386,7 +386,7 @@ FROM aggr -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_WITHIN_GROUP_EXPRESSION.WRONG_NUM_ORDERINGS", + "condition" : "INVALID_WITHIN_GROUP_EXPRESSION.WRONG_NUM_ORDERINGS", "sqlState" : "42K0K", "messageParameters" : { "actualNum" : "2", @@ -410,7 +410,7 @@ FROM aggr -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", + "condition" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", "sqlState" : "42605", "messageParameters" : { "actualNum" : "2", @@ -485,7 +485,7 @@ ORDER BY salary -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INVALID_WINDOW_SPEC_FOR_AGGREGATION_FUNC", + "condition" : "INVALID_WINDOW_SPEC_FOR_AGGREGATION_FUNC", "sqlState" : "42601", "messageParameters" : { "aggFunc" : "\"percentile_cont(salary, 0.25)\"" @@ -512,7 +512,7 @@ ORDER BY salary -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INVALID_WINDOW_SPEC_FOR_AGGREGATION_FUNC", + "condition" : "INVALID_WINDOW_SPEC_FOR_AGGREGATION_FUNC", "sqlState" : "42601", "messageParameters" : { "aggFunc" : "\"percentile_disc(salary, 0.25)\"" @@ -538,7 +538,7 @@ ORDER BY salary -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INVALID_WINDOW_SPEC_FOR_AGGREGATION_FUNC", + "condition" : "INVALID_WINDOW_SPEC_FOR_AGGREGATION_FUNC", "sqlState" : "42601", "messageParameters" : { "aggFunc" : "\"median(salary)\"" @@ -565,7 +565,7 @@ ORDER BY salary -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INVALID_WINDOW_SPEC_FOR_AGGREGATION_FUNC", + "condition" : "INVALID_WINDOW_SPEC_FOR_AGGREGATION_FUNC", "sqlState" : "42601", "messageParameters" : { "aggFunc" : "\"percentile_cont(salary, 0.25)\"" @@ -592,7 +592,7 @@ ORDER BY salary -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INVALID_WINDOW_SPEC_FOR_AGGREGATION_FUNC", + "condition" : "INVALID_WINDOW_SPEC_FOR_AGGREGATION_FUNC", "sqlState" : "42601", "messageParameters" : { "aggFunc" : "\"percentile_disc(salary, 0.25)\"" @@ -618,7 +618,7 @@ ORDER BY salary -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INVALID_WINDOW_SPEC_FOR_AGGREGATION_FUNC", + "condition" : "INVALID_WINDOW_SPEC_FOR_AGGREGATION_FUNC", "sqlState" : "42601", "messageParameters" : { "aggFunc" : "\"median(salary)\"" @@ -701,7 +701,7 @@ ORDER BY salary -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INVALID_WINDOW_SPEC_FOR_AGGREGATION_FUNC", + "condition" : "INVALID_WINDOW_SPEC_FOR_AGGREGATION_FUNC", "sqlState" : "42601", "messageParameters" : { "aggFunc" : "\"percentile_cont(salary, 0.25)\"" @@ -729,7 +729,7 @@ ORDER BY salary -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INVALID_WINDOW_SPEC_FOR_AGGREGATION_FUNC", + "condition" : "INVALID_WINDOW_SPEC_FOR_AGGREGATION_FUNC", "sqlState" : "42601", "messageParameters" : { "aggFunc" : "\"percentile_disc(salary, 0.25)\"" @@ -756,7 +756,7 @@ ORDER BY salary -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INVALID_WINDOW_SPEC_FOR_AGGREGATION_FUNC", + "condition" : "INVALID_WINDOW_SPEC_FOR_AGGREGATION_FUNC", "sqlState" : "42601", "messageParameters" : { "aggFunc" : "\"median(salary)\"" @@ -784,7 +784,7 @@ ORDER BY salary -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INVALID_WINDOW_SPEC_FOR_AGGREGATION_FUNC", + "condition" : "INVALID_WINDOW_SPEC_FOR_AGGREGATION_FUNC", "sqlState" : "42601", "messageParameters" : { "aggFunc" : "\"percentile_cont(salary, 0.25)\"" @@ -812,7 +812,7 @@ ORDER BY salary -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INVALID_WINDOW_SPEC_FOR_AGGREGATION_FUNC", + "condition" : "INVALID_WINDOW_SPEC_FOR_AGGREGATION_FUNC", "sqlState" : "42601", "messageParameters" : { "aggFunc" : "\"percentile_disc(salary, 0.25)\"" @@ -839,7 +839,7 @@ ORDER BY salary -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INVALID_WINDOW_SPEC_FOR_AGGREGATION_FUNC", + "condition" : "INVALID_WINDOW_SPEC_FOR_AGGREGATION_FUNC", "sqlState" : "42601", "messageParameters" : { "aggFunc" : "\"median(salary)\"" @@ -1166,7 +1166,7 @@ FROM values (12, 0.25), (13, 0.25), (22, 0.25) as v(a, b) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.NON_FOLDABLE_INPUT", + "condition" : "DATATYPE_MISMATCH.NON_FOLDABLE_INPUT", "sqlState" : "42K09", "messageParameters" : { "inputExpr" : "\"b\"", diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/pipe-operators.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/pipe-operators.sql.out index b231199cc473..8089d7c4e962 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/pipe-operators.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/pipe-operators.sql.out @@ -381,7 +381,7 @@ from t -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'from'", @@ -638,7 +638,7 @@ table t -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "PIPE_OPERATOR_CONTAINS_AGGREGATE_FUNCTION", + "condition" : "PIPE_OPERATOR_CONTAINS_AGGREGATE_FUNCTION", "sqlState" : "0A000", "messageParameters" : { "clause" : "SELECT", @@ -660,7 +660,7 @@ table t -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "PIPE_OPERATOR_CONTAINS_AGGREGATE_FUNCTION", + "condition" : "PIPE_OPERATOR_CONTAINS_AGGREGATE_FUNCTION", "sqlState" : "0A000", "messageParameters" : { "clause" : "SELECT", @@ -811,7 +811,7 @@ table t -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "PIPE_OPERATOR_CONTAINS_AGGREGATE_FUNCTION", + "condition" : "PIPE_OPERATOR_CONTAINS_AGGREGATE_FUNCTION", "sqlState" : "0A000", "messageParameters" : { "clause" : "EXTEND", @@ -833,7 +833,7 @@ table t -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'as'", @@ -848,7 +848,7 @@ table t -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_USAGE_OF_STAR_OR_REGEX", + "condition" : "INVALID_USAGE_OF_STAR_OR_REGEX", "sqlState" : "42000", "messageParameters" : { "elem" : "'*'", @@ -1042,7 +1042,7 @@ table t -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`z`", @@ -1064,7 +1064,7 @@ table t -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'as'", @@ -1079,7 +1079,7 @@ select col from st -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0035", + "condition" : "_LEGACY_ERROR_TEMP_0035", "messageParameters" : { "message" : "SQL pipe syntax |> SET operator with multi-part assignment key (only single-part keys are allowed)" }, @@ -1149,7 +1149,7 @@ table t -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`z`", @@ -1171,7 +1171,7 @@ table st -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'.'", @@ -1186,7 +1186,7 @@ table st -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`col.i1`", @@ -1208,7 +1208,7 @@ select 1 as x, 2 as y, 3 as z -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "EXCEPT_OVERLAPPING_COLUMNS", + "condition" : "EXCEPT_OVERLAPPING_COLUMNS", "sqlState" : "42702", "messageParameters" : { "columns" : "z, y, z" @@ -1287,7 +1287,7 @@ table t -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "','", @@ -1302,7 +1302,7 @@ table t -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'1'", @@ -1317,7 +1317,7 @@ table t -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_IDENTIFIER", + "condition" : "INVALID_IDENTIFIER", "sqlState" : "42602", "messageParameters" : { "ident" : "u-v" @@ -1331,7 +1331,7 @@ table t -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'@'", @@ -1346,7 +1346,7 @@ table t -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'#'", @@ -1478,7 +1478,7 @@ table t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INVALID_WHERE_CONDITION", + "condition" : "INVALID_WHERE_CONDITION", "sqlState" : "42903", "messageParameters" : { "condition" : "\"(sum(x) = 1)\"", @@ -1500,7 +1500,7 @@ table t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INVALID_WHERE_CONDITION", + "condition" : "INVALID_WHERE_CONDITION", "sqlState" : "42903", "messageParameters" : { "condition" : "\"((y = abc) OR ((length(y) + sum(x)) = 1))\"", @@ -1522,7 +1522,7 @@ table t -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "_LEGACY_ERROR_TEMP_1034", + "condition" : "_LEGACY_ERROR_TEMP_1034", "messageParameters" : { "clauseName" : "WHERE" } @@ -1536,7 +1536,7 @@ table t -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "NOT_ALLOWED_IN_PIPE_OPERATOR_WHERE.WINDOW_CLAUSE", + "condition" : "NOT_ALLOWED_IN_PIPE_OPERATOR_WHERE.WINDOW_CLAUSE", "sqlState" : "42601", "queryContext" : [ { "objectType" : "", @@ -1553,7 +1553,7 @@ select * from t where sum(x) over (partition by y) = 1 -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "_LEGACY_ERROR_TEMP_1034", + "condition" : "_LEGACY_ERROR_TEMP_1034", "messageParameters" : { "clauseName" : "WHERE" } @@ -1567,7 +1567,7 @@ table t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`y`", @@ -1591,7 +1591,7 @@ table t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`y`", @@ -1615,7 +1615,7 @@ table t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`y`", @@ -1639,7 +1639,7 @@ table t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`y`", @@ -1661,7 +1661,7 @@ org.apache.spark.sql.catalyst.ExtendedAnalysisException -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`y`", @@ -1830,7 +1830,7 @@ table courseSales -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`year`", @@ -1855,7 +1855,7 @@ table courseSales -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "NON_LITERAL_PIVOT_VALUES", + "condition" : "NON_LITERAL_PIVOT_VALUES", "sqlState" : "42K08", "messageParameters" : { "expression" : "\"course\"" @@ -1876,7 +1876,7 @@ table courseSales -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "NOT_ALLOWED_IN_FROM.UNPIVOT_WITH_PIVOT", + "condition" : "NOT_ALLOWED_IN_FROM.UNPIVOT_WITH_PIVOT", "sqlState" : "42601", "queryContext" : [ { "objectType" : "", @@ -1901,7 +1901,7 @@ table courseSales -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "NOT_ALLOWED_IN_FROM.UNPIVOT_WITH_PIVOT", + "condition" : "NOT_ALLOWED_IN_FROM.UNPIVOT_WITH_PIVOT", "sqlState" : "42601", "queryContext" : [ { "objectType" : "", @@ -1927,7 +1927,7 @@ table courseSales -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'pivot'", @@ -1952,7 +1952,7 @@ table courseSales -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'unpivot'", @@ -2009,7 +2009,7 @@ table t -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0014", + "condition" : "_LEGACY_ERROR_TEMP_0014", "queryContext" : [ { "objectType" : "", "objectName" : "", @@ -2026,7 +2026,7 @@ table t -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0064", + "condition" : "_LEGACY_ERROR_TEMP_0064", "messageParameters" : { "msg" : "Sampling fraction (-1.0) must be on interval [0, 1]" }, @@ -2046,7 +2046,7 @@ table t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INVALID_LIMIT_LIKE_EXPRESSION.IS_NEGATIVE", + "condition" : "INVALID_LIMIT_LIKE_EXPRESSION.IS_NEGATIVE", "sqlState" : "42K0E", "messageParameters" : { "expr" : "\"-5\"", @@ -2069,7 +2069,7 @@ table t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INVALID_LIMIT_LIKE_EXPRESSION.IS_UNFOLDABLE", + "condition" : "INVALID_LIMIT_LIKE_EXPRESSION.IS_UNFOLDABLE", "sqlState" : "42K0E", "messageParameters" : { "expr" : "\"x\"", @@ -2091,7 +2091,7 @@ table t -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0064", + "condition" : "_LEGACY_ERROR_TEMP_0064", "messageParameters" : { "msg" : "Sampling fraction (2.0) must be on interval [0, 1]" }, @@ -2111,7 +2111,7 @@ table t -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0015", + "condition" : "_LEGACY_ERROR_TEMP_0015", "messageParameters" : { "msg" : "byteLengthLiteral" }, @@ -2131,7 +2131,7 @@ table t -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0016", + "condition" : "_LEGACY_ERROR_TEMP_0016", "messageParameters" : { "bytesStr" : "200" }, @@ -2707,7 +2707,7 @@ table join_test_t1 -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'inner'", @@ -2723,7 +2723,7 @@ table join_test_t1 -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`join_test_t1`.`a`", @@ -2745,7 +2745,7 @@ table join_test_t1 jt -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'jt'", @@ -2928,7 +2928,7 @@ table t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "NUM_COLUMNS_MISMATCH", + "condition" : "NUM_COLUMNS_MISMATCH", "sqlState" : "42826", "messageParameters" : { "firstNumColumns" : "1", @@ -2952,7 +2952,7 @@ table t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "second", @@ -3110,7 +3110,7 @@ order by y -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'order'", @@ -3126,7 +3126,7 @@ table t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`x`", @@ -3149,7 +3149,7 @@ table t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`x`", @@ -3171,7 +3171,7 @@ table t -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "MULTIPLE_QUERY_RESULT_CLAUSES_WITH_PIPE_OPERATORS", + "condition" : "MULTIPLE_QUERY_RESULT_CLAUSES_WITH_PIPE_OPERATORS", "sqlState" : "42000", "messageParameters" : { "clause1" : "ORDER BY", @@ -3193,7 +3193,7 @@ table t -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "UNSUPPORTED_FEATURE.COMBINATION_QUERY_RESULT_CLAUSES", + "condition" : "UNSUPPORTED_FEATURE.COMBINATION_QUERY_RESULT_CLAUSES", "sqlState" : "0A000", "queryContext" : [ { "objectType" : "", @@ -3440,7 +3440,7 @@ table other -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "PIPE_OPERATOR_AGGREGATE_EXPRESSION_CONTAINS_NO_AGGREGATE_FUNCTION", + "condition" : "PIPE_OPERATOR_AGGREGATE_EXPRESSION_CONTAINS_NO_AGGREGATE_FUNCTION", "sqlState" : "0A000", "messageParameters" : { "expr" : "a#x" @@ -3461,7 +3461,7 @@ select 3 as x, 4 as y -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "UNSUPPORTED_FEATURE.PIPE_OPERATOR_AGGREGATE_UNSUPPORTED_CASE", + "condition" : "UNSUPPORTED_FEATURE.PIPE_OPERATOR_AGGREGATE_UNSUPPORTED_CASE", "sqlState" : "0A000", "messageParameters" : { "case" : "GROUP BY ALL" @@ -3483,7 +3483,7 @@ table courseSales -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "UNSUPPORTED_FEATURE.PIPE_OPERATOR_AGGREGATE_UNSUPPORTED_CASE", + "condition" : "UNSUPPORTED_FEATURE.PIPE_OPERATOR_AGGREGATE_UNSUPPORTED_CASE", "sqlState" : "0A000", "messageParameters" : { "case" : "GROUP BY ROLLUP" @@ -3505,7 +3505,7 @@ table courseSales -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "UNSUPPORTED_FEATURE.PIPE_OPERATOR_AGGREGATE_UNSUPPORTED_CASE", + "condition" : "UNSUPPORTED_FEATURE.PIPE_OPERATOR_AGGREGATE_UNSUPPORTED_CASE", "sqlState" : "0A000", "messageParameters" : { "case" : "GROUP BY CUBE" @@ -3527,7 +3527,7 @@ table courseSales -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "UNSUPPORTED_FEATURE.PIPE_OPERATOR_AGGREGATE_UNSUPPORTED_CASE", + "condition" : "UNSUPPORTED_FEATURE.PIPE_OPERATOR_AGGREGATE_UNSUPPORTED_CASE", "sqlState" : "0A000", "messageParameters" : { "case" : "GROUPING SETS" @@ -3550,7 +3550,7 @@ table courseSales -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "UNSUPPORTED_FEATURE.PIPE_OPERATOR_AGGREGATE_UNSUPPORTED_CASE", + "condition" : "UNSUPPORTED_FEATURE.PIPE_OPERATOR_AGGREGATE_UNSUPPORTED_CASE", "sqlState" : "0A000", "messageParameters" : { "case" : "GROUPING" @@ -3573,7 +3573,7 @@ table courseSales -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "UNSUPPORTED_FEATURE.PIPE_OPERATOR_AGGREGATE_UNSUPPORTED_CASE", + "condition" : "UNSUPPORTED_FEATURE.PIPE_OPERATOR_AGGREGATE_UNSUPPORTED_CASE", "sqlState" : "0A000", "messageParameters" : { "case" : "GROUPING_ID" @@ -3594,7 +3594,7 @@ select 1 as x, 2 as y -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "')'", @@ -3609,7 +3609,7 @@ table other -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "PIPE_OPERATOR_AGGREGATE_EXPRESSION_CONTAINS_NO_AGGREGATE_FUNCTION", + "condition" : "PIPE_OPERATOR_AGGREGATE_EXPRESSION_CONTAINS_NO_AGGREGATE_FUNCTION", "sqlState" : "0A000", "messageParameters" : { "expr" : "a#x" @@ -3630,7 +3630,7 @@ table other -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "PIPE_OPERATOR_CONTAINS_AGGREGATE_FUNCTION", + "condition" : "PIPE_OPERATOR_CONTAINS_AGGREGATE_FUNCTION", "sqlState" : "0A000", "messageParameters" : { "clause" : "SELECT", @@ -3652,7 +3652,7 @@ table other -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0035", + "condition" : "_LEGACY_ERROR_TEMP_0035", "messageParameters" : { "message" : "The AGGREGATE clause requires a list of aggregate expressions or a list of grouping expressions, or both" }, @@ -3672,7 +3672,7 @@ table other -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`group`", @@ -3694,7 +3694,7 @@ table other -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'group'", @@ -3709,7 +3709,7 @@ table other -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "UNSUPPORTED_FEATURE.PIPE_OPERATOR_AGGREGATE_UNSUPPORTED_CASE", + "condition" : "UNSUPPORTED_FEATURE.PIPE_OPERATOR_AGGREGATE_UNSUPPORTED_CASE", "sqlState" : "0A000", "messageParameters" : { "case" : "window functions; please update the query to move the window functions to a subsequent |> SELECT operator instead" @@ -3732,7 +3732,7 @@ select 1 x, 2 y, 3 z -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "AMBIGUOUS_REFERENCE", + "condition" : "AMBIGUOUS_REFERENCE", "sqlState" : "42704", "messageParameters" : { "name" : "`x`", @@ -3844,7 +3844,7 @@ table windowTestData -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "MISSING_WINDOW_SPECIFICATION", + "condition" : "MISSING_WINDOW_SPECIFICATION", "sqlState" : "42P20", "messageParameters" : { "docroot" : "https://spark.apache.org/docs/latest", @@ -3916,7 +3916,7 @@ table windowTestData -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "MISSING_WINDOW_SPECIFICATION", + "condition" : "MISSING_WINDOW_SPECIFICATION", "sqlState" : "42P20", "messageParameters" : { "docroot" : "https://spark.apache.org/docs/latest", @@ -3933,7 +3933,7 @@ table windowTestData -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "MISSING_WINDOW_SPECIFICATION", + "condition" : "MISSING_WINDOW_SPECIFICATION", "sqlState" : "42P20", "messageParameters" : { "docroot" : "https://spark.apache.org/docs/latest", @@ -3950,7 +3950,7 @@ table windowTestData -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "MISSING_WINDOW_SPECIFICATION", + "condition" : "MISSING_WINDOW_SPECIFICATION", "sqlState" : "42P20", "messageParameters" : { "docroot" : "https://spark.apache.org/docs/latest", @@ -3967,7 +3967,7 @@ table windowTestData -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "MISSING_WINDOW_SPECIFICATION", + "condition" : "MISSING_WINDOW_SPECIFICATION", "sqlState" : "42P20", "messageParameters" : { "docroot" : "https://spark.apache.org/docs/latest", @@ -3999,7 +3999,7 @@ limit 100 -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "TABLE_OR_VIEW_NOT_FOUND", + "condition" : "TABLE_OR_VIEW_NOT_FOUND", "sqlState" : "42P01", "messageParameters" : { "relationName" : "`store_returns`" @@ -4038,7 +4038,7 @@ from customer_total_return ctr1 -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "TABLE_OR_VIEW_NOT_FOUND", + "condition" : "TABLE_OR_VIEW_NOT_FOUND", "sqlState" : "42P01", "messageParameters" : { "relationName" : "`store_returns`" @@ -4138,7 +4138,7 @@ order by d_week_seq1 -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "TABLE_OR_VIEW_NOT_FOUND", + "condition" : "TABLE_OR_VIEW_NOT_FOUND", "sqlState" : "42P01", "messageParameters" : { "relationName" : "`web_sales`" @@ -4243,7 +4243,7 @@ table wswscs -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "TABLE_OR_VIEW_NOT_FOUND", + "condition" : "TABLE_OR_VIEW_NOT_FOUND", "sqlState" : "42P01", "messageParameters" : { "relationName" : "`web_sales`" @@ -4275,7 +4275,7 @@ limit 100 -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "TABLE_OR_VIEW_NOT_FOUND", + "condition" : "TABLE_OR_VIEW_NOT_FOUND", "sqlState" : "42P01", "messageParameters" : { "relationName" : "`date_dim`" @@ -4306,7 +4306,7 @@ table date_dim -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "TABLE_OR_VIEW_NOT_FOUND", + "condition" : "TABLE_OR_VIEW_NOT_FOUND", "sqlState" : "42P01", "messageParameters" : { "relationName" : "`date_dim`" @@ -4347,7 +4347,7 @@ limit 100 -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "TABLE_OR_VIEW_NOT_FOUND", + "condition" : "TABLE_OR_VIEW_NOT_FOUND", "sqlState" : "42P01", "messageParameters" : { "relationName" : "`web_sales`" @@ -4382,7 +4382,7 @@ table web_sales -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "TABLE_OR_VIEW_NOT_FOUND", + "condition" : "TABLE_OR_VIEW_NOT_FOUND", "sqlState" : "42P01", "messageParameters" : { "relationName" : "`web_sales`" @@ -4447,7 +4447,7 @@ limit 100 -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "TABLE_OR_VIEW_NOT_FOUND", + "condition" : "TABLE_OR_VIEW_NOT_FOUND", "sqlState" : "42P01", "messageParameters" : { "relationName" : "`store_sales`" @@ -4514,7 +4514,7 @@ from store_sales ss1 -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "TABLE_OR_VIEW_NOT_FOUND", + "condition" : "TABLE_OR_VIEW_NOT_FOUND", "sqlState" : "42P01", "messageParameters" : { "relationName" : "`store_sales`" @@ -4588,7 +4588,7 @@ limit 100 -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "TABLE_OR_VIEW_NOT_FOUND", + "condition" : "TABLE_OR_VIEW_NOT_FOUND", "sqlState" : "42P01", "messageParameters" : { "relationName" : "`web_sales`" @@ -4664,7 +4664,7 @@ table web_v1 -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "TABLE_OR_VIEW_NOT_FOUND", + "condition" : "TABLE_OR_VIEW_NOT_FOUND", "sqlState" : "42P01", "messageParameters" : { "relationName" : "`web_sales`" diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/pivot.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/pivot.sql.out index 93f2e240a019..03306257adf1 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/pivot.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/pivot.sql.out @@ -312,7 +312,7 @@ PIVOT ( -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "_LEGACY_ERROR_TEMP_1006", + "condition" : "_LEGACY_ERROR_TEMP_1006", "messageParameters" : { "sql" : "coursesales.earnings" } @@ -330,7 +330,7 @@ PIVOT ( -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "_LEGACY_ERROR_TEMP_1006", + "condition" : "_LEGACY_ERROR_TEMP_1006", "messageParameters" : { "sql" : "__auto_generated_subquery_name.year" } @@ -348,7 +348,7 @@ PIVOT ( -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`year`", @@ -398,7 +398,7 @@ PIVOT ( -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "NESTED_AGGREGATE_FUNCTION", + "condition" : "NESTED_AGGREGATE_FUNCTION", "sqlState" : "42607", "queryContext" : [ { "objectType" : "", @@ -487,7 +487,7 @@ PIVOT ( -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "PIVOT_VALUE_DATA_TYPE_MISMATCH", + "condition" : "PIVOT_VALUE_DATA_TYPE_MISMATCH", "sqlState" : "42K09", "messageParameters" : { "pivotType" : "struct", @@ -506,7 +506,7 @@ PIVOT ( -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`s`", @@ -531,7 +531,7 @@ PIVOT ( -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "NON_LITERAL_PIVOT_VALUES", + "condition" : "NON_LITERAL_PIVOT_VALUES", "sqlState" : "42K08", "messageParameters" : { "expression" : "\"course\"" @@ -740,7 +740,7 @@ PIVOT ( -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "INCOMPARABLE_PIVOT_COLUMN", + "condition" : "INCOMPARABLE_PIVOT_COLUMN", "sqlState" : "42818", "messageParameters" : { "columnName" : "`m`" @@ -761,7 +761,7 @@ PIVOT ( -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "INCOMPARABLE_PIVOT_COLUMN", + "condition" : "INCOMPARABLE_PIVOT_COLUMN", "sqlState" : "42818", "messageParameters" : { "columnName" : "`named_struct('course', __auto_generated_subquery_name`.`course, 'm', __auto_generated_subquery_name`.`m)`" diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/postgreSQL/aggregates_part1.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/postgreSQL/aggregates_part1.sql.out index 0577d73ea6a3..502932626a30 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/postgreSQL/aggregates_part1.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/postgreSQL/aggregates_part1.sql.out @@ -483,7 +483,7 @@ having exists (select 1 from onek b -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "UNSUPPORTED_SUBQUERY_EXPRESSION_CATEGORY.AGGREGATE_FUNCTION_MIXED_OUTER_LOCAL_REFERENCES", + "condition" : "UNSUPPORTED_SUBQUERY_EXPRESSION_CATEGORY.AGGREGATE_FUNCTION_MIXED_OUTER_LOCAL_REFERENCES", "sqlState" : "0A000", "messageParameters" : { "function" : "sum(DISTINCT (outer(a.four) + b.four))" @@ -505,7 +505,7 @@ from tenk1 o -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`o`.`unique1`", diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/postgreSQL/aggregates_part3.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/postgreSQL/aggregates_part3.sql.out index 2256213b63ff..aed4ebf821b7 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/postgreSQL/aggregates_part3.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/postgreSQL/aggregates_part3.sql.out @@ -4,7 +4,7 @@ select max(min(unique1)) from tenk1 -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "NESTED_AGGREGATE_FUNCTION", + "condition" : "NESTED_AGGREGATE_FUNCTION", "sqlState" : "42607", "queryContext" : [ { "objectType" : "", diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/postgreSQL/boolean.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/postgreSQL/boolean.sql.out index 277ab866f97f..c843eb1a0f7f 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/postgreSQL/boolean.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/postgreSQL/boolean.sql.out @@ -414,7 +414,7 @@ INSERT INTO BOOLTBL2 -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_INLINE_TABLE.FAILED_SQL_EXPRESSION_EVALUATION", + "condition" : "INVALID_INLINE_TABLE.FAILED_SQL_EXPRESSION_EVALUATION", "sqlState" : "42000", "messageParameters" : { "sqlExpr" : "\"XXX\"" diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/postgreSQL/create_view.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/postgreSQL/create_view.sql.out index e4e4be8fee7d..f182fa623acc 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/postgreSQL/create_view.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/postgreSQL/create_view.sql.out @@ -40,7 +40,7 @@ CREATE TABLE view_base_table (key int /* PRIMARY KEY */, data varchar(20)) USING -- !query analysis org.apache.spark.sql.catalyst.analysis.TableAlreadyExistsException { - "errorClass" : "TABLE_OR_VIEW_ALREADY_EXISTS", + "condition" : "TABLE_OR_VIEW_ALREADY_EXISTS", "sqlState" : "42P07", "messageParameters" : { "relationName" : "`spark_catalog`.`default`.`view_base_table`" @@ -54,7 +54,7 @@ CREATE VIEW key_dependent_view AS -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "MISSING_AGGREGATION", + "condition" : "MISSING_AGGREGATION", "sqlState" : "42803", "messageParameters" : { "expression" : "\"data\"", @@ -69,7 +69,7 @@ CREATE VIEW key_dependent_view_no_cols AS -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITHOUT_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITHOUT_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`FROM`" @@ -269,7 +269,7 @@ CREATE VIEW v1_temp AS SELECT * FROM temp_table -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_TEMP_OBJ_REFERENCE", + "condition" : "INVALID_TEMP_OBJ_REFERENCE", "sqlState" : "42K0F", "messageParameters" : { "obj" : "VIEW", @@ -315,7 +315,7 @@ CREATE VIEW temp_view_test.v3_temp AS SELECT * FROM temp_table -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_TEMP_OBJ_REFERENCE", + "condition" : "INVALID_TEMP_OBJ_REFERENCE", "sqlState" : "42K0F", "messageParameters" : { "obj" : "VIEW", @@ -360,7 +360,7 @@ CREATE VIEW v4_temp AS -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_TEMP_OBJ_REFERENCE", + "condition" : "INVALID_TEMP_OBJ_REFERENCE", "sqlState" : "42K0F", "messageParameters" : { "obj" : "VIEW", @@ -379,7 +379,7 @@ CREATE VIEW v5_temp AS -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_TEMP_OBJ_REFERENCE", + "condition" : "INVALID_TEMP_OBJ_REFERENCE", "sqlState" : "42K0F", "messageParameters" : { "obj" : "VIEW", @@ -491,7 +491,7 @@ CREATE VIEW v6_temp AS SELECT * FROM base_table WHERE id IN (SELECT id FROM temp -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_TEMP_OBJ_REFERENCE", + "condition" : "INVALID_TEMP_OBJ_REFERENCE", "sqlState" : "42K0F", "messageParameters" : { "obj" : "VIEW", @@ -507,7 +507,7 @@ CREATE VIEW v7_temp AS SELECT t1.id, t2.a FROM base_table t1, (SELECT * FROM tem -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_TEMP_OBJ_REFERENCE", + "condition" : "INVALID_TEMP_OBJ_REFERENCE", "sqlState" : "42K0F", "messageParameters" : { "obj" : "VIEW", @@ -523,7 +523,7 @@ CREATE VIEW v8_temp AS SELECT * FROM base_table WHERE EXISTS (SELECT 1 FROM temp -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_TEMP_OBJ_REFERENCE", + "condition" : "INVALID_TEMP_OBJ_REFERENCE", "sqlState" : "42K0F", "messageParameters" : { "obj" : "VIEW", @@ -539,7 +539,7 @@ CREATE VIEW v9_temp AS SELECT * FROM base_table WHERE NOT EXISTS (SELECT 1 FROM -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_TEMP_OBJ_REFERENCE", + "condition" : "INVALID_TEMP_OBJ_REFERENCE", "sqlState" : "42K0F", "messageParameters" : { "obj" : "VIEW", @@ -555,7 +555,7 @@ CREATE VIEW v10_temp AS SELECT * FROM v7_temp -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "TABLE_OR_VIEW_NOT_FOUND", + "condition" : "TABLE_OR_VIEW_NOT_FOUND", "sqlState" : "42P01", "messageParameters" : { "relationName" : "`v7_temp`" @@ -575,7 +575,7 @@ CREATE VIEW v11_temp AS SELECT t1.id, t2.a FROM base_table t1, v10_temp t2 -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "TABLE_OR_VIEW_NOT_FOUND", + "condition" : "TABLE_OR_VIEW_NOT_FOUND", "sqlState" : "42P01", "messageParameters" : { "relationName" : "`v10_temp`" @@ -595,7 +595,7 @@ CREATE VIEW v12_temp AS SELECT true FROM v11_temp -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "TABLE_OR_VIEW_NOT_FOUND", + "condition" : "TABLE_OR_VIEW_NOT_FOUND", "sqlState" : "42P01", "messageParameters" : { "relationName" : "`v11_temp`" @@ -670,7 +670,7 @@ CREATE VIEW temporal1 AS SELECT * FROM t1 CROSS JOIN tt -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_TEMP_OBJ_REFERENCE", + "condition" : "INVALID_TEMP_OBJ_REFERENCE", "sqlState" : "42K0F", "messageParameters" : { "obj" : "VIEW", @@ -704,7 +704,7 @@ CREATE VIEW temporal2 AS SELECT * FROM t1 INNER JOIN tt ON t1.num = tt.num2 -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_TEMP_OBJ_REFERENCE", + "condition" : "INVALID_TEMP_OBJ_REFERENCE", "sqlState" : "42K0F", "messageParameters" : { "obj" : "VIEW", @@ -738,7 +738,7 @@ CREATE VIEW temporal3 AS SELECT * FROM t1 LEFT JOIN tt ON t1.num = tt.num2 -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_TEMP_OBJ_REFERENCE", + "condition" : "INVALID_TEMP_OBJ_REFERENCE", "sqlState" : "42K0F", "messageParameters" : { "obj" : "VIEW", @@ -772,7 +772,7 @@ CREATE VIEW temporal4 AS SELECT * FROM t1 LEFT JOIN tt ON t1.num = tt.num2 AND t -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_TEMP_OBJ_REFERENCE", + "condition" : "INVALID_TEMP_OBJ_REFERENCE", "sqlState" : "42K0F", "messageParameters" : { "obj" : "VIEW", @@ -788,7 +788,7 @@ CREATE VIEW temporal5 AS SELECT * FROM t1 WHERE num IN (SELECT num FROM t1 WHERE -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_TEMP_OBJ_REFERENCE", + "condition" : "INVALID_TEMP_OBJ_REFERENCE", "sqlState" : "42K0F", "messageParameters" : { "obj" : "VIEW", diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/postgreSQL/date.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/postgreSQL/date.sql.out index d3ac6a3eb2b5..98a3831a7d12 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/postgreSQL/date.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/postgreSQL/date.sql.out @@ -169,7 +169,7 @@ SELECT date '1999 Jan 08' -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_TYPED_LITERAL", + "condition" : "INVALID_TYPED_LITERAL", "sqlState" : "42604", "messageParameters" : { "value" : "'1999 Jan 08'", @@ -190,7 +190,7 @@ SELECT date '1999 08 Jan' -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_TYPED_LITERAL", + "condition" : "INVALID_TYPED_LITERAL", "sqlState" : "42604", "messageParameters" : { "value" : "'1999 08 Jan'", @@ -223,7 +223,7 @@ SELECT date '1999 01 08' -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_TYPED_LITERAL", + "condition" : "INVALID_TYPED_LITERAL", "sqlState" : "42604", "messageParameters" : { "value" : "'1999 01 08'", @@ -244,7 +244,7 @@ SELECT date '1999 08 01' -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_TYPED_LITERAL", + "condition" : "INVALID_TYPED_LITERAL", "sqlState" : "42604", "messageParameters" : { "value" : "'1999 08 01'", @@ -271,7 +271,7 @@ SELECT date '1999 Jan 08' -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_TYPED_LITERAL", + "condition" : "INVALID_TYPED_LITERAL", "sqlState" : "42604", "messageParameters" : { "value" : "'1999 Jan 08'", @@ -292,7 +292,7 @@ SELECT date '1999 08 Jan' -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_TYPED_LITERAL", + "condition" : "INVALID_TYPED_LITERAL", "sqlState" : "42604", "messageParameters" : { "value" : "'1999 08 Jan'", @@ -325,7 +325,7 @@ SELECT date '1999 01 08' -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_TYPED_LITERAL", + "condition" : "INVALID_TYPED_LITERAL", "sqlState" : "42604", "messageParameters" : { "value" : "'1999 01 08'", @@ -346,7 +346,7 @@ SELECT date '1999 08 01' -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_TYPED_LITERAL", + "condition" : "INVALID_TYPED_LITERAL", "sqlState" : "42604", "messageParameters" : { "value" : "'1999 08 01'", @@ -379,7 +379,7 @@ SELECT date '1999 Jan 08' -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_TYPED_LITERAL", + "condition" : "INVALID_TYPED_LITERAL", "sqlState" : "42604", "messageParameters" : { "value" : "'1999 Jan 08'", @@ -400,7 +400,7 @@ SELECT date '1999 08 Jan' -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_TYPED_LITERAL", + "condition" : "INVALID_TYPED_LITERAL", "sqlState" : "42604", "messageParameters" : { "value" : "'1999 08 Jan'", @@ -433,7 +433,7 @@ SELECT date '1999 01 08' -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_TYPED_LITERAL", + "condition" : "INVALID_TYPED_LITERAL", "sqlState" : "42604", "messageParameters" : { "value" : "'1999 01 08'", @@ -454,7 +454,7 @@ SELECT date '1999 08 01' -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_TYPED_LITERAL", + "condition" : "INVALID_TYPED_LITERAL", "sqlState" : "42604", "messageParameters" : { "value" : "'1999 08 01'", diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/postgreSQL/int8.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/postgreSQL/int8.sql.out index 72972469fa6e..97d8cd326eb1 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/postgreSQL/int8.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/postgreSQL/int8.sql.out @@ -634,7 +634,7 @@ SELECT * FROM range(bigint('+4567890123456789'), bigint('+4567890123456799'), 0) -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "FAILED_FUNCTION_CALL", + "condition" : "FAILED_FUNCTION_CALL", "sqlState" : "38000", "messageParameters" : { "funcName" : "`range`" diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/postgreSQL/interval.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/postgreSQL/interval.sql.out index 1add0830d9b7..fb837793ccd5 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/postgreSQL/interval.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/postgreSQL/interval.sql.out @@ -88,7 +88,7 @@ SELECT interval '1 2:03' day to hour -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_INTERVAL_FORMAT.UNMATCHED_FORMAT_STRING_WITH_NOTICE", + "condition" : "INVALID_INTERVAL_FORMAT.UNMATCHED_FORMAT_STRING_WITH_NOTICE", "sqlState" : "22006", "messageParameters" : { "input" : "1 2:03", @@ -111,7 +111,7 @@ SELECT interval '1 2:03:04' day to hour -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_INTERVAL_FORMAT.UNMATCHED_FORMAT_STRING_WITH_NOTICE", + "condition" : "INVALID_INTERVAL_FORMAT.UNMATCHED_FORMAT_STRING_WITH_NOTICE", "sqlState" : "22006", "messageParameters" : { "input" : "1 2:03:04", @@ -141,7 +141,7 @@ SELECT interval '1 2:03:04' day to minute -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_INTERVAL_FORMAT.UNMATCHED_FORMAT_STRING_WITH_NOTICE", + "condition" : "INVALID_INTERVAL_FORMAT.UNMATCHED_FORMAT_STRING_WITH_NOTICE", "sqlState" : "22006", "messageParameters" : { "input" : "1 2:03:04", @@ -164,7 +164,7 @@ SELECT interval '1 2:03' day to second -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_INTERVAL_FORMAT.UNMATCHED_FORMAT_STRING_WITH_NOTICE", + "condition" : "INVALID_INTERVAL_FORMAT.UNMATCHED_FORMAT_STRING_WITH_NOTICE", "sqlState" : "22006", "messageParameters" : { "input" : "1 2:03", @@ -194,7 +194,7 @@ SELECT interval '1 2:03' hour to minute -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_INTERVAL_FORMAT.UNMATCHED_FORMAT_STRING_WITH_NOTICE", + "condition" : "INVALID_INTERVAL_FORMAT.UNMATCHED_FORMAT_STRING_WITH_NOTICE", "sqlState" : "22006", "messageParameters" : { "input" : "1 2:03", @@ -217,7 +217,7 @@ SELECT interval '1 2:03:04' hour to minute -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_INTERVAL_FORMAT.UNMATCHED_FORMAT_STRING_WITH_NOTICE", + "condition" : "INVALID_INTERVAL_FORMAT.UNMATCHED_FORMAT_STRING_WITH_NOTICE", "sqlState" : "22006", "messageParameters" : { "input" : "1 2:03:04", @@ -240,7 +240,7 @@ SELECT interval '1 2:03' hour to second -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_INTERVAL_FORMAT.UNMATCHED_FORMAT_STRING_WITH_NOTICE", + "condition" : "INVALID_INTERVAL_FORMAT.UNMATCHED_FORMAT_STRING_WITH_NOTICE", "sqlState" : "22006", "messageParameters" : { "input" : "1 2:03", @@ -263,7 +263,7 @@ SELECT interval '1 2:03:04' hour to second -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_INTERVAL_FORMAT.UNMATCHED_FORMAT_STRING_WITH_NOTICE", + "condition" : "INVALID_INTERVAL_FORMAT.UNMATCHED_FORMAT_STRING_WITH_NOTICE", "sqlState" : "22006", "messageParameters" : { "input" : "1 2:03:04", @@ -286,7 +286,7 @@ SELECT interval '1 2:03' minute to second -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_INTERVAL_FORMAT.UNMATCHED_FORMAT_STRING_WITH_NOTICE", + "condition" : "INVALID_INTERVAL_FORMAT.UNMATCHED_FORMAT_STRING_WITH_NOTICE", "sqlState" : "22006", "messageParameters" : { "input" : "1 2:03", @@ -309,7 +309,7 @@ SELECT interval '1 2:03:04' minute to second -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_INTERVAL_FORMAT.UNMATCHED_FORMAT_STRING_WITH_NOTICE", + "condition" : "INVALID_INTERVAL_FORMAT.UNMATCHED_FORMAT_STRING_WITH_NOTICE", "sqlState" : "22006", "messageParameters" : { "input" : "1 2:03:04", diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/postgreSQL/join.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/postgreSQL/join.sql.out index 144fb69812ee..6f862b36f9de 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/postgreSQL/join.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/postgreSQL/join.sql.out @@ -348,7 +348,7 @@ SELECT '' AS `xxx`, i, k, t -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "AMBIGUOUS_REFERENCE", + "condition" : "AMBIGUOUS_REFERENCE", "sqlState" : "42704", "messageParameters" : { "name" : "`i`", @@ -3306,7 +3306,7 @@ select * from -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "AMBIGUOUS_REFERENCE", + "condition" : "AMBIGUOUS_REFERENCE", "sqlState" : "42704", "messageParameters" : { "name" : "`f1`", @@ -3328,7 +3328,7 @@ select * from -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`y`.`f1`", @@ -3385,7 +3385,7 @@ select t1.uunique1 from -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`t1`.`uunique1`", @@ -3407,7 +3407,7 @@ select t2.uunique1 from -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`t2`.`uunique1`", @@ -3429,7 +3429,7 @@ select uunique1 from -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`uunique1`", @@ -3753,7 +3753,7 @@ select f1,g from int4_tbl a, (select f1 as g) ss -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITHOUT_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITHOUT_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`f1`" @@ -3773,7 +3773,7 @@ select f1,g from int4_tbl a, (select a.f1 as g) ss -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITHOUT_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITHOUT_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`a`.`f1`" @@ -3793,7 +3793,7 @@ select f1,g from int4_tbl a cross join (select f1 as g) ss -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITHOUT_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITHOUT_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`f1`" @@ -3813,7 +3813,7 @@ select f1,g from int4_tbl a cross join (select a.f1 as g) ss -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITHOUT_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITHOUT_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`a`.`f1`" diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/postgreSQL/limit.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/postgreSQL/limit.sql.out index 2a5062bd65b7..b9a96f8cce40 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/postgreSQL/limit.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/postgreSQL/limit.sql.out @@ -141,7 +141,7 @@ select * from int8_tbl limit (case when random() < 0.5 then bigint(null) end) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INVALID_LIMIT_LIKE_EXPRESSION.IS_UNFOLDABLE", + "condition" : "INVALID_LIMIT_LIKE_EXPRESSION.IS_UNFOLDABLE", "sqlState" : "42K0E", "messageParameters" : { "expr" : "\"CASE WHEN (_nondeterministic < 0.5) THEN NULL END\"", @@ -162,7 +162,7 @@ select * from int8_tbl offset (case when random() < 0.5 then bigint(null) end) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INVALID_LIMIT_LIKE_EXPRESSION.IS_UNFOLDABLE", + "condition" : "INVALID_LIMIT_LIKE_EXPRESSION.IS_UNFOLDABLE", "sqlState" : "42K0E", "messageParameters" : { "expr" : "\"CASE WHEN (_nondeterministic < 0.5) THEN NULL END\"", diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/postgreSQL/numeric.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/postgreSQL/numeric.sql.out index 6c2ae2329175..f525113aa926 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/postgreSQL/numeric.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/postgreSQL/numeric.sql.out @@ -3560,7 +3560,7 @@ INSERT INTO num_exp_power_10_ln VALUES (7,1716699575118597095.423308199106402476 -- !query analysis org.apache.spark.SparkArithmeticException { - "errorClass" : "DECIMAL_PRECISION_EXCEEDS_MAX_PRECISION", + "condition" : "DECIMAL_PRECISION_EXCEEDS_MAX_PRECISION", "sqlState" : "22003", "messageParameters" : { "maxPrecision" : "38", @@ -3841,7 +3841,7 @@ INSERT INTO num_result SELECT t1.id, t2.id, t1.val, t2.val, t1.val * t2.val -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "INSERT_COLUMN_ARITY_MISMATCH.TOO_MANY_DATA_COLUMNS", + "condition" : "INSERT_COLUMN_ARITY_MISMATCH.TOO_MANY_DATA_COLUMNS", "sqlState" : "21S01", "messageParameters" : { "dataColumns" : "`id`, `id`, `val`, `val`, `(val * val)`", @@ -4521,7 +4521,7 @@ SELECT '' AS to_number_2, to_number('-34,338,492.654,878', '99G999G999D999G999' -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INVALID_FORMAT.THOUSANDS_SEPS_MUST_BEFORE_DEC", + "condition" : "INVALID_FORMAT.THOUSANDS_SEPS_MUST_BEFORE_DEC", "sqlState" : "42601", "messageParameters" : { "format" : "'99G999G999D999G999'" @@ -4576,7 +4576,7 @@ SELECT '' AS to_number_15, to_number('123,000','999G') -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INVALID_FORMAT.CONT_THOUSANDS_SEPS", + "condition" : "INVALID_FORMAT.CONT_THOUSANDS_SEPS", "sqlState" : "42601", "messageParameters" : { "format" : "'999G'" diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/postgreSQL/select_having.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/postgreSQL/select_having.sql.out index 215e93abb5a5..948558affa4d 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/postgreSQL/select_having.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/postgreSQL/select_having.sql.out @@ -144,7 +144,7 @@ SELECT a FROM test_having HAVING min(a) < max(a) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "MISSING_GROUP_BY", + "condition" : "MISSING_GROUP_BY", "sqlState" : "42803", "queryContext" : [ { "objectType" : "", @@ -161,7 +161,7 @@ SELECT 1 AS one FROM test_having HAVING a > 1 -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`a`", diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/postgreSQL/select_implicit.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/postgreSQL/select_implicit.sql.out index d2e8345fc942..2d0d9fabe594 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/postgreSQL/select_implicit.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/postgreSQL/select_implicit.sql.out @@ -109,7 +109,7 @@ SELECT count(*) FROM test_missing_target GROUP BY a ORDER BY b -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`b`", @@ -188,7 +188,7 @@ SELECT c, count(*) FROM test_missing_target GROUP BY 3 -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "GROUP_BY_POS_OUT_OF_RANGE", + "condition" : "GROUP_BY_POS_OUT_OF_RANGE", "sqlState" : "42805", "messageParameters" : { "index" : "3", @@ -211,7 +211,7 @@ SELECT count(*) FROM test_missing_target x, test_missing_target y -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "AMBIGUOUS_REFERENCE", + "condition" : "AMBIGUOUS_REFERENCE", "sqlState" : "42704", "messageParameters" : { "name" : "`b`", @@ -321,7 +321,7 @@ SELECT count(a) FROM test_missing_target GROUP BY a ORDER BY b -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`b`", @@ -385,7 +385,7 @@ SELECT count(x.a) FROM test_missing_target x, test_missing_target y -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "AMBIGUOUS_REFERENCE", + "condition" : "AMBIGUOUS_REFERENCE", "sqlState" : "42704", "messageParameters" : { "name" : "`b`", @@ -425,7 +425,7 @@ SELECT count(b) FROM test_missing_target x, test_missing_target y -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "AMBIGUOUS_REFERENCE", + "condition" : "AMBIGUOUS_REFERENCE", "sqlState" : "42704", "messageParameters" : { "name" : "`b`", diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/postgreSQL/text.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/postgreSQL/text.sql.out index 474c2401f40d..217f452ed2b1 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/postgreSQL/text.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/postgreSQL/text.sql.out @@ -232,7 +232,7 @@ select format_string('%0$s', 'Hello') -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_PARAMETER_VALUE.ZERO_INDEX", + "condition" : "INVALID_PARAMETER_VALUE.ZERO_INDEX", "sqlState" : "22023", "messageParameters" : { "functionName" : "`format_string`", diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/postgreSQL/union.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/postgreSQL/union.sql.out index 05002a7c4538..73697cde834f 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/postgreSQL/union.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/postgreSQL/union.sql.out @@ -998,7 +998,7 @@ SELECT q1 FROM int8_tbl EXCEPT SELECT q2 FROM int8_tbl ORDER BY q2 LIMIT 1 -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`q2`", diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/postgreSQL/window_part2.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/postgreSQL/window_part2.sql.out index 330e1c1cad7e..f9c1af1c5d32 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/postgreSQL/window_part2.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/postgreSQL/window_part2.sql.out @@ -446,7 +446,7 @@ window w as (order by f_numeric range between -- !query analysis org.apache.spark.SparkNumberFormatException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'NaN'", diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/postgreSQL/window_part3.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/postgreSQL/window_part3.sql.out index 2b2d69db348d..db223603c8fd 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/postgreSQL/window_part3.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/postgreSQL/window_part3.sql.out @@ -65,7 +65,7 @@ insert into datetimes values -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_INLINE_TABLE.FAILED_SQL_EXPRESSION_EVALUATION", + "condition" : "INVALID_INLINE_TABLE.FAILED_SQL_EXPRESSION_EVALUATION", "sqlState" : "42000", "messageParameters" : { "sqlExpr" : "\"CAST(11:00 BST AS TIMESTAMP)\"" @@ -234,7 +234,7 @@ from t1 where f1 = f2 -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.RANGE_FRAME_WITHOUT_ORDER", + "condition" : "DATATYPE_MISMATCH.RANGE_FRAME_WITHOUT_ORDER", "sqlState" : "42K09", "messageParameters" : { "sqlExpr" : "\"(PARTITION BY f1 RANGE BETWEEN 1 PRECEDING AND 1 FOLLOWING)\"" @@ -306,7 +306,7 @@ SELECT * FROM empsalary WHERE row_number() OVER (ORDER BY salary) < 10 -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "_LEGACY_ERROR_TEMP_1034", + "condition" : "_LEGACY_ERROR_TEMP_1034", "messageParameters" : { "clauseName" : "WHERE" } @@ -318,7 +318,7 @@ SELECT * FROM empsalary INNER JOIN tenk1 ON row_number() OVER (ORDER BY salary) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNSUPPORTED_EXPR_FOR_OPERATOR", + "condition" : "UNSUPPORTED_EXPR_FOR_OPERATOR", "sqlState" : "42K0E", "messageParameters" : { "invalidExprSqls" : "\"row_number() OVER (ORDER BY salary ASC NULLS FIRST ROWS BETWEEN UNBOUNDED PRECEDING AND CURRENT ROW)\"" @@ -338,7 +338,7 @@ SELECT rank() OVER (ORDER BY 1), count(*) FROM empsalary GROUP BY 1 -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNSUPPORTED_EXPR_FOR_OPERATOR", + "condition" : "UNSUPPORTED_EXPR_FOR_OPERATOR", "sqlState" : "42K0E", "messageParameters" : { "invalidExprSqls" : "\"RANK() OVER (ORDER BY 1 ASC NULLS FIRST ROWS BETWEEN UNBOUNDED PRECEDING AND CURRENT ROW)\"" @@ -358,7 +358,7 @@ SELECT * FROM rank() OVER (ORDER BY random()) -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'BY'", @@ -372,7 +372,7 @@ SELECT * FROM empsalary WHERE (rank() OVER (ORDER BY random())) > 10 -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "_LEGACY_ERROR_TEMP_1034", + "condition" : "_LEGACY_ERROR_TEMP_1034", "messageParameters" : { "clauseName" : "WHERE" } @@ -384,7 +384,7 @@ SELECT * FROM empsalary WHERE rank() OVER (ORDER BY random()) -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "_LEGACY_ERROR_TEMP_1034", + "condition" : "_LEGACY_ERROR_TEMP_1034", "messageParameters" : { "clauseName" : "WHERE" } @@ -396,7 +396,7 @@ select rank() OVER (PARTITION BY four, ORDER BY ten) FROM tenk1 -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'BY'", @@ -410,7 +410,7 @@ SELECT range(1, 100) OVER () FROM empsalary -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "UNRESOLVED_ROUTINE", + "condition" : "UNRESOLVED_ROUTINE", "sqlState" : "42883", "messageParameters" : { "routineName" : "`range`", @@ -431,7 +431,7 @@ SELECT ntile(0) OVER (ORDER BY ten), ten, four FROM tenk1 -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.VALUE_OUT_OF_RANGE", + "condition" : "DATATYPE_MISMATCH.VALUE_OUT_OF_RANGE", "sqlState" : "42K09", "messageParameters" : { "currentValue" : "0", @@ -454,7 +454,7 @@ SELECT nth_value(four, 0) OVER (ORDER BY ten), ten, four FROM tenk1 -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.VALUE_OUT_OF_RANGE", + "condition" : "DATATYPE_MISMATCH.VALUE_OUT_OF_RANGE", "sqlState" : "42K09", "messageParameters" : { "currentValue" : "0L", diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/postgreSQL/window_part4.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/postgreSQL/window_part4.sql.out index f042116182f7..f37830b7322d 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/postgreSQL/window_part4.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/postgreSQL/window_part4.sql.out @@ -500,7 +500,7 @@ FROM (VALUES(1,1),(2,2),(3,(cast('nan' as int))),(4,3),(5,4)) t(a,b) -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_INLINE_TABLE.FAILED_SQL_EXPRESSION_EVALUATION", + "condition" : "INVALID_INLINE_TABLE.FAILED_SQL_EXPRESSION_EVALUATION", "sqlState" : "42000", "messageParameters" : { "sqlExpr" : "\"CAST(nan AS INT)\"" diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/postgreSQL/with.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/postgreSQL/with.sql.out index 1219562068dd..4a220f59ac52 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/postgreSQL/with.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/postgreSQL/with.sql.out @@ -242,7 +242,7 @@ SELECT * FROM outermost ORDER BY 1 -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "TABLE_OR_VIEW_NOT_FOUND", + "condition" : "TABLE_OR_VIEW_NOT_FOUND", "sqlState" : "42P01", "messageParameters" : { "relationName" : "`outermost`" @@ -373,7 +373,7 @@ create table foo (with baz) -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "UNSUPPORTED_DATATYPE", + "condition" : "UNSUPPORTED_DATATYPE", "sqlState" : "0A000", "messageParameters" : { "typeName" : "\"BAZ\"" @@ -393,7 +393,7 @@ create table foo (with ordinality) -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "UNSUPPORTED_DATATYPE", + "condition" : "UNSUPPORTED_DATATYPE", "sqlState" : "0A000", "messageParameters" : { "typeName" : "\"ORDINALITY\"" @@ -426,7 +426,7 @@ WITH test AS (SELECT 42) INSERT INTO test VALUES (1) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "TABLE_OR_VIEW_NOT_FOUND", + "condition" : "TABLE_OR_VIEW_NOT_FOUND", "sqlState" : "42P01", "messageParameters" : { "relationName" : "`test`" diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/query_regex_column.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/query_regex_column.sql.out index d62e4b792136..d6282fb2eab9 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/query_regex_column.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/query_regex_column.sql.out @@ -36,7 +36,7 @@ SELECT `(a)?+.+` FROM testData2 WHERE a = 1 -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`(a)?+.+`", @@ -57,7 +57,7 @@ SELECT t.`(a)?+.+` FROM testData2 t WHERE a = 1 -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`t`.`(a)?+.+`", @@ -78,7 +78,7 @@ SELECT `(a|b)` FROM testData2 WHERE a = 2 -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`(a|b)`", @@ -99,7 +99,7 @@ SELECT `(a|b)?+.+` FROM testData2 WHERE a = 2 -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`(a|b)?+.+`", @@ -120,7 +120,7 @@ SELECT SUM(`(a|b)?+.+`) FROM testData2 -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`(a|b)?+.+`", @@ -141,7 +141,7 @@ SELECT SUM(`(a)`) FROM testData2 -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`(a)`", @@ -458,7 +458,7 @@ SELECT SUM(a) FROM testdata3 GROUP BY `(a)` -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`(a)`", @@ -479,7 +479,7 @@ SELECT SUM(a) FROM testdata3 GROUP BY `(a)?+.+` -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`(a)?+.+`", diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/random.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/random.sql.out index 96a4b2ec91c7..c46708e5e866 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/random.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/random.sql.out @@ -28,7 +28,7 @@ SELECT rand(1.0) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"1.0\"", @@ -76,7 +76,7 @@ SELECT rand('1') -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"1\"", @@ -280,7 +280,7 @@ SELECT uniform(0, 1, cast(NULL as float)) AS result -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"CAST(NULL AS FLOAT)\"", @@ -304,7 +304,7 @@ SELECT uniform(10, 20, col) AS result FROM VALUES (0), (1), (2) tab(col) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.NON_FOLDABLE_INPUT", + "condition" : "DATATYPE_MISMATCH.NON_FOLDABLE_INPUT", "sqlState" : "42K09", "messageParameters" : { "inputExpr" : "\"col\"", @@ -327,7 +327,7 @@ SELECT uniform(col, 10, 0) AS result FROM VALUES (0), (1), (2) tab(col) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.NON_FOLDABLE_INPUT", + "condition" : "DATATYPE_MISMATCH.NON_FOLDABLE_INPUT", "sqlState" : "42K09", "messageParameters" : { "inputExpr" : "\"col\"", @@ -350,7 +350,7 @@ SELECT uniform(10) AS result -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", + "condition" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", "sqlState" : "42605", "messageParameters" : { "actualNum" : "1", @@ -373,7 +373,7 @@ SELECT uniform(10, 20, 30, 40) AS result -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", + "condition" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", "sqlState" : "42605", "messageParameters" : { "actualNum" : "4", @@ -396,7 +396,7 @@ SELECT uniform(10.0F, 20.0F, 0.0F) AS result -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"0.0\"", @@ -420,7 +420,7 @@ SELECT uniform(10.0F, 20.0F, 0.0D) AS result -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"0.0\"", @@ -444,7 +444,7 @@ SELECT uniform(cast(10 as decimal(10, 3)), cast(20 as decimal(10, 3)), cast(0 as -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"CAST(0 AS DECIMAL(10,3))\"", @@ -468,7 +468,7 @@ SELECT uniform('abc', 10, 0) AS result -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"abc\"", @@ -492,7 +492,7 @@ SELECT uniform(0, 'def', 0) AS result -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"def\"", @@ -516,7 +516,7 @@ SELECT uniform(0, 10, 'ghi') AS result -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"ghi\"", @@ -642,7 +642,7 @@ SELECT randstr(col, 0) AS result FROM VALUES (0), (1), (2) tab(col) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.NON_FOLDABLE_INPUT", + "condition" : "DATATYPE_MISMATCH.NON_FOLDABLE_INPUT", "sqlState" : "42K09", "messageParameters" : { "inputExpr" : "\"col\"", @@ -665,7 +665,7 @@ SELECT randstr(10, col) AS result FROM VALUES (0), (1), (2) tab(col) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.NON_FOLDABLE_INPUT", + "condition" : "DATATYPE_MISMATCH.NON_FOLDABLE_INPUT", "sqlState" : "42K09", "messageParameters" : { "inputExpr" : "\"col\"", @@ -688,7 +688,7 @@ SELECT randstr(10, 0, 1) AS result -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", + "condition" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", "sqlState" : "42605", "messageParameters" : { "actualNum" : "3", @@ -717,7 +717,7 @@ SELECT randstr(10, "a") AS result FROM VALUES (0) tab(a) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"a\"", @@ -741,7 +741,7 @@ SELECT randstr(10, 1.5) AS result FROM VALUES (0) tab(a) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"1.5\"", diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/regexp-functions.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/regexp-functions.sql.out index cd6b95e75f1d..0ac2e804370a 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/regexp-functions.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/regexp-functions.sql.out @@ -236,7 +236,7 @@ SELECT regexp_replace('healthy, wealthy, and wise', '\\w+thy', 'something', -2) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.VALUE_OUT_OF_RANGE", + "condition" : "DATATYPE_MISMATCH.VALUE_OUT_OF_RANGE", "sqlState" : "42K09", "messageParameters" : { "currentValue" : "-2", @@ -259,7 +259,7 @@ SELECT regexp_replace('healthy, wealthy, and wise', '\\w+thy', 'something', 0) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.VALUE_OUT_OF_RANGE", + "condition" : "DATATYPE_MISMATCH.VALUE_OUT_OF_RANGE", "sqlState" : "42K09", "messageParameters" : { "currentValue" : "0", diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/selectExcept.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/selectExcept.sql.out index e02562e29835..2e7ba1bcf8bb 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/selectExcept.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/selectExcept.sql.out @@ -29,7 +29,7 @@ CREATE TABLE ids (id INT) USING CSV -- !query analysis org.apache.spark.sql.catalyst.analysis.TableAlreadyExistsException { - "errorClass" : "TABLE_OR_VIEW_ALREADY_EXISTS", + "condition" : "TABLE_OR_VIEW_ALREADY_EXISTS", "sqlState" : "42P07", "messageParameters" : { "relationName" : "`spark_catalog`.`default`.`ids`" @@ -237,7 +237,7 @@ SELECT * EXCEPT name FROM tbl_view -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'name'", @@ -251,7 +251,7 @@ SELECT * EXCEPT() name FROM tbl_view -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "')'", @@ -265,7 +265,7 @@ SELECT * EXCEPT(invalid_column) FROM tbl_view -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`invalid_column`", @@ -286,7 +286,7 @@ SELECT * EXCEPT(id, invalid_column) FROM tbl_view -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`invalid_column`", @@ -307,7 +307,7 @@ SELECT * EXCEPT(id, id) FROM tbl_view -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "EXCEPT_OVERLAPPING_COLUMNS", + "condition" : "EXCEPT_OVERLAPPING_COLUMNS", "sqlState" : "42702", "messageParameters" : { "columns" : "id, id" @@ -327,7 +327,7 @@ SELECT * EXCEPT(data.s2, data.s2.f2) FROM tbl_view -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "EXCEPT_OVERLAPPING_COLUMNS", + "condition" : "EXCEPT_OVERLAPPING_COLUMNS", "sqlState" : "42702", "messageParameters" : { "columns" : "data.s2, data.s2.f2" diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/show-tables.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/show-tables.sql.out index 8ae12b928d72..fb985d6b6b0c 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/show-tables.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/show-tables.sql.out @@ -98,7 +98,7 @@ SHOW TABLE EXTENDED -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "end of input", @@ -118,7 +118,7 @@ SHOW TABLE EXTENDED PARTITION(c='Us', d=1) -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'PARTITION'", @@ -132,7 +132,7 @@ SHOW TABLE EXTENDED LIKE 'show_t*' PARTITION(c='Us', d=1) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "TABLE_OR_VIEW_NOT_FOUND", + "condition" : "TABLE_OR_VIEW_NOT_FOUND", "sqlState" : "42P01", "messageParameters" : { "relationName" : "`show_t*`" @@ -152,7 +152,7 @@ SHOW TABLE EXTENDED LIKE 'show_t1' PARTITION(c='Us') -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "_LEGACY_ERROR_TEMP_1232", + "condition" : "_LEGACY_ERROR_TEMP_1232", "messageParameters" : { "partitionColumnNames" : "c, d", "specKeys" : "c", @@ -166,7 +166,7 @@ SHOW TABLE EXTENDED LIKE 'show_t1' PARTITION(a='Us', d=1) -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "PARTITIONS_NOT_FOUND", + "condition" : "PARTITIONS_NOT_FOUND", "sqlState" : "428FT", "messageParameters" : { "partitionList" : "`a`", @@ -180,7 +180,7 @@ SHOW TABLE EXTENDED LIKE 'show_t1' PARTITION(c='Ch', d=1) -- !query analysis org.apache.spark.sql.catalyst.analysis.NoSuchPartitionException { - "errorClass" : "PARTITIONS_NOT_FOUND", + "condition" : "PARTITIONS_NOT_FOUND", "sqlState" : "428FT", "messageParameters" : { "partitionList" : "PARTITION (`c` = Ch, `d` = 1)", diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/show-views.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/show-views.sql.out index d092590b143b..e1057dafbebd 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/show-views.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/show-views.sql.out @@ -105,7 +105,7 @@ SHOW VIEWS IN wrongdb LIKE 'view_*' -- !query analysis org.apache.spark.sql.catalyst.analysis.NoSuchNamespaceException { - "errorClass" : "SCHEMA_NOT_FOUND", + "condition" : "SCHEMA_NOT_FOUND", "sqlState" : "42704", "messageParameters" : { "schemaName" : "`spark_catalog`.`wrongdb`" diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/show_columns.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/show_columns.sql.out index 76c3b88a3ce6..85b757cddf67 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/show_columns.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/show_columns.sql.out @@ -68,7 +68,7 @@ SHOW COLUMNS IN badtable FROM showdb -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "TABLE_OR_VIEW_NOT_FOUND", + "condition" : "TABLE_OR_VIEW_NOT_FOUND", "sqlState" : "42P01", "messageParameters" : { "relationName" : "`showdb`.`badtable`" @@ -94,7 +94,7 @@ SHOW COLUMNS IN showdb.showcolumn1 FROM baddb -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "SHOW_COLUMNS_WITH_CONFLICT_NAMESPACE", + "condition" : "SHOW_COLUMNS_WITH_CONFLICT_NAMESPACE", "sqlState" : "42K05", "messageParameters" : { "namespaceA" : "`baddb`", @@ -114,7 +114,7 @@ SHOW COLUMNS IN showdb.showcolumn3 -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "TABLE_OR_VIEW_NOT_FOUND", + "condition" : "TABLE_OR_VIEW_NOT_FOUND", "sqlState" : "42P01", "messageParameters" : { "relationName" : "`showdb`.`showcolumn3`" @@ -134,7 +134,7 @@ SHOW COLUMNS IN showcolumn3 FROM showdb -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "TABLE_OR_VIEW_NOT_FOUND", + "condition" : "TABLE_OR_VIEW_NOT_FOUND", "sqlState" : "42P01", "messageParameters" : { "relationName" : "`showdb`.`showcolumn3`" @@ -154,7 +154,7 @@ SHOW COLUMNS IN showcolumn4 -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "TABLE_OR_VIEW_NOT_FOUND", + "condition" : "TABLE_OR_VIEW_NOT_FOUND", "sqlState" : "42P01", "messageParameters" : { "relationName" : "`showcolumn4`" diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/sql-compatibility-functions.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/sql-compatibility-functions.sql.out index a18e4ede957c..723a4ccda708 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/sql-compatibility-functions.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/sql-compatibility-functions.sql.out @@ -81,7 +81,7 @@ SELECT string(1, 2) -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", + "condition" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", "sqlState" : "42605", "messageParameters" : { "actualNum" : "2", diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/sql-on-files.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/sql-on-files.sql.out index 329fbd49b92d..a70bdc9f0521 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/sql-on-files.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/sql-on-files.sql.out @@ -19,7 +19,7 @@ SELECT * FROM parquet.`` -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_EMPTY_LOCATION", + "condition" : "INVALID_EMPTY_LOCATION", "sqlState" : "42K05", "messageParameters" : { "location" : "" @@ -39,7 +39,7 @@ SELECT * FROM parquet.`/file/not/found` -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "PATH_NOT_FOUND", + "condition" : "PATH_NOT_FOUND", "sqlState" : "42K03", "messageParameters" : { "path" : "file:/file/not/found" @@ -74,7 +74,7 @@ SELECT * FROM orc.`` -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_EMPTY_LOCATION", + "condition" : "INVALID_EMPTY_LOCATION", "sqlState" : "42K05", "messageParameters" : { "location" : "" @@ -94,7 +94,7 @@ SELECT * FROM orc.`/file/not/found` -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "PATH_NOT_FOUND", + "condition" : "PATH_NOT_FOUND", "sqlState" : "42K03", "messageParameters" : { "path" : "file:/file/not/found" @@ -129,7 +129,7 @@ SELECT * FROM csv.`` -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_EMPTY_LOCATION", + "condition" : "INVALID_EMPTY_LOCATION", "sqlState" : "42K05", "messageParameters" : { "location" : "" @@ -149,7 +149,7 @@ SELECT * FROM csv.`/file/not/found` -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "PATH_NOT_FOUND", + "condition" : "PATH_NOT_FOUND", "sqlState" : "42K03", "messageParameters" : { "path" : "file:/file/not/found" @@ -184,7 +184,7 @@ SELECT * FROM json.`` -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_EMPTY_LOCATION", + "condition" : "INVALID_EMPTY_LOCATION", "sqlState" : "42K05", "messageParameters" : { "location" : "" @@ -204,7 +204,7 @@ SELECT * FROM json.`/file/not/found` -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "PATH_NOT_FOUND", + "condition" : "PATH_NOT_FOUND", "sqlState" : "42K03", "messageParameters" : { "path" : "file:/file/not/found" @@ -238,7 +238,7 @@ SELECT * FROM json.`https://raw.githubusercontent.com/apache/spark/refs/heads/ma -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "FAILED_READ_FILE.UNSUPPORTED_FILE_SYSTEM", + "condition" : "FAILED_READ_FILE.UNSUPPORTED_FILE_SYSTEM", "sqlState" : "KD001", "messageParameters" : { "fileSystemClass" : "org.apache.hadoop.fs.http.HttpsFileSystem", diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/sql-session-variables.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/sql-session-variables.sql.out index 77e7174167ec..a18a88982150 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/sql-session-variables.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/sql-session-variables.sql.out @@ -125,7 +125,7 @@ DECLARE VARIABLE IF NOT EXISTS var1 INT -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'EXISTS'", @@ -175,7 +175,7 @@ SELECT var1 -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITHOUT_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITHOUT_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`var1`" @@ -195,7 +195,7 @@ DROP TEMPORARY VARIABLE var1 -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "VARIABLE_NOT_FOUND", + "condition" : "VARIABLE_NOT_FOUND", "sqlState" : "42883", "messageParameters" : { "variableName" : "`system`.`session`.`var1`" @@ -222,7 +222,7 @@ DROP VARIABLE var1 -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'VARIABLE'", @@ -236,7 +236,7 @@ DROP VARIABLE system.session.var1 -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'VARIABLE'", @@ -395,7 +395,7 @@ DROP TEMPORARY VARIABLE var1 -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "VARIABLE_NOT_FOUND", + "condition" : "VARIABLE_NOT_FOUND", "sqlState" : "42883", "messageParameters" : { "variableName" : "`system`.`session`.`var1`" @@ -422,7 +422,7 @@ DROP TEMPORARY VARIABLE var1 -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "VARIABLE_NOT_FOUND", + "condition" : "VARIABLE_NOT_FOUND", "sqlState" : "42883", "messageParameters" : { "variableName" : "`system`.`session`.`var1`" @@ -449,7 +449,7 @@ DROP TEMPORARY VARIABLE var1 -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "VARIABLE_NOT_FOUND", + "condition" : "VARIABLE_NOT_FOUND", "sqlState" : "42883", "messageParameters" : { "variableName" : "`system`.`session`.`var1`" @@ -476,7 +476,7 @@ DROP TEMPORARY VARIABLE var1 -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "VARIABLE_NOT_FOUND", + "condition" : "VARIABLE_NOT_FOUND", "sqlState" : "42883", "messageParameters" : { "variableName" : "`system`.`session`.`var1`" @@ -516,7 +516,7 @@ DECLARE OR REPLACE VARIABLE builtin.var1 INT -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "UNRESOLVED_VARIABLE", + "condition" : "UNRESOLVED_VARIABLE", "sqlState" : "42883", "messageParameters" : { "searchPath" : "`system`.`session`", @@ -530,7 +530,7 @@ DECLARE OR REPLACE VARIABLE system.sesion.var1 INT -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "UNRESOLVED_VARIABLE", + "condition" : "UNRESOLVED_VARIABLE", "sqlState" : "42883", "messageParameters" : { "searchPath" : "`system`.`session`", @@ -544,7 +544,7 @@ DECLARE OR REPLACE VARIABLE sys.session.var1 INT -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "UNRESOLVED_VARIABLE", + "condition" : "UNRESOLVED_VARIABLE", "sqlState" : "42883", "messageParameters" : { "searchPath" : "`system`.`session`", @@ -565,7 +565,7 @@ SELECT var -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITHOUT_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITHOUT_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`var`" @@ -585,7 +585,7 @@ SELECT ses.var1 -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITHOUT_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITHOUT_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`ses`.`var1`" @@ -605,7 +605,7 @@ SELECT b.sesson.var1 -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITHOUT_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITHOUT_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`b`.`sesson`.`var1`" @@ -625,7 +625,7 @@ SELECT builtn.session.var1 -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITHOUT_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITHOUT_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`builtn`.`session`.`var1`" @@ -645,7 +645,7 @@ SET VARIABLE ses.var1 = 1 -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "UNRESOLVED_VARIABLE", + "condition" : "UNRESOLVED_VARIABLE", "sqlState" : "42883", "messageParameters" : { "searchPath" : "`SYSTEM`.`SESSION`", @@ -659,7 +659,7 @@ SET VARIABLE builtn.session.var1 = 1 -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "UNRESOLVED_VARIABLE", + "condition" : "UNRESOLVED_VARIABLE", "sqlState" : "42883", "messageParameters" : { "searchPath" : "`SYSTEM`.`SESSION`", @@ -799,7 +799,7 @@ DECLARE var1 -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_SQL_SYNTAX.VARIABLE_TYPE_OR_DEFAULT_REQUIRED", + "condition" : "INVALID_SQL_SYNTAX.VARIABLE_TYPE_OR_DEFAULT_REQUIRED", "sqlState" : "42000", "queryContext" : [ { "objectType" : "", @@ -824,7 +824,7 @@ DECLARE OR REPLACE VARIABLE var1 INT DEFAULT (SELECT c1 FROM VALUES(1) AS T(c1)) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INVALID_DEFAULT_VALUE.SUBQUERY_EXPRESSION", + "condition" : "INVALID_DEFAULT_VALUE.SUBQUERY_EXPRESSION", "sqlState" : "42623", "messageParameters" : { "colName" : "`system`.`session`.`var1`", @@ -839,7 +839,7 @@ DECLARE OR REPLACE VARIABLE var1 INT DEFAULT 'hello' -- !query analysis org.apache.spark.SparkNumberFormatException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'hello'", @@ -861,7 +861,7 @@ DECLARE OR REPLACE VARIABLE var1 INT DEFAULT 1 / 0 -- !query analysis org.apache.spark.SparkArithmeticException { - "errorClass" : "DIVIDE_BY_ZERO", + "condition" : "DIVIDE_BY_ZERO", "sqlState" : "22012", "messageParameters" : { "config" : "\"spark.sql.ansi.enabled\"" @@ -881,7 +881,7 @@ DECLARE OR REPLACE VARIABLE var1 SMALLINT DEFAULT 100000 -- !query analysis org.apache.spark.SparkArithmeticException { - "errorClass" : "CAST_OVERFLOW", + "condition" : "CAST_OVERFLOW", "sqlState" : "22003", "messageParameters" : { "sourceType" : "\"INT\"", @@ -997,7 +997,7 @@ SET VARIABLE var1 = (SELECT c1 FROM VALUES(1.0E10) AS T(c1)) -- !query analysis org.apache.spark.SparkArithmeticException { - "errorClass" : "CAST_OVERFLOW", + "condition" : "CAST_OVERFLOW", "sqlState" : "22003", "messageParameters" : { "sourceType" : "\"DOUBLE\"", @@ -1019,7 +1019,7 @@ SET VARIABLE var1 = (SELECT c1 FROM VALUES(1), (2) AS T(c1)) -- !query analysis org.apache.spark.SparkException { - "errorClass" : "SCALAR_SUBQUERY_TOO_MANY_ROWS", + "condition" : "SCALAR_SUBQUERY_TOO_MANY_ROWS", "sqlState" : "21000", "queryContext" : [ { "objectType" : "", @@ -1036,7 +1036,7 @@ SET VARIABLE var1 = (SELECT c1, c1 FROM VALUES(1), (2) AS T(c1)) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INVALID_SUBQUERY_EXPRESSION.SCALAR_SUBQUERY_RETURN_MORE_THAN_ONE_OUTPUT_COLUMN", + "condition" : "INVALID_SUBQUERY_EXPRESSION.SCALAR_SUBQUERY_RETURN_MORE_THAN_ONE_OUTPUT_COLUMN", "sqlState" : "42823", "messageParameters" : { "number" : "2" @@ -1056,7 +1056,7 @@ SET VARIABLE var1 = (SELECT c1 FROM VALUES('hello') AS T(c1)) -- !query analysis org.apache.spark.SparkNumberFormatException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'hello'", @@ -1213,7 +1213,7 @@ SET VARIABLE var1 = 1, var2 = 0, vAr1 = 1 -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "DUPLICATE_ASSIGNMENTS", + "condition" : "DUPLICATE_ASSIGNMENTS", "sqlState" : "42701", "messageParameters" : { "nameList" : "`var1`" @@ -1360,7 +1360,7 @@ SET VARIABLE () = (SELECT 1) -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_SET_SYNTAX", + "condition" : "INVALID_SET_SYNTAX", "sqlState" : "42000", "queryContext" : [ { "objectType" : "", @@ -1377,7 +1377,7 @@ SET VARIABLE (var1, var2, var3) = (SELECT c1, c2, c3 FROM VALUES(100, 110, 120), -- !query analysis org.apache.spark.SparkException { - "errorClass" : "ROW_SUBQUERY_TOO_MANY_ROWS", + "condition" : "ROW_SUBQUERY_TOO_MANY_ROWS", "sqlState" : "21000" } @@ -1387,7 +1387,7 @@ SET VARIABLE (var1, var2, var3) = (SELECT c1, c2 FROM VALUES(100, 110, 120) AS T -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "ASSIGNMENT_ARITY_MISMATCH", + "condition" : "ASSIGNMENT_ARITY_MISMATCH", "sqlState" : "42802", "messageParameters" : { "numExpr" : "2", @@ -1401,7 +1401,7 @@ SET VARIABLE (var1, var2, var3) = (SELECT c1, c2, c3, c1 FROM VALUES(100, 110, 1 -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "ASSIGNMENT_ARITY_MISMATCH", + "condition" : "ASSIGNMENT_ARITY_MISMATCH", "sqlState" : "42802", "messageParameters" : { "numExpr" : "4", @@ -1415,7 +1415,7 @@ SET VARIABLE (var1, var2, var1) = (SELECT c1, c2, c3, c1 FROM VALUES(100, 110, 1 -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "DUPLICATE_ASSIGNMENTS", + "condition" : "DUPLICATE_ASSIGNMENTS", "sqlState" : "42701", "messageParameters" : { "nameList" : "`var1`" @@ -1649,7 +1649,7 @@ SET VARIABLE (var1) = (SELECT DEFAULT FROM VALUES(1),(2),(3) AS T(c1) OFFSET 1) -- !query analysis org.apache.spark.SparkException { - "errorClass" : "ROW_SUBQUERY_TOO_MANY_ROWS", + "condition" : "ROW_SUBQUERY_TOO_MANY_ROWS", "sqlState" : "21000" } @@ -1674,7 +1674,7 @@ SET VARIABLE (var1) = (WITH v1(c1) AS (VALUES(1) AS T(c1)) SELECT DEFAULT FROM V -- !query analysis org.apache.spark.SparkException { - "errorClass" : "ROW_SUBQUERY_TOO_MANY_ROWS", + "condition" : "ROW_SUBQUERY_TOO_MANY_ROWS", "sqlState" : "21000" } @@ -1691,7 +1691,7 @@ SET VARIABLE var1 = 'Hello' || DEFAULT -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "DEFAULT_PLACEMENT_INVALID", + "condition" : "DEFAULT_PLACEMENT_INVALID", "sqlState" : "42608" } @@ -1708,7 +1708,7 @@ SET VARIABLE (var1) = (WITH v1(c1) AS (VALUES(1) AS T(c1)) SELECT DEFAULT + 1 FR -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "DEFAULT_PLACEMENT_INVALID", + "condition" : "DEFAULT_PLACEMENT_INVALID", "sqlState" : "42608" } @@ -1718,7 +1718,7 @@ SET VARIABLE var1 = session.default -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITHOUT_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITHOUT_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`session`.`default`" @@ -1798,7 +1798,7 @@ SET var1 = 5 -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "UNSUPPORTED_FEATURE.SET_VARIABLE_USING_SET", + "condition" : "UNSUPPORTED_FEATURE.SET_VARIABLE_USING_SET", "sqlState" : "0A000", "messageParameters" : { "variableName" : "`var1`" @@ -1811,7 +1811,7 @@ SET session.var1 = 5 -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "UNSUPPORTED_FEATURE.SET_VARIABLE_USING_SET", + "condition" : "UNSUPPORTED_FEATURE.SET_VARIABLE_USING_SET", "sqlState" : "0A000", "messageParameters" : { "variableName" : "`session`.`var1`" @@ -1824,7 +1824,7 @@ SET system.session.var1 = 5 -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "UNSUPPORTED_FEATURE.SET_VARIABLE_USING_SET", + "condition" : "UNSUPPORTED_FEATURE.SET_VARIABLE_USING_SET", "sqlState" : "0A000", "messageParameters" : { "variableName" : "`system`.`session`.`var1`" @@ -1837,7 +1837,7 @@ SET vAr1 = 5 -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "UNSUPPORTED_FEATURE.SET_VARIABLE_USING_SET", + "condition" : "UNSUPPORTED_FEATURE.SET_VARIABLE_USING_SET", "sqlState" : "0A000", "messageParameters" : { "variableName" : "`vAr1`" @@ -1850,7 +1850,7 @@ SET seSSion.var1 = 5 -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "UNSUPPORTED_FEATURE.SET_VARIABLE_USING_SET", + "condition" : "UNSUPPORTED_FEATURE.SET_VARIABLE_USING_SET", "sqlState" : "0A000", "messageParameters" : { "variableName" : "`seSSion`.`var1`" @@ -1863,7 +1863,7 @@ SET sYStem.session.var1 = 5 -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "UNSUPPORTED_FEATURE.SET_VARIABLE_USING_SET", + "condition" : "UNSUPPORTED_FEATURE.SET_VARIABLE_USING_SET", "sqlState" : "0A000", "messageParameters" : { "variableName" : "`sYStem`.`session`.`var1`" @@ -2015,7 +2015,7 @@ SELECT sum(c1) FILTER (c1 != var1) FROM VALUES(1, 2), (2, 3) AS T(c1, c2) -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'('", @@ -2105,7 +2105,7 @@ CREATE OR REPLACE VIEW v AS SELECT var1 AS c1 -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_TEMP_OBJ_REFERENCE", + "condition" : "INVALID_TEMP_OBJ_REFERENCE", "sqlState" : "42K0F", "messageParameters" : { "obj" : "VIEW", diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/sql-udf.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/sql-udf.sql.out index b3c10e929f29..3316642de1f8 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/sql-udf.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/sql-udf.sql.out @@ -4,7 +4,7 @@ CREATE FUNCTION foo1a0() RETURNS INT RETURN 1 -- !query analysis org.apache.spark.sql.catalyst.analysis.FunctionAlreadyExistsException { - "errorClass" : "ROUTINE_ALREADY_EXISTS", + "condition" : "ROUTINE_ALREADY_EXISTS", "sqlState" : "42723", "messageParameters" : { "existingRoutineType" : "routine", @@ -27,7 +27,7 @@ SELECT foo1a0(1) -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", + "condition" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", "sqlState" : "42605", "messageParameters" : { "actualNum" : "1", @@ -50,7 +50,7 @@ CREATE FUNCTION foo1a1(a INT) RETURNS INT RETURN 1 -- !query analysis org.apache.spark.sql.catalyst.analysis.FunctionAlreadyExistsException { - "errorClass" : "ROUTINE_ALREADY_EXISTS", + "condition" : "ROUTINE_ALREADY_EXISTS", "sqlState" : "42723", "messageParameters" : { "existingRoutineType" : "routine", @@ -73,7 +73,7 @@ SELECT foo1a1(1, 2) -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", + "condition" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", "sqlState" : "42605", "messageParameters" : { "actualNum" : "2", @@ -96,7 +96,7 @@ CREATE FUNCTION foo1a2(a INT, b INT, c INT, d INT) RETURNS INT RETURN 1 -- !query analysis org.apache.spark.sql.catalyst.analysis.FunctionAlreadyExistsException { - "errorClass" : "ROUTINE_ALREADY_EXISTS", + "condition" : "ROUTINE_ALREADY_EXISTS", "sqlState" : "42723", "messageParameters" : { "existingRoutineType" : "routine", @@ -119,7 +119,7 @@ CREATE FUNCTION foo2_1a(a INT) RETURNS INT RETURN a -- !query analysis org.apache.spark.sql.catalyst.analysis.FunctionAlreadyExistsException { - "errorClass" : "ROUTINE_ALREADY_EXISTS", + "condition" : "ROUTINE_ALREADY_EXISTS", "sqlState" : "42723", "messageParameters" : { "existingRoutineType" : "routine", @@ -142,7 +142,7 @@ CREATE FUNCTION foo2_1b(a INT, b INT) RETURNS INT RETURN a + b -- !query analysis org.apache.spark.sql.catalyst.analysis.FunctionAlreadyExistsException { - "errorClass" : "ROUTINE_ALREADY_EXISTS", + "condition" : "ROUTINE_ALREADY_EXISTS", "sqlState" : "42723", "messageParameters" : { "existingRoutineType" : "routine", @@ -165,7 +165,7 @@ CREATE FUNCTION foo2_1c(a INT, b INT) RETURNS INT RETURN 10 * (a + b) + 100 * (a -- !query analysis org.apache.spark.sql.catalyst.analysis.FunctionAlreadyExistsException { - "errorClass" : "ROUTINE_ALREADY_EXISTS", + "condition" : "ROUTINE_ALREADY_EXISTS", "sqlState" : "42723", "messageParameters" : { "existingRoutineType" : "routine", @@ -188,7 +188,7 @@ CREATE FUNCTION foo2_1d(a INT, b INT) RETURNS INT RETURN ABS(a) - LENGTH(CAST(b -- !query analysis org.apache.spark.sql.catalyst.analysis.FunctionAlreadyExistsException { - "errorClass" : "ROUTINE_ALREADY_EXISTS", + "condition" : "ROUTINE_ALREADY_EXISTS", "sqlState" : "42723", "messageParameters" : { "existingRoutineType" : "routine", @@ -211,7 +211,7 @@ CREATE FUNCTION foo2_2a(a INT) RETURNS INT RETURN SELECT a -- !query analysis org.apache.spark.sql.catalyst.analysis.FunctionAlreadyExistsException { - "errorClass" : "ROUTINE_ALREADY_EXISTS", + "condition" : "ROUTINE_ALREADY_EXISTS", "sqlState" : "42723", "messageParameters" : { "existingRoutineType" : "routine", @@ -234,7 +234,7 @@ CREATE FUNCTION foo2_2b(a INT) RETURNS INT RETURN 1 + (SELECT a) -- !query analysis org.apache.spark.sql.catalyst.analysis.FunctionAlreadyExistsException { - "errorClass" : "ROUTINE_ALREADY_EXISTS", + "condition" : "ROUTINE_ALREADY_EXISTS", "sqlState" : "42723", "messageParameters" : { "existingRoutineType" : "routine", @@ -259,7 +259,7 @@ CREATE FUNCTION foo2_2c(a INT) RETURNS INT RETURN 1 + (SELECT (SELECT a)) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITHOUT_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITHOUT_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`a`" @@ -279,7 +279,7 @@ CREATE FUNCTION foo2_2d(a INT) RETURNS INT RETURN 1 + (SELECT (SELECT (SELECT (S -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITHOUT_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITHOUT_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`a`" @@ -302,7 +302,7 @@ SELECT a + 1 FROM (VALUES 1) AS V(c1) -- !query analysis org.apache.spark.sql.catalyst.analysis.FunctionAlreadyExistsException { - "errorClass" : "ROUTINE_ALREADY_EXISTS", + "condition" : "ROUTINE_ALREADY_EXISTS", "sqlState" : "42723", "messageParameters" : { "existingRoutineType" : "routine", @@ -320,7 +320,7 @@ SELECT a + 1 FROM (VALUES 1) AS V(a) -- !query analysis org.apache.spark.sql.catalyst.analysis.FunctionAlreadyExistsException { - "errorClass" : "ROUTINE_ALREADY_EXISTS", + "condition" : "ROUTINE_ALREADY_EXISTS", "sqlState" : "42723", "messageParameters" : { "existingRoutineType" : "routine", @@ -338,7 +338,7 @@ SELECT a FROM (VALUES 1) AS V(a) -- !query analysis org.apache.spark.sql.catalyst.analysis.FunctionAlreadyExistsException { - "errorClass" : "ROUTINE_ALREADY_EXISTS", + "condition" : "ROUTINE_ALREADY_EXISTS", "sqlState" : "42723", "messageParameters" : { "existingRoutineType" : "routine", @@ -444,7 +444,7 @@ CREATE FUNCTION foo2_3(a INT, b INT) RETURNS INT RETURN a + b -- !query analysis org.apache.spark.sql.catalyst.analysis.FunctionAlreadyExistsException { - "errorClass" : "ROUTINE_ALREADY_EXISTS", + "condition" : "ROUTINE_ALREADY_EXISTS", "sqlState" : "42723", "messageParameters" : { "existingRoutineType" : "routine", @@ -513,7 +513,7 @@ SELECT array_sort(a, (i, j) -> rank[i] - rank[j])[0] FROM (SELECT MAP('a', 1, 'b -- !query analysis org.apache.spark.sql.catalyst.analysis.FunctionAlreadyExistsException { - "errorClass" : "ROUTINE_ALREADY_EXISTS", + "condition" : "ROUTINE_ALREADY_EXISTS", "sqlState" : "42723", "messageParameters" : { "existingRoutineType" : "routine", @@ -541,7 +541,7 @@ SELECT v || ' ' || v FROM (SELECT upper(m[k]) AS v) -- !query analysis org.apache.spark.sql.catalyst.analysis.FunctionAlreadyExistsException { - "errorClass" : "ROUTINE_ALREADY_EXISTS", + "condition" : "ROUTINE_ALREADY_EXISTS", "sqlState" : "42723", "messageParameters" : { "existingRoutineType" : "routine", diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/string-functions.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/string-functions.sql.out index 98664dedf820..02c7fb9292f4 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/string-functions.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/string-functions.sql.out @@ -4,7 +4,7 @@ select concat_ws() -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", + "condition" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", "sqlState" : "42605", "messageParameters" : { "actualNum" : "0", @@ -20,7 +20,7 @@ select format_string() -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", + "condition" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", "sqlState" : "42605", "messageParameters" : { "actualNum" : "0", @@ -779,7 +779,7 @@ select decode() -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", + "condition" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", "sqlState" : "42605", "messageParameters" : { "actualNum" : "0", @@ -802,7 +802,7 @@ select decode(encode('abc', 'utf-8')) -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", + "condition" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", "sqlState" : "42605", "messageParameters" : { "actualNum" : "1", @@ -1508,7 +1508,7 @@ select to_binary('abc', 1) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.INVALID_ARG_VALUE", + "condition" : "DATATYPE_MISMATCH.INVALID_ARG_VALUE", "sqlState" : "42K09", "messageParameters" : { "inputName" : "fmt", @@ -1532,7 +1532,7 @@ select to_binary('abc', 'invalidFormat') -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.INVALID_ARG_VALUE", + "condition" : "DATATYPE_MISMATCH.INVALID_ARG_VALUE", "sqlState" : "42K09", "messageParameters" : { "inputName" : "fmt", @@ -1564,7 +1564,7 @@ SELECT to_binary('abc', fmtField) FROM fmtTable -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.NON_FOLDABLE_INPUT", + "condition" : "DATATYPE_MISMATCH.NON_FOLDABLE_INPUT", "sqlState" : "42K09", "messageParameters" : { "inputExpr" : "\"fmtField\"", diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/subquery/in-subquery/in-basic.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/subquery/in-subquery/in-basic.sql.out index 4d59da8c393a..2014dbce8f99 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/subquery/in-subquery/in-basic.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/subquery/in-subquery/in-basic.sql.out @@ -52,7 +52,7 @@ select 1 from tab_a where (a1, b1) not in (select (a2, b2) from tab_b) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.IN_SUBQUERY_LENGTH_MISMATCH", + "condition" : "DATATYPE_MISMATCH.IN_SUBQUERY_LENGTH_MISMATCH", "sqlState" : "42K09", "messageParameters" : { "leftColumns" : "\"a1\", \"b1\"", diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/subquery/negative-cases/invalid-correlation.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/subquery/negative-cases/invalid-correlation.sql.out index 95b38e1c7e0f..cf6b1d0aa39d 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/subquery/negative-cases/invalid-correlation.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/subquery/negative-cases/invalid-correlation.sql.out @@ -51,7 +51,7 @@ AND t2b = (SELECT max(avg) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "MISSING_GROUP_BY", + "condition" : "MISSING_GROUP_BY", "sqlState" : "42803", "queryContext" : [ { "objectType" : "", @@ -76,7 +76,7 @@ WHERE t1a IN (SELECT min(t2a) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "MISSING_ATTRIBUTES.RESOLVED_ATTRIBUTE_MISSING_FROM_INPUT", + "condition" : "MISSING_ATTRIBUTES.RESOLVED_ATTRIBUTE_MISSING_FROM_INPUT", "sqlState" : "XX000", "messageParameters" : { "input" : "\"min(t2a)\", \"t2c\"", @@ -104,7 +104,7 @@ HAVING EXISTS (SELECT t2a -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "UNSUPPORTED_SUBQUERY_EXPRESSION_CATEGORY.AGGREGATE_FUNCTION_MIXED_OUTER_LOCAL_REFERENCES", + "condition" : "UNSUPPORTED_SUBQUERY_EXPRESSION_CATEGORY.AGGREGATE_FUNCTION_MIXED_OUTER_LOCAL_REFERENCES", "sqlState" : "0A000", "messageParameters" : { "function" : "min((outer(t1.t1a) + t2.t2a))" @@ -131,7 +131,7 @@ WHERE t1a IN (SELECT t2a -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "UNSUPPORTED_SUBQUERY_EXPRESSION_CATEGORY.AGGREGATE_FUNCTION_MIXED_OUTER_LOCAL_REFERENCES", + "condition" : "UNSUPPORTED_SUBQUERY_EXPRESSION_CATEGORY.AGGREGATE_FUNCTION_MIXED_OUTER_LOCAL_REFERENCES", "sqlState" : "0A000", "messageParameters" : { "function" : "min((outer(t2.t2a) + t3.t3a))" @@ -156,7 +156,7 @@ WHERE t1a IN (SELECT t2a -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNSUPPORTED_SUBQUERY_EXPRESSION_CATEGORY.CORRELATED_REFERENCE", + "condition" : "UNSUPPORTED_SUBQUERY_EXPRESSION_CATEGORY.CORRELATED_REFERENCE", "sqlState" : "0A000", "messageParameters" : { "sqlExprs" : "\"min(t2a) AS `min(outer(t2.t2a))`\"" @@ -192,7 +192,7 @@ ON EXISTS (SELECT 1 FROM t2 WHERE t2a > t1a) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`t1a`", diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/subquery/negative-cases/subq-input-typecheck.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/subquery/negative-cases/subq-input-typecheck.sql.out index dbbb3e2d7062..5aef8586a7e8 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/subquery/negative-cases/subq-input-typecheck.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/subquery/negative-cases/subq-input-typecheck.sql.out @@ -75,7 +75,7 @@ FROM t1 -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INVALID_SUBQUERY_EXPRESSION.SCALAR_SUBQUERY_RETURN_MORE_THAN_ONE_OUTPUT_COLUMN", + "condition" : "INVALID_SUBQUERY_EXPRESSION.SCALAR_SUBQUERY_RETURN_MORE_THAN_ONE_OUTPUT_COLUMN", "sqlState" : "42823", "messageParameters" : { "number" : "2" @@ -101,7 +101,7 @@ FROM t1 -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INVALID_SUBQUERY_EXPRESSION.SCALAR_SUBQUERY_RETURN_MORE_THAN_ONE_OUTPUT_COLUMN", + "condition" : "INVALID_SUBQUERY_EXPRESSION.SCALAR_SUBQUERY_RETURN_MORE_THAN_ONE_OUTPUT_COLUMN", "sqlState" : "42823", "messageParameters" : { "number" : "2" @@ -125,7 +125,7 @@ t1a IN (SELECT t2a, t2b -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.IN_SUBQUERY_LENGTH_MISMATCH", + "condition" : "DATATYPE_MISMATCH.IN_SUBQUERY_LENGTH_MISMATCH", "sqlState" : "42K09", "messageParameters" : { "leftColumns" : "\"t1a\"", @@ -153,7 +153,7 @@ WHERE -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.IN_SUBQUERY_LENGTH_MISMATCH", + "condition" : "DATATYPE_MISMATCH.IN_SUBQUERY_LENGTH_MISMATCH", "sqlState" : "42K09", "messageParameters" : { "leftColumns" : "\"t1a\", \"t1b\"", @@ -182,7 +182,7 @@ WHERE -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.IN_SUBQUERY_DATA_TYPE_MISMATCH", + "condition" : "DATATYPE_MISMATCH.IN_SUBQUERY_DATA_TYPE_MISMATCH", "sqlState" : "42K09", "messageParameters" : { "leftType" : "\"DOUBLE\", \"STRING\", \"STRING\"", diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/subquery/scalar-subquery/scalar-subquery-group-by.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/subquery/scalar-subquery/scalar-subquery-group-by.sql.out index 01de7beda551..73a0a82fc86f 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/subquery/scalar-subquery/scalar-subquery-group-by.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/subquery/scalar-subquery/scalar-subquery-group-by.sql.out @@ -42,7 +42,7 @@ select * from x where (select count(*) from y where y1 = x1 group by x1) = 1 -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNSUPPORTED_SUBQUERY_EXPRESSION_CATEGORY.CORRELATED_REFERENCE", + "condition" : "UNSUPPORTED_SUBQUERY_EXPRESSION_CATEGORY.CORRELATED_REFERENCE", "sqlState" : "0A000", "messageParameters" : { "sqlExprs" : "\"x1\"" @@ -62,7 +62,7 @@ select * from x where (select count(*) from y where y1 > x1 group by x1) = 1 -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNSUPPORTED_SUBQUERY_EXPRESSION_CATEGORY.CORRELATED_REFERENCE", + "condition" : "UNSUPPORTED_SUBQUERY_EXPRESSION_CATEGORY.CORRELATED_REFERENCE", "sqlState" : "0A000", "messageParameters" : { "sqlExprs" : "\"x1\"" @@ -153,7 +153,7 @@ select * from x where (select count(*) from y where y1 > x1 group by y1) = 1 -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNSUPPORTED_SUBQUERY_EXPRESSION_CATEGORY.NON_CORRELATED_COLUMNS_IN_GROUP_BY", + "condition" : "UNSUPPORTED_SUBQUERY_EXPRESSION_CATEGORY.NON_CORRELATED_COLUMNS_IN_GROUP_BY", "sqlState" : "0A000", "messageParameters" : { "value" : "y1" @@ -173,7 +173,7 @@ select *, (select count(*) from y where y1 + y2 = x1 group by y1) from x -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNSUPPORTED_SUBQUERY_EXPRESSION_CATEGORY.NON_CORRELATED_COLUMNS_IN_GROUP_BY", + "condition" : "UNSUPPORTED_SUBQUERY_EXPRESSION_CATEGORY.NON_CORRELATED_COLUMNS_IN_GROUP_BY", "sqlState" : "0A000", "messageParameters" : { "value" : "y1" @@ -193,7 +193,7 @@ select *, (select count(*) from y where x1 = y1 and y2 + 10 = x1 + 1 group by y2 -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNSUPPORTED_SUBQUERY_EXPRESSION_CATEGORY.NON_CORRELATED_COLUMNS_IN_GROUP_BY", + "condition" : "UNSUPPORTED_SUBQUERY_EXPRESSION_CATEGORY.NON_CORRELATED_COLUMNS_IN_GROUP_BY", "sqlState" : "0A000", "messageParameters" : { "value" : "y2" @@ -292,7 +292,7 @@ select *, (select count(*) from y left join (select * from z where z1 = x1) sub -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNSUPPORTED_SUBQUERY_EXPRESSION_CATEGORY.ACCESSING_OUTER_QUERY_COLUMN_IS_NOT_ALLOWED", + "condition" : "UNSUPPORTED_SUBQUERY_EXPRESSION_CATEGORY.ACCESSING_OUTER_QUERY_COLUMN_IS_NOT_ALLOWED", "sqlState" : "0A000", "messageParameters" : { "treeNode" : "Filter (z1#x = outer(x1#x))\n+- SubqueryAlias z\n +- View (`z`, [z1#x, z2#x])\n +- Project [cast(col1#x as int) AS z1#x, cast(col2#x as int) AS z2#x]\n +- LocalRelation [col1#x, col2#x]\n" diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/subquery/scalar-subquery/scalar-subquery-predicate.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/subquery/scalar-subquery/scalar-subquery-predicate.sql.out index f8f23009258a..36387235e34d 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/subquery/scalar-subquery/scalar-subquery-predicate.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/subquery/scalar-subquery/scalar-subquery-predicate.sql.out @@ -1562,7 +1562,7 @@ SELECT * FROM t0 WHERE t0a < -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNSUPPORTED_SUBQUERY_EXPRESSION_CATEGORY.CORRELATED_REFERENCE", + "condition" : "UNSUPPORTED_SUBQUERY_EXPRESSION_CATEGORY.CORRELATED_REFERENCE", "sqlState" : "0A000", "messageParameters" : { "sqlExprs" : "\"sum(t0a) AS d\"" diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/subquery/scalar-subquery/scalar-subquery-set-op.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/subquery/scalar-subquery/scalar-subquery-set-op.sql.out index eaeff0ba9ded..7b45deaa9f1f 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/subquery/scalar-subquery/scalar-subquery-set-op.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/subquery/scalar-subquery/scalar-subquery-set-op.sql.out @@ -301,7 +301,7 @@ FROM t0 -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNSUPPORTED_SUBQUERY_EXPRESSION_CATEGORY.CORRELATED_REFERENCE", + "condition" : "UNSUPPORTED_SUBQUERY_EXPRESSION_CATEGORY.CORRELATED_REFERENCE", "sqlState" : "0A000", "messageParameters" : { "sqlExprs" : "\"sum(t0a) AS d\"" @@ -605,7 +605,7 @@ FROM t0 -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNSUPPORTED_SUBQUERY_EXPRESSION_CATEGORY.CORRELATED_REFERENCE", + "condition" : "UNSUPPORTED_SUBQUERY_EXPRESSION_CATEGORY.CORRELATED_REFERENCE", "sqlState" : "0A000", "messageParameters" : { "sqlExprs" : "\"sum(t0a) AS d\"" @@ -901,7 +901,7 @@ FROM t0 -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNSUPPORTED_SUBQUERY_EXPRESSION_CATEGORY.CORRELATED_REFERENCE", + "condition" : "UNSUPPORTED_SUBQUERY_EXPRESSION_CATEGORY.CORRELATED_REFERENCE", "sqlState" : "0A000", "messageParameters" : { "sqlExprs" : "\"sum(t0a) AS d\"" @@ -1197,7 +1197,7 @@ FROM t0 -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNSUPPORTED_SUBQUERY_EXPRESSION_CATEGORY.CORRELATED_REFERENCE", + "condition" : "UNSUPPORTED_SUBQUERY_EXPRESSION_CATEGORY.CORRELATED_REFERENCE", "sqlState" : "0A000", "messageParameters" : { "sqlExprs" : "\"sum(t0a) AS d\"" @@ -1493,7 +1493,7 @@ FROM t0 -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNSUPPORTED_SUBQUERY_EXPRESSION_CATEGORY.CORRELATED_REFERENCE", + "condition" : "UNSUPPORTED_SUBQUERY_EXPRESSION_CATEGORY.CORRELATED_REFERENCE", "sqlState" : "0A000", "messageParameters" : { "sqlExprs" : "\"sum(t0a) AS d\"" @@ -1789,7 +1789,7 @@ FROM t0 -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNSUPPORTED_SUBQUERY_EXPRESSION_CATEGORY.CORRELATED_REFERENCE", + "condition" : "UNSUPPORTED_SUBQUERY_EXPRESSION_CATEGORY.CORRELATED_REFERENCE", "sqlState" : "0A000", "messageParameters" : { "sqlExprs" : "\"sum(t0a) AS d\"" diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/subquery/subquery-nested-data.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/subquery/subquery-nested-data.sql.out index 9cbd10e4d2ca..ab45afe31bd6 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/subquery/subquery-nested-data.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/subquery/subquery-nested-data.sql.out @@ -244,7 +244,7 @@ select * from x join lateral (select xm[1], sum(ym[1]), xm[1] - sum(ym[1]) from -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNSUPPORTED_SUBQUERY_EXPRESSION_CATEGORY.CORRELATED_REFERENCE", + "condition" : "UNSUPPORTED_SUBQUERY_EXPRESSION_CATEGORY.CORRELATED_REFERENCE", "sqlState" : "0A000", "messageParameters" : { "sqlExprs" : "\"xm[1]\",\"xm[1] AS `outer(spark_catalog.default.x.xm)[1]`\",\"(xm[1] - sum(ym[1])) AS `(outer(spark_catalog.default.x.xm)[1] - sum(ym[1]))`\"" diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/table-aliases.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/table-aliases.sql.out index b7eacd1b1b0e..bca87e0b5da5 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/table-aliases.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/table-aliases.sql.out @@ -57,7 +57,7 @@ SELECT * FROM testData AS t(col1, col2, col3) -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "ASSIGNMENT_ARITY_MISMATCH", + "condition" : "ASSIGNMENT_ARITY_MISMATCH", "sqlState" : "42802", "messageParameters" : { "numExpr" : "3", @@ -78,7 +78,7 @@ SELECT * FROM testData AS t(col1) -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "ASSIGNMENT_ARITY_MISMATCH", + "condition" : "ASSIGNMENT_ARITY_MISMATCH", "sqlState" : "42802", "messageParameters" : { "numExpr" : "1", @@ -99,7 +99,7 @@ SELECT a AS col1, b AS col2 FROM testData AS t(c, d) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`a`", @@ -224,7 +224,7 @@ SELECT src1.* FROM src1 a ORDER BY id LIMIT 1 -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "CANNOT_RESOLVE_STAR_EXPAND", + "condition" : "CANNOT_RESOLVE_STAR_EXPAND", "sqlState" : "42704", "messageParameters" : { "columns" : "`id`, `v1`", @@ -245,7 +245,7 @@ SELECT src1.id FROM (SELECT * FROM src1 ORDER BY id LIMIT 1) a -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`src1`.`id`", diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/table-valued-functions.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/table-valued-functions.sql.out index 438e98f559db..3c8e5ac66e65 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/table-valued-functions.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/table-valued-functions.sql.out @@ -4,7 +4,7 @@ select * from dummy(3) -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "UNRESOLVABLE_TABLE_VALUED_FUNCTION", + "condition" : "UNRESOLVABLE_TABLE_VALUED_FUNCTION", "sqlState" : "42883", "messageParameters" : { "name" : "`dummy`" @@ -52,7 +52,7 @@ select * from range(1, 1, 1, 1, 1) -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", + "condition" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", "sqlState" : "42605", "messageParameters" : { "actualNum" : "5", @@ -75,7 +75,7 @@ select * from range(1, null) -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "UNEXPECTED_INPUT_TYPE", + "condition" : "UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "functionName" : "`range`", @@ -99,7 +99,7 @@ select * from range(array(1, 2, 3)) -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "UNEXPECTED_INPUT_TYPE", + "condition" : "UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "functionName" : "`range`", @@ -123,7 +123,7 @@ select * from range(0, 5, 0) -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "FAILED_FUNCTION_CALL", + "condition" : "FAILED_FUNCTION_CALL", "sqlState" : "38000", "messageParameters" : { "funcName" : "`range`" @@ -159,7 +159,7 @@ select * from range(0, (select 1)) -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "NON_FOLDABLE_ARGUMENT", + "condition" : "NON_FOLDABLE_ARGUMENT", "sqlState" : "42K08", "messageParameters" : { "funcName" : "`range`", @@ -181,7 +181,7 @@ select * from values (0, 1) t(c1, c2), lateral range(0, c2) -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "NON_FOLDABLE_ARGUMENT", + "condition" : "NON_FOLDABLE_ARGUMENT", "sqlState" : "42K08", "messageParameters" : { "funcName" : "`range`", @@ -261,7 +261,7 @@ select * from explode(null) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"NULL\"", @@ -285,7 +285,7 @@ select * from explode(null) t(c1) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"NULL\"", @@ -309,7 +309,7 @@ select * from explode(1) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"1\"", @@ -333,7 +333,7 @@ select * from explode(1, 2) -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", + "condition" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", "sqlState" : "42605", "messageParameters" : { "actualNum" : "2", @@ -356,7 +356,7 @@ select * from explode(explode(array(1))) -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "UNSUPPORTED_GENERATOR.NESTED_IN_EXPRESSIONS", + "condition" : "UNSUPPORTED_GENERATOR.NESTED_IN_EXPRESSIONS", "sqlState" : "42K0E", "messageParameters" : { "expression" : "\"explode(explode(array(1)))\"" @@ -376,7 +376,7 @@ select * from explode(array(1, 2)) t(c1, c2) -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "NUM_TABLE_VALUE_ALIASES_MISMATCH", + "condition" : "NUM_TABLE_VALUE_ALIASES_MISMATCH", "sqlState" : "42826", "messageParameters" : { "aliasesNum" : "2", @@ -476,7 +476,7 @@ select * from inline(null) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"NULL\"", @@ -500,7 +500,7 @@ select * from inline(array(struct(1, 2), struct(2, 3))) t(a, b, c) -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "NUM_TABLE_VALUE_ALIASES_MISMATCH", + "condition" : "NUM_TABLE_VALUE_ALIASES_MISMATCH", "sqlState" : "42826", "messageParameters" : { "aliasesNum" : "3", @@ -590,7 +590,7 @@ select * from posexplode(1) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"1\"", @@ -614,7 +614,7 @@ select * from posexplode(1, 2) -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", + "condition" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", "sqlState" : "42605", "messageParameters" : { "actualNum" : "2", @@ -637,7 +637,7 @@ select * from posexplode(explode(array(1))) -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "UNSUPPORTED_GENERATOR.NESTED_IN_EXPRESSIONS", + "condition" : "UNSUPPORTED_GENERATOR.NESTED_IN_EXPRESSIONS", "sqlState" : "42K0E", "messageParameters" : { "expression" : "\"posexplode(explode(array(1)))\"" @@ -657,7 +657,7 @@ select * from posexplode(array(1, 2)) t(x) -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "NUM_TABLE_VALUE_ALIASES_MISMATCH", + "condition" : "NUM_TABLE_VALUE_ALIASES_MISMATCH", "sqlState" : "42826", "messageParameters" : { "aliasesNum" : "1", @@ -753,7 +753,7 @@ select * from json_tuple() -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", + "condition" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", "sqlState" : "42605", "messageParameters" : { "actualNum" : "0", @@ -769,7 +769,7 @@ select * from json_tuple('{"a": 1}') -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", + "condition" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", "sqlState" : "42605", "messageParameters" : { "actualNum" : "1", @@ -785,7 +785,7 @@ select * from json_tuple('{"a": 1}', 1) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.NON_STRING_TYPE", + "condition" : "DATATYPE_MISMATCH.NON_STRING_TYPE", "sqlState" : "42K09", "messageParameters" : { "funcName" : "`json_tuple`", @@ -806,7 +806,7 @@ select * from json_tuple('{"a": 1}', null) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.NON_STRING_TYPE", + "condition" : "DATATYPE_MISMATCH.NON_STRING_TYPE", "sqlState" : "42K09", "messageParameters" : { "funcName" : "`json_tuple`", @@ -827,7 +827,7 @@ select * from json_tuple('{"a": 1, "b": 2}', 'a', 'b') AS t(x) -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "NUM_TABLE_VALUE_ALIASES_MISMATCH", + "condition" : "NUM_TABLE_VALUE_ALIASES_MISMATCH", "sqlState" : "42826", "messageParameters" : { "aliasesNum" : "1", @@ -905,7 +905,7 @@ select * from stack() -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", + "condition" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", "sqlState" : "42605", "messageParameters" : { "actualNum" : "0", @@ -921,7 +921,7 @@ select * from stack(2, 1, 2, 3) t(a, b, c) -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "NUM_TABLE_VALUE_ALIASES_MISMATCH", + "condition" : "NUM_TABLE_VALUE_ALIASES_MISMATCH", "sqlState" : "42826", "messageParameters" : { "aliasesNum" : "3", @@ -943,7 +943,7 @@ select * from stack(2, 1, '1.1', 'a', 2, 2.2, 'b') -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.STACK_COLUMN_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.STACK_COLUMN_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "columnIndex" : "1", @@ -968,7 +968,7 @@ select * from stack(2, explode(array(1, 2, 3))) -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "UNSUPPORTED_GENERATOR.NESTED_IN_EXPRESSIONS", + "condition" : "UNSUPPORTED_GENERATOR.NESTED_IN_EXPRESSIONS", "sqlState" : "42K0E", "messageParameters" : { "expression" : "\"stack(2, explode(array(1, 2, 3)))\"" diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/tablesample-negative.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/tablesample-negative.sql.out index 57ae6fb7dc7d..5627c8ef3c66 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/tablesample-negative.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/tablesample-negative.sql.out @@ -26,7 +26,7 @@ SELECT mydb1.t1 FROM t1 TABLESAMPLE (-1 PERCENT) -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0064", + "condition" : "_LEGACY_ERROR_TEMP_0064", "messageParameters" : { "msg" : "Sampling fraction (-0.01) must be on interval [0, 1]" }, @@ -45,7 +45,7 @@ SELECT mydb1.t1 FROM t1 TABLESAMPLE (101 PERCENT) -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0064", + "condition" : "_LEGACY_ERROR_TEMP_0064", "messageParameters" : { "msg" : "Sampling fraction (1.01) must be on interval [0, 1]" }, diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/timestamp-ntz.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/timestamp-ntz.sql.out index e92a392e22b6..4fe9ffa55988 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/timestamp-ntz.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/timestamp-ntz.sql.out @@ -39,7 +39,7 @@ SELECT make_timestamp_ntz(2021, 07, 11, 6, 30, 45.678, 'CET') -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", + "condition" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", "sqlState" : "42605", "messageParameters" : { "actualNum" : "7", diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/timestamp.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/timestamp.sql.out index 560974d28c54..8950230de72d 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/timestamp.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/timestamp.sql.out @@ -10,7 +10,7 @@ select timestamp '2019-01-01中文' -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_TYPED_LITERAL", + "condition" : "INVALID_TYPED_LITERAL", "sqlState" : "42604", "messageParameters" : { "value" : "'2019-01-01中文'", @@ -31,7 +31,7 @@ select timestamp'4294967297' -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_TYPED_LITERAL", + "condition" : "INVALID_TYPED_LITERAL", "sqlState" : "42604", "messageParameters" : { "value" : "'4294967297'", @@ -52,7 +52,7 @@ select timestamp'2021-01-01T12:30:4294967297.123456' -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_TYPED_LITERAL", + "condition" : "INVALID_TYPED_LITERAL", "sqlState" : "42604", "messageParameters" : { "value" : "'2021-01-01T12:30:4294967297.123456'", @@ -597,7 +597,7 @@ select timestamp'2011-11-11 11:11:11' + '1' -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"TIMESTAMP\"", @@ -619,7 +619,7 @@ select '1' + timestamp'2011-11-11 11:11:11' -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"TIMESTAMP\"", @@ -641,7 +641,7 @@ select timestamp'2011-11-11 11:11:11' + null -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP\"", @@ -663,7 +663,7 @@ select null + timestamp'2011-11-11 11:11:11' -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"VOID\"", @@ -770,7 +770,7 @@ select timestampadd('MONTH', -1, timestamp'2022-02-14 01:02:03') -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", + "condition" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", "sqlState" : "22023", "messageParameters" : { "functionName" : "`timestampadd`", @@ -792,7 +792,7 @@ select timestampadd('SECOND', -1, date'2022-02-15') -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", + "condition" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", "sqlState" : "22023", "messageParameters" : { "functionName" : "`timestampadd`", @@ -838,7 +838,7 @@ select timestampdiff('MINUTE', timestamp'2022-02-14 01:02:03', timestamp'2022-02 -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", + "condition" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", "sqlState" : "22023", "messageParameters" : { "functionName" : "`timestampdiff`", @@ -860,7 +860,7 @@ select timestampdiff('YEAR', date'2022-02-15', date'2023-02-15') -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", + "condition" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", "sqlState" : "22023", "messageParameters" : { "functionName" : "`timestampdiff`", @@ -906,7 +906,7 @@ select timediff('MINUTE', timestamp'2023-02-14 01:02:03', timestamp'2023-02-14 0 -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", + "condition" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", "sqlState" : "22023", "messageParameters" : { "functionName" : "`timediff`", @@ -928,7 +928,7 @@ select timediff('YEAR', date'2020-02-15', date'2023-02-15') -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", + "condition" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", "sqlState" : "22023", "messageParameters" : { "functionName" : "`timediff`", diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/timestampNTZ/timestamp-ansi.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/timestampNTZ/timestamp-ansi.sql.out index f900293918c6..3e185decc257 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/timestampNTZ/timestamp-ansi.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/timestampNTZ/timestamp-ansi.sql.out @@ -11,7 +11,7 @@ select timestamp '2019-01-01中文' -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_TYPED_LITERAL", + "condition" : "INVALID_TYPED_LITERAL", "sqlState" : "42604", "messageParameters" : { "value" : "'2019-01-01中文'", @@ -32,7 +32,7 @@ select timestamp'4294967297' -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_TYPED_LITERAL", + "condition" : "INVALID_TYPED_LITERAL", "sqlState" : "42604", "messageParameters" : { "value" : "'4294967297'", @@ -53,7 +53,7 @@ select timestamp'2021-01-01T12:30:4294967297.123456' -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_TYPED_LITERAL", + "condition" : "INVALID_TYPED_LITERAL", "sqlState" : "42604", "messageParameters" : { "value" : "'2021-01-01T12:30:4294967297.123456'", @@ -611,7 +611,7 @@ select timestamp'2011-11-11 11:11:11' + '1' -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"TIMESTAMP_NTZ\"", @@ -633,7 +633,7 @@ select '1' + timestamp'2011-11-11 11:11:11' -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"TIMESTAMP_NTZ\"", @@ -655,7 +655,7 @@ select timestamp'2011-11-11 11:11:11' + null -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP_NTZ\"", @@ -677,7 +677,7 @@ select null + timestamp'2011-11-11 11:11:11' -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"VOID\"", @@ -787,7 +787,7 @@ select timestampadd('MONTH', -1, timestamp'2022-02-14 01:02:03') -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", + "condition" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", "sqlState" : "22023", "messageParameters" : { "functionName" : "`timestampadd`", @@ -809,7 +809,7 @@ select timestampadd('SECOND', -1, date'2022-02-15') -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", + "condition" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", "sqlState" : "22023", "messageParameters" : { "functionName" : "`timestampadd`", @@ -857,7 +857,7 @@ select timestampdiff('MINUTE', timestamp'2022-02-14 01:02:03', timestamp'2022-02 -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", + "condition" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", "sqlState" : "22023", "messageParameters" : { "functionName" : "`timestampdiff`", @@ -879,7 +879,7 @@ select timestampdiff('YEAR', date'2022-02-15', date'2023-02-15') -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", + "condition" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", "sqlState" : "22023", "messageParameters" : { "functionName" : "`timestampdiff`", @@ -927,7 +927,7 @@ select timediff('MINUTE', timestamp'2023-02-14 01:02:03', timestamp'2023-02-14 0 -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", + "condition" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", "sqlState" : "22023", "messageParameters" : { "functionName" : "`timediff`", @@ -949,7 +949,7 @@ select timediff('YEAR', date'2020-02-15', date'2023-02-15') -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", + "condition" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", "sqlState" : "22023", "messageParameters" : { "functionName" : "`timediff`", diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/timestampNTZ/timestamp.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/timestampNTZ/timestamp.sql.out index 6dd91c8cdccf..0134892e0caa 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/timestampNTZ/timestamp.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/timestampNTZ/timestamp.sql.out @@ -11,7 +11,7 @@ select timestamp '2019-01-01中文' -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_TYPED_LITERAL", + "condition" : "INVALID_TYPED_LITERAL", "sqlState" : "42604", "messageParameters" : { "value" : "'2019-01-01中文'", @@ -32,7 +32,7 @@ select timestamp'4294967297' -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_TYPED_LITERAL", + "condition" : "INVALID_TYPED_LITERAL", "sqlState" : "42604", "messageParameters" : { "value" : "'4294967297'", @@ -53,7 +53,7 @@ select timestamp'2021-01-01T12:30:4294967297.123456' -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_TYPED_LITERAL", + "condition" : "INVALID_TYPED_LITERAL", "sqlState" : "42604", "messageParameters" : { "value" : "'2021-01-01T12:30:4294967297.123456'", @@ -555,7 +555,7 @@ select timestamp'2011-11-11 11:11:11' - '2011-11-11 11:11:10' -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"2011-11-11 11:11:10\"", @@ -579,7 +579,7 @@ select '2011-11-11 11:11:11' - timestamp'2011-11-11 11:11:10' -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"2011-11-11 11:11:11\"", @@ -623,7 +623,7 @@ select str - timestamp'2011-11-11 11:11:11' from ts_view -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"str\"", @@ -647,7 +647,7 @@ select timestamp'2011-11-11 11:11:11' - str from ts_view -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"str\"", @@ -671,7 +671,7 @@ select timestamp'2011-11-11 11:11:11' + '1' -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP_NTZ\"", @@ -693,7 +693,7 @@ select '1' + timestamp'2011-11-11 11:11:11' -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DOUBLE\"", @@ -715,7 +715,7 @@ select timestamp'2011-11-11 11:11:11' + null -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP_NTZ\"", @@ -737,7 +737,7 @@ select null + timestamp'2011-11-11 11:11:11' -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"VOID\"", @@ -847,7 +847,7 @@ select timestampadd('MONTH', -1, timestamp'2022-02-14 01:02:03') -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", + "condition" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", "sqlState" : "22023", "messageParameters" : { "functionName" : "`timestampadd`", @@ -869,7 +869,7 @@ select timestampadd('SECOND', -1, date'2022-02-15') -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", + "condition" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", "sqlState" : "22023", "messageParameters" : { "functionName" : "`timestampadd`", @@ -917,7 +917,7 @@ select timestampdiff('MINUTE', timestamp'2022-02-14 01:02:03', timestamp'2022-02 -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", + "condition" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", "sqlState" : "22023", "messageParameters" : { "functionName" : "`timestampdiff`", @@ -939,7 +939,7 @@ select timestampdiff('YEAR', date'2022-02-15', date'2023-02-15') -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", + "condition" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", "sqlState" : "22023", "messageParameters" : { "functionName" : "`timestampdiff`", @@ -987,7 +987,7 @@ select timediff('MINUTE', timestamp'2023-02-14 01:02:03', timestamp'2023-02-14 0 -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", + "condition" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", "sqlState" : "22023", "messageParameters" : { "functionName" : "`timediff`", @@ -1009,7 +1009,7 @@ select timediff('YEAR', date'2020-02-15', date'2023-02-15') -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", + "condition" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", "sqlState" : "22023", "messageParameters" : { "functionName" : "`timediff`", diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/timezone.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/timezone.sql.out index 5b55a0c21893..3038f848705f 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/timezone.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/timezone.sql.out @@ -34,7 +34,7 @@ SET TIME ZONE -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0045", + "condition" : "_LEGACY_ERROR_TEMP_0045", "queryContext" : [ { "objectType" : "", "objectName" : "", @@ -50,7 +50,7 @@ SET TIME ZONE 'invalid/zone' -- !query analysis org.apache.spark.SparkIllegalArgumentException { - "errorClass" : "INVALID_CONF_VALUE.TIME_ZONE", + "condition" : "INVALID_CONF_VALUE.TIME_ZONE", "sqlState" : "22022", "messageParameters" : { "confName" : "spark.sql.session.timeZone", @@ -64,7 +64,7 @@ SET TIME ZONE INTERVAL 3 DAYS -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_INTERVAL_FORMAT.TIMEZONE_INTERVAL_OUT_OF_RANGE", + "condition" : "INVALID_INTERVAL_FORMAT.TIMEZONE_INTERVAL_OUT_OF_RANGE", "sqlState" : "22006", "messageParameters" : { "input" : "3" @@ -84,7 +84,7 @@ SET TIME ZONE INTERVAL 24 HOURS -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_INTERVAL_FORMAT.TIMEZONE_INTERVAL_OUT_OF_RANGE", + "condition" : "INVALID_INTERVAL_FORMAT.TIMEZONE_INTERVAL_OUT_OF_RANGE", "sqlState" : "22006", "messageParameters" : { "input" : "24" @@ -104,7 +104,7 @@ SET TIME ZONE INTERVAL '19:40:32' HOUR TO SECOND -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_INTERVAL_FORMAT.TIMEZONE_INTERVAL_OUT_OF_RANGE", + "condition" : "INVALID_INTERVAL_FORMAT.TIMEZONE_INTERVAL_OUT_OF_RANGE", "sqlState" : "22006", "messageParameters" : { "input" : "19" @@ -124,7 +124,7 @@ SET TIME ZONE INTERVAL 10 HOURS 'GMT+1' -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0045", + "condition" : "_LEGACY_ERROR_TEMP_0045", "queryContext" : [ { "objectType" : "", "objectName" : "", @@ -140,7 +140,7 @@ SET TIME ZONE INTERVAL 10 HOURS 1 MILLISECOND -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_INTERVAL_FORMAT.TIMEZONE_INTERVAL_OUT_OF_RANGE", + "condition" : "INVALID_INTERVAL_FORMAT.TIMEZONE_INTERVAL_OUT_OF_RANGE", "sqlState" : "22006", "messageParameters" : { "input" : "36000" diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/to_from_avro.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/to_from_avro.sql.out index 8275e4f1c0ff..9fc724641d30 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/to_from_avro.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/to_from_avro.sql.out @@ -31,7 +31,7 @@ select from_avro(s, 42, map()) from t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.TYPE_CHECK_FAILURE_WITH_HINT", + "condition" : "DATATYPE_MISMATCH.TYPE_CHECK_FAILURE_WITH_HINT", "sqlState" : "42K09", "messageParameters" : { "hint" : "", @@ -53,7 +53,7 @@ select from_avro(s, avro_schema, 42) from t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.TYPE_CHECK_FAILURE_WITH_HINT", + "condition" : "DATATYPE_MISMATCH.TYPE_CHECK_FAILURE_WITH_HINT", "sqlState" : "42K09", "messageParameters" : { "hint" : "", @@ -75,7 +75,7 @@ select to_avro(s, 42) from t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.TYPE_CHECK_FAILURE_WITH_HINT", + "condition" : "DATATYPE_MISMATCH.TYPE_CHECK_FAILURE_WITH_HINT", "sqlState" : "42K09", "messageParameters" : { "hint" : "", @@ -97,7 +97,7 @@ select to_avro(s, avro_schema) as result from t -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "AVRO_NOT_LOADED_SQL_FUNCTIONS_UNUSABLE", + "condition" : "AVRO_NOT_LOADED_SQL_FUNCTIONS_UNUSABLE", "sqlState" : "22KD3", "messageParameters" : { "functionName" : "TO_AVRO" @@ -110,7 +110,7 @@ select from_avro(result, avro_schema, map()).u from (select null as result) -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "AVRO_NOT_LOADED_SQL_FUNCTIONS_UNUSABLE", + "condition" : "AVRO_NOT_LOADED_SQL_FUNCTIONS_UNUSABLE", "sqlState" : "22KD3", "messageParameters" : { "functionName" : "FROM_AVRO" diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/transform.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/transform.sql.out index d5376913f9ae..267500842487 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/transform.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/transform.sql.out @@ -495,7 +495,7 @@ SELECT a, b, decode(c, 'UTF-8'), d, e, f, g, h, i, j, k, l FROM ( -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0064", + "condition" : "_LEGACY_ERROR_TEMP_0064", "messageParameters" : { "msg" : "LINES TERMINATED BY only supports newline '\\n' right now: @" }, @@ -902,7 +902,7 @@ WHERE a <= 4 -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "UNSUPPORTED_FEATURE.TRANSFORM_DISTINCT_ALL", + "condition" : "UNSUPPORTED_FEATURE.TRANSFORM_DISTINCT_ALL", "sqlState" : "0A000", "queryContext" : [ { "objectType" : "", @@ -922,7 +922,7 @@ WHERE a <= 4 -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "UNSUPPORTED_FEATURE.TRANSFORM_DISTINCT_ALL", + "condition" : "UNSUPPORTED_FEATURE.TRANSFORM_DISTINCT_ALL", "sqlState" : "0A000", "queryContext" : [ { "objectType" : "", @@ -943,7 +943,7 @@ GROUP BY b -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'AS'", @@ -961,7 +961,7 @@ GROUP BY b -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'b_1'", @@ -979,7 +979,7 @@ GROUP BY b -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'AS'", diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/try_arithmetic.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/try_arithmetic.sql.out index 30654d1d71e2..891eca603da7 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/try_arithmetic.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/try_arithmetic.sql.out @@ -150,7 +150,7 @@ SELECT try_add(interval 2 year, interval 2 second) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"INTERVAL '2' YEAR\"", diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/try_reflect.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/try_reflect.sql.out index 7b395b9e0a05..21dabadf81dd 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/try_reflect.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/try_reflect.sql.out @@ -46,7 +46,7 @@ SELECT try_reflect("java.wrongclass.Math", "max", 2, 3) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_CLASS_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_CLASS_TYPE", "sqlState" : "42K09", "messageParameters" : { "className" : "java.wrongclass.Math", @@ -67,7 +67,7 @@ SELECT try_reflect("java.lang.Math", "wrongmethod", 2, 3) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_STATIC_METHOD", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_STATIC_METHOD", "sqlState" : "42K09", "messageParameters" : { "className" : "java.lang.Math", @@ -89,7 +89,7 @@ SELECT try_reflect("java.lang.Math") -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", + "condition" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", "sqlState" : "42605", "messageParameters" : { "actualNum" : "1", @@ -105,7 +105,7 @@ SELECT try_reflect("java.lang.Math", "round", 2.5) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"2.5\"", @@ -129,7 +129,7 @@ SELECT try_reflect("java.lang.Object", "toString") -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_STATIC_METHOD", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_STATIC_METHOD", "sqlState" : "42K09", "messageParameters" : { "className" : "java.lang.Object", diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/typeCoercion/native/binaryComparison.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/typeCoercion/native/binaryComparison.sql.out index 1395cc350db7..4df7c020ccfb 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/typeCoercion/native/binaryComparison.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/typeCoercion/native/binaryComparison.sql.out @@ -12,7 +12,7 @@ SELECT cast(1 as binary) = '1' FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", + "condition" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", "sqlState" : "42K09", "messageParameters" : { "config" : "\"spark.sql.ansi.enabled\"", @@ -36,7 +36,7 @@ SELECT cast(1 as binary) > '2' FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", + "condition" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", "sqlState" : "42K09", "messageParameters" : { "config" : "\"spark.sql.ansi.enabled\"", @@ -60,7 +60,7 @@ SELECT cast(1 as binary) >= '2' FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", + "condition" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", "sqlState" : "42K09", "messageParameters" : { "config" : "\"spark.sql.ansi.enabled\"", @@ -84,7 +84,7 @@ SELECT cast(1 as binary) < '2' FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", + "condition" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", "sqlState" : "42K09", "messageParameters" : { "config" : "\"spark.sql.ansi.enabled\"", @@ -108,7 +108,7 @@ SELECT cast(1 as binary) <= '2' FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", + "condition" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", "sqlState" : "42K09", "messageParameters" : { "config" : "\"spark.sql.ansi.enabled\"", @@ -132,7 +132,7 @@ SELECT cast(1 as binary) <> '2' FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", + "condition" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", "sqlState" : "42K09", "messageParameters" : { "config" : "\"spark.sql.ansi.enabled\"", @@ -156,7 +156,7 @@ SELECT cast(1 as binary) = cast(null as string) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", + "condition" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", "sqlState" : "42K09", "messageParameters" : { "config" : "\"spark.sql.ansi.enabled\"", @@ -180,7 +180,7 @@ SELECT cast(1 as binary) > cast(null as string) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", + "condition" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", "sqlState" : "42K09", "messageParameters" : { "config" : "\"spark.sql.ansi.enabled\"", @@ -204,7 +204,7 @@ SELECT cast(1 as binary) >= cast(null as string) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", + "condition" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", "sqlState" : "42K09", "messageParameters" : { "config" : "\"spark.sql.ansi.enabled\"", @@ -228,7 +228,7 @@ SELECT cast(1 as binary) < cast(null as string) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", + "condition" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", "sqlState" : "42K09", "messageParameters" : { "config" : "\"spark.sql.ansi.enabled\"", @@ -252,7 +252,7 @@ SELECT cast(1 as binary) <= cast(null as string) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", + "condition" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", "sqlState" : "42K09", "messageParameters" : { "config" : "\"spark.sql.ansi.enabled\"", @@ -276,7 +276,7 @@ SELECT cast(1 as binary) <> cast(null as string) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", + "condition" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", "sqlState" : "42K09", "messageParameters" : { "config" : "\"spark.sql.ansi.enabled\"", @@ -300,7 +300,7 @@ SELECT '1' = cast(1 as binary) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", + "condition" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", "sqlState" : "42K09", "messageParameters" : { "config" : "\"spark.sql.ansi.enabled\"", @@ -324,7 +324,7 @@ SELECT '2' > cast(1 as binary) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", + "condition" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", "sqlState" : "42K09", "messageParameters" : { "config" : "\"spark.sql.ansi.enabled\"", @@ -348,7 +348,7 @@ SELECT '2' >= cast(1 as binary) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", + "condition" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", "sqlState" : "42K09", "messageParameters" : { "config" : "\"spark.sql.ansi.enabled\"", @@ -372,7 +372,7 @@ SELECT '2' < cast(1 as binary) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", + "condition" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", "sqlState" : "42K09", "messageParameters" : { "config" : "\"spark.sql.ansi.enabled\"", @@ -396,7 +396,7 @@ SELECT '2' <= cast(1 as binary) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", + "condition" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", "sqlState" : "42K09", "messageParameters" : { "config" : "\"spark.sql.ansi.enabled\"", @@ -420,7 +420,7 @@ SELECT '2' <> cast(1 as binary) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", + "condition" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", "sqlState" : "42K09", "messageParameters" : { "config" : "\"spark.sql.ansi.enabled\"", @@ -444,7 +444,7 @@ SELECT cast(null as string) = cast(1 as binary) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", + "condition" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", "sqlState" : "42K09", "messageParameters" : { "config" : "\"spark.sql.ansi.enabled\"", @@ -468,7 +468,7 @@ SELECT cast(null as string) > cast(1 as binary) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", + "condition" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", "sqlState" : "42K09", "messageParameters" : { "config" : "\"spark.sql.ansi.enabled\"", @@ -492,7 +492,7 @@ SELECT cast(null as string) >= cast(1 as binary) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", + "condition" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", "sqlState" : "42K09", "messageParameters" : { "config" : "\"spark.sql.ansi.enabled\"", @@ -516,7 +516,7 @@ SELECT cast(null as string) < cast(1 as binary) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", + "condition" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", "sqlState" : "42K09", "messageParameters" : { "config" : "\"spark.sql.ansi.enabled\"", @@ -540,7 +540,7 @@ SELECT cast(null as string) <= cast(1 as binary) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", + "condition" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", "sqlState" : "42K09", "messageParameters" : { "config" : "\"spark.sql.ansi.enabled\"", @@ -564,7 +564,7 @@ SELECT cast(null as string) <> cast(1 as binary) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", + "condition" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", "sqlState" : "42K09", "messageParameters" : { "config" : "\"spark.sql.ansi.enabled\"", diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/typeCoercion/native/booleanEquality.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/typeCoercion/native/booleanEquality.sql.out index 4fe85374f4b1..7c7bff97aad3 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/typeCoercion/native/booleanEquality.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/typeCoercion/native/booleanEquality.sql.out @@ -12,7 +12,7 @@ SELECT true = cast(1 as tinyint) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BOOLEAN\"", @@ -34,7 +34,7 @@ SELECT true = cast(1 as smallint) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BOOLEAN\"", @@ -56,7 +56,7 @@ SELECT true = cast(1 as int) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BOOLEAN\"", @@ -78,7 +78,7 @@ SELECT true = cast(1 as bigint) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BOOLEAN\"", @@ -100,7 +100,7 @@ SELECT true = cast(1 as float) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BOOLEAN\"", @@ -122,7 +122,7 @@ SELECT true = cast(1 as double) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BOOLEAN\"", @@ -144,7 +144,7 @@ SELECT true = cast(1 as decimal(10, 0)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BOOLEAN\"", @@ -177,7 +177,7 @@ SELECT true = cast('1' as binary) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BOOLEAN\"", @@ -210,7 +210,7 @@ SELECT true = cast('2017-12-11 09:30:00.0' as timestamp) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BOOLEAN\"", @@ -232,7 +232,7 @@ SELECT true = cast('2017-12-11 09:30:00' as date) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BOOLEAN\"", @@ -254,7 +254,7 @@ SELECT true <=> cast(1 as tinyint) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BOOLEAN\"", @@ -276,7 +276,7 @@ SELECT true <=> cast(1 as smallint) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BOOLEAN\"", @@ -298,7 +298,7 @@ SELECT true <=> cast(1 as int) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BOOLEAN\"", @@ -320,7 +320,7 @@ SELECT true <=> cast(1 as bigint) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BOOLEAN\"", @@ -342,7 +342,7 @@ SELECT true <=> cast(1 as float) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BOOLEAN\"", @@ -364,7 +364,7 @@ SELECT true <=> cast(1 as double) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BOOLEAN\"", @@ -386,7 +386,7 @@ SELECT true <=> cast(1 as decimal(10, 0)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BOOLEAN\"", @@ -419,7 +419,7 @@ SELECT true <=> cast('1' as binary) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BOOLEAN\"", @@ -452,7 +452,7 @@ SELECT true <=> cast('2017-12-11 09:30:00.0' as timestamp) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BOOLEAN\"", @@ -474,7 +474,7 @@ SELECT true <=> cast('2017-12-11 09:30:00' as date) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BOOLEAN\"", @@ -496,7 +496,7 @@ SELECT cast(1 as tinyint) = true FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TINYINT\"", @@ -518,7 +518,7 @@ SELECT cast(1 as smallint) = true FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"SMALLINT\"", @@ -540,7 +540,7 @@ SELECT cast(1 as int) = true FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"INT\"", @@ -562,7 +562,7 @@ SELECT cast(1 as bigint) = true FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BIGINT\"", @@ -584,7 +584,7 @@ SELECT cast(1 as float) = true FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"FLOAT\"", @@ -606,7 +606,7 @@ SELECT cast(1 as double) = true FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DOUBLE\"", @@ -628,7 +628,7 @@ SELECT cast(1 as decimal(10, 0)) = true FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(10,0)\"", @@ -661,7 +661,7 @@ SELECT cast('1' as binary) = true FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BINARY\"", @@ -694,7 +694,7 @@ SELECT cast('2017-12-11 09:30:00.0' as timestamp) = true FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP\"", @@ -716,7 +716,7 @@ SELECT cast('2017-12-11 09:30:00' as date) = true FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DATE\"", @@ -738,7 +738,7 @@ SELECT cast(1 as tinyint) <=> true FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TINYINT\"", @@ -760,7 +760,7 @@ SELECT cast(1 as smallint) <=> true FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"SMALLINT\"", @@ -782,7 +782,7 @@ SELECT cast(1 as int) <=> true FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"INT\"", @@ -804,7 +804,7 @@ SELECT cast(1 as bigint) <=> true FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BIGINT\"", @@ -826,7 +826,7 @@ SELECT cast(1 as float) <=> true FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"FLOAT\"", @@ -848,7 +848,7 @@ SELECT cast(1 as double) <=> true FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DOUBLE\"", @@ -870,7 +870,7 @@ SELECT cast(1 as decimal(10, 0)) <=> true FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(10,0)\"", @@ -903,7 +903,7 @@ SELECT cast('1' as binary) <=> true FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BINARY\"", @@ -936,7 +936,7 @@ SELECT cast('2017-12-11 09:30:00.0' as timestamp) <=> true FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP\"", @@ -958,7 +958,7 @@ SELECT cast('2017-12-11 09:30:00' as date) <=> true FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DATE\"", @@ -980,7 +980,7 @@ SELECT false = cast(0 as tinyint) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BOOLEAN\"", @@ -1002,7 +1002,7 @@ SELECT false = cast(0 as smallint) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BOOLEAN\"", @@ -1024,7 +1024,7 @@ SELECT false = cast(0 as int) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BOOLEAN\"", @@ -1046,7 +1046,7 @@ SELECT false = cast(0 as bigint) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BOOLEAN\"", @@ -1068,7 +1068,7 @@ SELECT false = cast(0 as float) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BOOLEAN\"", @@ -1090,7 +1090,7 @@ SELECT false = cast(0 as double) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BOOLEAN\"", @@ -1112,7 +1112,7 @@ SELECT false = cast(0 as decimal(10, 0)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BOOLEAN\"", @@ -1145,7 +1145,7 @@ SELECT false = cast('0' as binary) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BOOLEAN\"", @@ -1178,7 +1178,7 @@ SELECT false = cast('2017-12-11 09:30:00.0' as timestamp) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BOOLEAN\"", @@ -1200,7 +1200,7 @@ SELECT false = cast('2017-12-11 09:30:00' as date) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BOOLEAN\"", @@ -1222,7 +1222,7 @@ SELECT false <=> cast(0 as tinyint) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BOOLEAN\"", @@ -1244,7 +1244,7 @@ SELECT false <=> cast(0 as smallint) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BOOLEAN\"", @@ -1266,7 +1266,7 @@ SELECT false <=> cast(0 as int) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BOOLEAN\"", @@ -1288,7 +1288,7 @@ SELECT false <=> cast(0 as bigint) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BOOLEAN\"", @@ -1310,7 +1310,7 @@ SELECT false <=> cast(0 as float) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BOOLEAN\"", @@ -1332,7 +1332,7 @@ SELECT false <=> cast(0 as double) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BOOLEAN\"", @@ -1354,7 +1354,7 @@ SELECT false <=> cast(0 as decimal(10, 0)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BOOLEAN\"", @@ -1387,7 +1387,7 @@ SELECT false <=> cast('0' as binary) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BOOLEAN\"", @@ -1420,7 +1420,7 @@ SELECT false <=> cast('2017-12-11 09:30:00.0' as timestamp) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BOOLEAN\"", @@ -1442,7 +1442,7 @@ SELECT false <=> cast('2017-12-11 09:30:00' as date) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BOOLEAN\"", @@ -1464,7 +1464,7 @@ SELECT cast(0 as tinyint) = false FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TINYINT\"", @@ -1486,7 +1486,7 @@ SELECT cast(0 as smallint) = false FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"SMALLINT\"", @@ -1508,7 +1508,7 @@ SELECT cast(0 as int) = false FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"INT\"", @@ -1530,7 +1530,7 @@ SELECT cast(0 as bigint) = false FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BIGINT\"", @@ -1552,7 +1552,7 @@ SELECT cast(0 as float) = false FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"FLOAT\"", @@ -1574,7 +1574,7 @@ SELECT cast(0 as double) = false FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DOUBLE\"", @@ -1596,7 +1596,7 @@ SELECT cast(0 as decimal(10, 0)) = false FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(10,0)\"", @@ -1629,7 +1629,7 @@ SELECT cast('0' as binary) = false FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BINARY\"", @@ -1662,7 +1662,7 @@ SELECT cast('2017-12-11 09:30:00.0' as timestamp) = false FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP\"", @@ -1684,7 +1684,7 @@ SELECT cast('2017-12-11 09:30:00' as date) = false FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DATE\"", @@ -1706,7 +1706,7 @@ SELECT cast(0 as tinyint) <=> false FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TINYINT\"", @@ -1728,7 +1728,7 @@ SELECT cast(0 as smallint) <=> false FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"SMALLINT\"", @@ -1750,7 +1750,7 @@ SELECT cast(0 as int) <=> false FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"INT\"", @@ -1772,7 +1772,7 @@ SELECT cast(0 as bigint) <=> false FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BIGINT\"", @@ -1794,7 +1794,7 @@ SELECT cast(0 as float) <=> false FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"FLOAT\"", @@ -1816,7 +1816,7 @@ SELECT cast(0 as double) <=> false FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DOUBLE\"", @@ -1838,7 +1838,7 @@ SELECT cast(0 as decimal(10, 0)) <=> false FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(10,0)\"", @@ -1871,7 +1871,7 @@ SELECT cast('0' as binary) <=> false FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BINARY\"", @@ -1904,7 +1904,7 @@ SELECT cast('2017-12-11 09:30:00.0' as timestamp) <=> false FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP\"", @@ -1926,7 +1926,7 @@ SELECT cast('2017-12-11 09:30:00' as date) <=> false FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DATE\"", diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/typeCoercion/native/caseWhenCoercion.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/typeCoercion/native/caseWhenCoercion.sql.out index 45ce3bdd96d9..c0e20572f5aa 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/typeCoercion/native/caseWhenCoercion.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/typeCoercion/native/caseWhenCoercion.sql.out @@ -100,7 +100,7 @@ SELECT CASE WHEN true THEN cast(1 as tinyint) ELSE cast('2' as binary) END FROM -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"TINYINT\", \"BINARY\"]", @@ -122,7 +122,7 @@ SELECT CASE WHEN true THEN cast(1 as tinyint) ELSE cast(2 as boolean) END FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"TINYINT\", \"BOOLEAN\"]", @@ -144,7 +144,7 @@ SELECT CASE WHEN true THEN cast(1 as tinyint) ELSE cast('2017-12-11 09:30:00.0' -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"TINYINT\", \"TIMESTAMP\"]", @@ -166,7 +166,7 @@ SELECT CASE WHEN true THEN cast(1 as tinyint) ELSE cast('2017-12-11 09:30:00' as -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"TINYINT\", \"DATE\"]", @@ -276,7 +276,7 @@ SELECT CASE WHEN true THEN cast(1 as smallint) ELSE cast('2' as binary) END FROM -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"SMALLINT\", \"BINARY\"]", @@ -298,7 +298,7 @@ SELECT CASE WHEN true THEN cast(1 as smallint) ELSE cast(2 as boolean) END FROM -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"SMALLINT\", \"BOOLEAN\"]", @@ -320,7 +320,7 @@ SELECT CASE WHEN true THEN cast(1 as smallint) ELSE cast('2017-12-11 09:30:00.0' -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"SMALLINT\", \"TIMESTAMP\"]", @@ -342,7 +342,7 @@ SELECT CASE WHEN true THEN cast(1 as smallint) ELSE cast('2017-12-11 09:30:00' a -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"SMALLINT\", \"DATE\"]", @@ -452,7 +452,7 @@ SELECT CASE WHEN true THEN cast(1 as int) ELSE cast('2' as binary) END FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"INT\", \"BINARY\"]", @@ -474,7 +474,7 @@ SELECT CASE WHEN true THEN cast(1 as int) ELSE cast(2 as boolean) END FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"INT\", \"BOOLEAN\"]", @@ -496,7 +496,7 @@ SELECT CASE WHEN true THEN cast(1 as int) ELSE cast('2017-12-11 09:30:00.0' as t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"INT\", \"TIMESTAMP\"]", @@ -518,7 +518,7 @@ SELECT CASE WHEN true THEN cast(1 as int) ELSE cast('2017-12-11 09:30:00' as dat -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"INT\", \"DATE\"]", @@ -628,7 +628,7 @@ SELECT CASE WHEN true THEN cast(1 as bigint) ELSE cast('2' as binary) END FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BIGINT\", \"BINARY\"]", @@ -650,7 +650,7 @@ SELECT CASE WHEN true THEN cast(1 as bigint) ELSE cast(2 as boolean) END FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BIGINT\", \"BOOLEAN\"]", @@ -672,7 +672,7 @@ SELECT CASE WHEN true THEN cast(1 as bigint) ELSE cast('2017-12-11 09:30:00.0' a -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BIGINT\", \"TIMESTAMP\"]", @@ -694,7 +694,7 @@ SELECT CASE WHEN true THEN cast(1 as bigint) ELSE cast('2017-12-11 09:30:00' as -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BIGINT\", \"DATE\"]", @@ -804,7 +804,7 @@ SELECT CASE WHEN true THEN cast(1 as float) ELSE cast('2' as binary) END FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"FLOAT\", \"BINARY\"]", @@ -826,7 +826,7 @@ SELECT CASE WHEN true THEN cast(1 as float) ELSE cast(2 as boolean) END FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"FLOAT\", \"BOOLEAN\"]", @@ -848,7 +848,7 @@ SELECT CASE WHEN true THEN cast(1 as float) ELSE cast('2017-12-11 09:30:00.0' as -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"FLOAT\", \"TIMESTAMP\"]", @@ -870,7 +870,7 @@ SELECT CASE WHEN true THEN cast(1 as float) ELSE cast('2017-12-11 09:30:00' as d -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"FLOAT\", \"DATE\"]", @@ -980,7 +980,7 @@ SELECT CASE WHEN true THEN cast(1 as double) ELSE cast('2' as binary) END FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DOUBLE\", \"BINARY\"]", @@ -1002,7 +1002,7 @@ SELECT CASE WHEN true THEN cast(1 as double) ELSE cast(2 as boolean) END FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DOUBLE\", \"BOOLEAN\"]", @@ -1024,7 +1024,7 @@ SELECT CASE WHEN true THEN cast(1 as double) ELSE cast('2017-12-11 09:30:00.0' a -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DOUBLE\", \"TIMESTAMP\"]", @@ -1046,7 +1046,7 @@ SELECT CASE WHEN true THEN cast(1 as double) ELSE cast('2017-12-11 09:30:00' as -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DOUBLE\", \"DATE\"]", @@ -1156,7 +1156,7 @@ SELECT CASE WHEN true THEN cast(1 as decimal(10, 0)) ELSE cast('2' as binary) EN -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DECIMAL(10,0)\", \"BINARY\"]", @@ -1178,7 +1178,7 @@ SELECT CASE WHEN true THEN cast(1 as decimal(10, 0)) ELSE cast(2 as boolean) END -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DECIMAL(10,0)\", \"BOOLEAN\"]", @@ -1200,7 +1200,7 @@ SELECT CASE WHEN true THEN cast(1 as decimal(10, 0)) ELSE cast('2017-12-11 09:30 -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DECIMAL(10,0)\", \"TIMESTAMP\"]", @@ -1222,7 +1222,7 @@ SELECT CASE WHEN true THEN cast(1 as decimal(10, 0)) ELSE cast('2017-12-11 09:30 -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DECIMAL(10,0)\", \"DATE\"]", @@ -1376,7 +1376,7 @@ SELECT CASE WHEN true THEN cast('1' as binary) ELSE cast(2 as tinyint) END FROM -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BINARY\", \"TINYINT\"]", @@ -1398,7 +1398,7 @@ SELECT CASE WHEN true THEN cast('1' as binary) ELSE cast(2 as smallint) END FROM -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BINARY\", \"SMALLINT\"]", @@ -1420,7 +1420,7 @@ SELECT CASE WHEN true THEN cast('1' as binary) ELSE cast(2 as int) END FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BINARY\", \"INT\"]", @@ -1442,7 +1442,7 @@ SELECT CASE WHEN true THEN cast('1' as binary) ELSE cast(2 as bigint) END FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BINARY\", \"BIGINT\"]", @@ -1464,7 +1464,7 @@ SELECT CASE WHEN true THEN cast('1' as binary) ELSE cast(2 as float) END FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BINARY\", \"FLOAT\"]", @@ -1486,7 +1486,7 @@ SELECT CASE WHEN true THEN cast('1' as binary) ELSE cast(2 as double) END FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BINARY\", \"DOUBLE\"]", @@ -1508,7 +1508,7 @@ SELECT CASE WHEN true THEN cast('1' as binary) ELSE cast(2 as decimal(10, 0)) EN -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BINARY\", \"DECIMAL(10,0)\"]", @@ -1552,7 +1552,7 @@ SELECT CASE WHEN true THEN cast('1' as binary) ELSE cast(2 as boolean) END FROM -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BINARY\", \"BOOLEAN\"]", @@ -1574,7 +1574,7 @@ SELECT CASE WHEN true THEN cast('1' as binary) ELSE cast('2017-12-11 09:30:00.0' -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BINARY\", \"TIMESTAMP\"]", @@ -1596,7 +1596,7 @@ SELECT CASE WHEN true THEN cast('1' as binary) ELSE cast('2017-12-11 09:30:00' a -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BINARY\", \"DATE\"]", @@ -1618,7 +1618,7 @@ SELECT CASE WHEN true THEN cast(1 as boolean) ELSE cast(2 as tinyint) END FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BOOLEAN\", \"TINYINT\"]", @@ -1640,7 +1640,7 @@ SELECT CASE WHEN true THEN cast(1 as boolean) ELSE cast(2 as smallint) END FROM -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BOOLEAN\", \"SMALLINT\"]", @@ -1662,7 +1662,7 @@ SELECT CASE WHEN true THEN cast(1 as boolean) ELSE cast(2 as int) END FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BOOLEAN\", \"INT\"]", @@ -1684,7 +1684,7 @@ SELECT CASE WHEN true THEN cast(1 as boolean) ELSE cast(2 as bigint) END FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BOOLEAN\", \"BIGINT\"]", @@ -1706,7 +1706,7 @@ SELECT CASE WHEN true THEN cast(1 as boolean) ELSE cast(2 as float) END FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BOOLEAN\", \"FLOAT\"]", @@ -1728,7 +1728,7 @@ SELECT CASE WHEN true THEN cast(1 as boolean) ELSE cast(2 as double) END FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BOOLEAN\", \"DOUBLE\"]", @@ -1750,7 +1750,7 @@ SELECT CASE WHEN true THEN cast(1 as boolean) ELSE cast(2 as decimal(10, 0)) END -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BOOLEAN\", \"DECIMAL(10,0)\"]", @@ -1783,7 +1783,7 @@ SELECT CASE WHEN true THEN cast(1 as boolean) ELSE cast('2' as binary) END FROM -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BOOLEAN\", \"BINARY\"]", @@ -1816,7 +1816,7 @@ SELECT CASE WHEN true THEN cast(1 as boolean) ELSE cast('2017-12-11 09:30:00.0' -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BOOLEAN\", \"TIMESTAMP\"]", @@ -1838,7 +1838,7 @@ SELECT CASE WHEN true THEN cast(1 as boolean) ELSE cast('2017-12-11 09:30:00' as -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BOOLEAN\", \"DATE\"]", @@ -1860,7 +1860,7 @@ SELECT CASE WHEN true THEN cast('2017-12-12 09:30:00.0' as timestamp) ELSE cast( -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"TIMESTAMP\", \"TINYINT\"]", @@ -1882,7 +1882,7 @@ SELECT CASE WHEN true THEN cast('2017-12-12 09:30:00.0' as timestamp) ELSE cast( -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"TIMESTAMP\", \"SMALLINT\"]", @@ -1904,7 +1904,7 @@ SELECT CASE WHEN true THEN cast('2017-12-12 09:30:00.0' as timestamp) ELSE cast( -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"TIMESTAMP\", \"INT\"]", @@ -1926,7 +1926,7 @@ SELECT CASE WHEN true THEN cast('2017-12-12 09:30:00.0' as timestamp) ELSE cast( -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"TIMESTAMP\", \"BIGINT\"]", @@ -1948,7 +1948,7 @@ SELECT CASE WHEN true THEN cast('2017-12-12 09:30:00.0' as timestamp) ELSE cast( -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"TIMESTAMP\", \"FLOAT\"]", @@ -1970,7 +1970,7 @@ SELECT CASE WHEN true THEN cast('2017-12-12 09:30:00.0' as timestamp) ELSE cast( -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"TIMESTAMP\", \"DOUBLE\"]", @@ -1992,7 +1992,7 @@ SELECT CASE WHEN true THEN cast('2017-12-12 09:30:00.0' as timestamp) ELSE cast( -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"TIMESTAMP\", \"DECIMAL(10,0)\"]", @@ -2025,7 +2025,7 @@ SELECT CASE WHEN true THEN cast('2017-12-12 09:30:00.0' as timestamp) ELSE cast( -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"TIMESTAMP\", \"BINARY\"]", @@ -2047,7 +2047,7 @@ SELECT CASE WHEN true THEN cast('2017-12-12 09:30:00.0' as timestamp) ELSE cast( -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"TIMESTAMP\", \"BOOLEAN\"]", @@ -2091,7 +2091,7 @@ SELECT CASE WHEN true THEN cast('2017-12-12 09:30:00' as date) ELSE cast(2 as ti -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DATE\", \"TINYINT\"]", @@ -2113,7 +2113,7 @@ SELECT CASE WHEN true THEN cast('2017-12-12 09:30:00' as date) ELSE cast(2 as sm -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DATE\", \"SMALLINT\"]", @@ -2135,7 +2135,7 @@ SELECT CASE WHEN true THEN cast('2017-12-12 09:30:00' as date) ELSE cast(2 as in -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DATE\", \"INT\"]", @@ -2157,7 +2157,7 @@ SELECT CASE WHEN true THEN cast('2017-12-12 09:30:00' as date) ELSE cast(2 as bi -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DATE\", \"BIGINT\"]", @@ -2179,7 +2179,7 @@ SELECT CASE WHEN true THEN cast('2017-12-12 09:30:00' as date) ELSE cast(2 as fl -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DATE\", \"FLOAT\"]", @@ -2201,7 +2201,7 @@ SELECT CASE WHEN true THEN cast('2017-12-12 09:30:00' as date) ELSE cast(2 as do -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DATE\", \"DOUBLE\"]", @@ -2223,7 +2223,7 @@ SELECT CASE WHEN true THEN cast('2017-12-12 09:30:00' as date) ELSE cast(2 as de -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DATE\", \"DECIMAL(10,0)\"]", @@ -2256,7 +2256,7 @@ SELECT CASE WHEN true THEN cast('2017-12-12 09:30:00' as date) ELSE cast('2' as -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DATE\", \"BINARY\"]", @@ -2278,7 +2278,7 @@ SELECT CASE WHEN true THEN cast('2017-12-12 09:30:00' as date) ELSE cast(2 as bo -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DATE\", \"BOOLEAN\"]", diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/typeCoercion/native/dateTimeOperations.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/typeCoercion/native/dateTimeOperations.sql.out index c2787e581688..d15ea0529990 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/typeCoercion/native/dateTimeOperations.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/typeCoercion/native/dateTimeOperations.sql.out @@ -12,7 +12,7 @@ select cast(1 as tinyint) + interval 2 day -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"CAST(1 AS TINYINT)\"", @@ -36,7 +36,7 @@ select cast(1 as smallint) + interval 2 day -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"CAST(1 AS SMALLINT)\"", @@ -60,7 +60,7 @@ select cast(1 as int) + interval 2 day -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"CAST(1 AS INT)\"", @@ -84,7 +84,7 @@ select cast(1 as bigint) + interval 2 day -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"CAST(1 AS BIGINT)\"", @@ -108,7 +108,7 @@ select cast(1 as float) + interval 2 day -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"CAST(1 AS FLOAT)\"", @@ -132,7 +132,7 @@ select cast(1 as double) + interval 2 day -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"CAST(1 AS DOUBLE)\"", @@ -156,7 +156,7 @@ select cast(1 as decimal(10, 0)) + interval 2 day -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"CAST(1 AS DECIMAL(10,0))\"", @@ -194,7 +194,7 @@ select cast('1' as binary) + interval 2 day -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"CAST(1 AS BINARY)\"", @@ -218,7 +218,7 @@ select cast(1 as boolean) + interval 2 day -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"CAST(1 AS BOOLEAN)\"", @@ -256,7 +256,7 @@ select interval 2 day + cast(1 as tinyint) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"CAST(1 AS TINYINT)\"", @@ -280,7 +280,7 @@ select interval 2 day + cast(1 as smallint) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"CAST(1 AS SMALLINT)\"", @@ -304,7 +304,7 @@ select interval 2 day + cast(1 as int) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"CAST(1 AS INT)\"", @@ -328,7 +328,7 @@ select interval 2 day + cast(1 as bigint) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"CAST(1 AS BIGINT)\"", @@ -352,7 +352,7 @@ select interval 2 day + cast(1 as float) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"CAST(1 AS FLOAT)\"", @@ -376,7 +376,7 @@ select interval 2 day + cast(1 as double) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"CAST(1 AS DOUBLE)\"", @@ -400,7 +400,7 @@ select interval 2 day + cast(1 as decimal(10, 0)) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"CAST(1 AS DECIMAL(10,0))\"", @@ -438,7 +438,7 @@ select interval 2 day + cast('1' as binary) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"CAST(1 AS BINARY)\"", @@ -462,7 +462,7 @@ select interval 2 day + cast(1 as boolean) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"CAST(1 AS BOOLEAN)\"", @@ -500,7 +500,7 @@ select cast(1 as tinyint) - interval 2 day -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"CAST(1 AS TINYINT)\"", @@ -524,7 +524,7 @@ select cast(1 as smallint) - interval 2 day -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"CAST(1 AS SMALLINT)\"", @@ -548,7 +548,7 @@ select cast(1 as int) - interval 2 day -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"CAST(1 AS INT)\"", @@ -572,7 +572,7 @@ select cast(1 as bigint) - interval 2 day -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"CAST(1 AS BIGINT)\"", @@ -596,7 +596,7 @@ select cast(1 as float) - interval 2 day -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"CAST(1 AS FLOAT)\"", @@ -620,7 +620,7 @@ select cast(1 as double) - interval 2 day -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"CAST(1 AS DOUBLE)\"", @@ -644,7 +644,7 @@ select cast(1 as decimal(10, 0)) - interval 2 day -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"CAST(1 AS DECIMAL(10,0))\"", @@ -682,7 +682,7 @@ select cast('1' as binary) - interval 2 day -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"CAST(1 AS BINARY)\"", @@ -706,7 +706,7 @@ select cast(1 as boolean) - interval 2 day -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"CAST(1 AS BOOLEAN)\"", diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/typeCoercion/native/decimalPrecision.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/typeCoercion/native/decimalPrecision.sql.out index 4458e15e53cf..78e72ae58586 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/typeCoercion/native/decimalPrecision.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/typeCoercion/native/decimalPrecision.sql.out @@ -320,7 +320,7 @@ SELECT cast('1' as binary) + cast(1 as decimal(3, 0)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BINARY\"", @@ -342,7 +342,7 @@ SELECT cast('1' as binary) + cast(1 as decimal(5, 0)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BINARY\"", @@ -364,7 +364,7 @@ SELECT cast('1' as binary) + cast(1 as decimal(10, 0)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BINARY\"", @@ -386,7 +386,7 @@ SELECT cast('1' as binary) + cast(1 as decimal(20, 0)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BINARY\"", @@ -408,7 +408,7 @@ SELECT cast('2017-12-11 09:30:00.0' as timestamp) + cast(1 as decimal(3, 0)) FRO -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP\"", @@ -430,7 +430,7 @@ SELECT cast('2017-12-11 09:30:00.0' as timestamp) + cast(1 as decimal(5, 0)) FRO -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP\"", @@ -452,7 +452,7 @@ SELECT cast('2017-12-11 09:30:00.0' as timestamp) + cast(1 as decimal(10, 0)) FR -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP\"", @@ -474,7 +474,7 @@ SELECT cast('2017-12-11 09:30:00.0' as timestamp) + cast(1 as decimal(20, 0)) FR -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP\"", @@ -496,7 +496,7 @@ SELECT cast('2017-12-11 09:30:00' as date) + cast(1 as decimal(3, 0)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"CAST(1 AS DECIMAL(3,0))\"", @@ -520,7 +520,7 @@ SELECT cast('2017-12-11 09:30:00' as date) + cast(1 as decimal(5, 0)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"CAST(1 AS DECIMAL(5,0))\"", @@ -544,7 +544,7 @@ SELECT cast('2017-12-11 09:30:00' as date) + cast(1 as decimal(10, 0)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"CAST(1 AS DECIMAL(10,0))\"", @@ -568,7 +568,7 @@ SELECT cast('2017-12-11 09:30:00' as date) + cast(1 as decimal(20, 0)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"CAST(1 AS DECIMAL(20,0))\"", @@ -944,7 +944,7 @@ SELECT cast(1 as decimal(3, 0)) + cast('1' as binary) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(3,0)\"", @@ -966,7 +966,7 @@ SELECT cast(1 as decimal(5, 0)) + cast('1' as binary) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(5,0)\"", @@ -988,7 +988,7 @@ SELECT cast(1 as decimal(10, 0)) + cast('1' as binary) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(10,0)\"", @@ -1010,7 +1010,7 @@ SELECT cast(1 as decimal(20, 0)) + cast('1' as binary) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(20,0)\"", @@ -1032,7 +1032,7 @@ SELECT cast(1 as decimal(3, 0)) + cast(1 as boolean) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(3,0)\"", @@ -1054,7 +1054,7 @@ SELECT cast(1 as decimal(5, 0)) + cast(1 as boolean) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(5,0)\"", @@ -1076,7 +1076,7 @@ SELECT cast(1 as decimal(10, 0)) + cast(1 as boolean) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(10,0)\"", @@ -1098,7 +1098,7 @@ SELECT cast(1 as decimal(20, 0)) + cast(1 as boolean) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(20,0)\"", @@ -1120,7 +1120,7 @@ SELECT cast(1 as decimal(3, 0)) + cast('2017-12-11 09:30:00.0' as timestamp) FR -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(3,0)\"", @@ -1142,7 +1142,7 @@ SELECT cast(1 as decimal(5, 0)) + cast('2017-12-11 09:30:00.0' as timestamp) FR -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(5,0)\"", @@ -1164,7 +1164,7 @@ SELECT cast(1 as decimal(10, 0)) + cast('2017-12-11 09:30:00.0' as timestamp) FR -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(10,0)\"", @@ -1186,7 +1186,7 @@ SELECT cast(1 as decimal(20, 0)) + cast('2017-12-11 09:30:00.0' as timestamp) FR -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(20,0)\"", @@ -1208,7 +1208,7 @@ SELECT cast(1 as decimal(3, 0)) + cast('2017-12-11 09:30:00' as date) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"CAST(1 AS DECIMAL(3,0))\"", @@ -1232,7 +1232,7 @@ SELECT cast(1 as decimal(5, 0)) + cast('2017-12-11 09:30:00' as date) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"CAST(1 AS DECIMAL(5,0))\"", @@ -1256,7 +1256,7 @@ SELECT cast(1 as decimal(10, 0)) + cast('2017-12-11 09:30:00' as date) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"CAST(1 AS DECIMAL(10,0))\"", @@ -1280,7 +1280,7 @@ SELECT cast(1 as decimal(20, 0)) + cast('2017-12-11 09:30:00' as date) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"CAST(1 AS DECIMAL(20,0))\"", @@ -1612,7 +1612,7 @@ SELECT cast('1' as binary) - cast(1 as decimal(3, 0)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BINARY\"", @@ -1634,7 +1634,7 @@ SELECT cast('1' as binary) - cast(1 as decimal(5, 0)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BINARY\"", @@ -1656,7 +1656,7 @@ SELECT cast('1' as binary) - cast(1 as decimal(10, 0)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BINARY\"", @@ -1678,7 +1678,7 @@ SELECT cast('1' as binary) - cast(1 as decimal(20, 0)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BINARY\"", @@ -1700,7 +1700,7 @@ SELECT cast('2017-12-11 09:30:00.0' as timestamp) - cast(1 as decimal(3, 0)) FRO -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"CAST(1 AS DECIMAL(3,0))\"", @@ -1724,7 +1724,7 @@ SELECT cast('2017-12-11 09:30:00.0' as timestamp) - cast(1 as decimal(5, 0)) FRO -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"CAST(1 AS DECIMAL(5,0))\"", @@ -1748,7 +1748,7 @@ SELECT cast('2017-12-11 09:30:00.0' as timestamp) - cast(1 as decimal(10, 0)) FR -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"CAST(1 AS DECIMAL(10,0))\"", @@ -1772,7 +1772,7 @@ SELECT cast('2017-12-11 09:30:00.0' as timestamp) - cast(1 as decimal(20, 0)) FR -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"CAST(1 AS DECIMAL(20,0))\"", @@ -1796,7 +1796,7 @@ SELECT cast('2017-12-11 09:30:00' as date) - cast(1 as decimal(3, 0)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"CAST(1 AS DECIMAL(3,0))\"", @@ -1820,7 +1820,7 @@ SELECT cast('2017-12-11 09:30:00' as date) - cast(1 as decimal(5, 0)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"CAST(1 AS DECIMAL(5,0))\"", @@ -1844,7 +1844,7 @@ SELECT cast('2017-12-11 09:30:00' as date) - cast(1 as decimal(10, 0)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"CAST(1 AS DECIMAL(10,0))\"", @@ -1868,7 +1868,7 @@ SELECT cast('2017-12-11 09:30:00' as date) - cast(1 as decimal(20, 0)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"CAST(1 AS DECIMAL(20,0))\"", @@ -2244,7 +2244,7 @@ SELECT cast(1 as decimal(3, 0)) - cast('1' as binary) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(3,0)\"", @@ -2266,7 +2266,7 @@ SELECT cast(1 as decimal(5, 0)) - cast('1' as binary) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(5,0)\"", @@ -2288,7 +2288,7 @@ SELECT cast(1 as decimal(10, 0)) - cast('1' as binary) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(10,0)\"", @@ -2310,7 +2310,7 @@ SELECT cast(1 as decimal(20, 0)) - cast('1' as binary) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(20,0)\"", @@ -2332,7 +2332,7 @@ SELECT cast(1 as decimal(3, 0)) - cast(1 as boolean) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(3,0)\"", @@ -2354,7 +2354,7 @@ SELECT cast(1 as decimal(5, 0)) - cast(1 as boolean) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(5,0)\"", @@ -2376,7 +2376,7 @@ SELECT cast(1 as decimal(10, 0)) - cast(1 as boolean) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(10,0)\"", @@ -2398,7 +2398,7 @@ SELECT cast(1 as decimal(20, 0)) - cast(1 as boolean) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(20,0)\"", @@ -2420,7 +2420,7 @@ SELECT cast(1 as decimal(3, 0)) - cast('2017-12-11 09:30:00.0' as timestamp) FR -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"CAST(1 AS DECIMAL(3,0))\"", @@ -2444,7 +2444,7 @@ SELECT cast(1 as decimal(5, 0)) - cast('2017-12-11 09:30:00.0' as timestamp) FR -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"CAST(1 AS DECIMAL(5,0))\"", @@ -2468,7 +2468,7 @@ SELECT cast(1 as decimal(10, 0)) - cast('2017-12-11 09:30:00.0' as timestamp) FR -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"CAST(1 AS DECIMAL(10,0))\"", @@ -2492,7 +2492,7 @@ SELECT cast(1 as decimal(20, 0)) - cast('2017-12-11 09:30:00.0' as timestamp) FR -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"CAST(1 AS DECIMAL(20,0))\"", @@ -2516,7 +2516,7 @@ SELECT cast(1 as decimal(3, 0)) - cast('2017-12-11 09:30:00' as date) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"CAST(1 AS DECIMAL(3,0))\"", @@ -2540,7 +2540,7 @@ SELECT cast(1 as decimal(5, 0)) - cast('2017-12-11 09:30:00' as date) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"CAST(1 AS DECIMAL(5,0))\"", @@ -2564,7 +2564,7 @@ SELECT cast(1 as decimal(10, 0)) - cast('2017-12-11 09:30:00' as date) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"CAST(1 AS DECIMAL(10,0))\"", @@ -2588,7 +2588,7 @@ SELECT cast(1 as decimal(20, 0)) - cast('2017-12-11 09:30:00' as date) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"CAST(1 AS DECIMAL(20,0))\"", @@ -2920,7 +2920,7 @@ SELECT cast('1' as binary) * cast(1 as decimal(3, 0)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BINARY\"", @@ -2942,7 +2942,7 @@ SELECT cast('1' as binary) * cast(1 as decimal(5, 0)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BINARY\"", @@ -2964,7 +2964,7 @@ SELECT cast('1' as binary) * cast(1 as decimal(10, 0)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BINARY\"", @@ -2986,7 +2986,7 @@ SELECT cast('1' as binary) * cast(1 as decimal(20, 0)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BINARY\"", @@ -3008,7 +3008,7 @@ SELECT cast('2017*12*11 09:30:00.0' as timestamp) * cast(1 as decimal(3, 0)) FRO -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP\"", @@ -3030,7 +3030,7 @@ SELECT cast('2017*12*11 09:30:00.0' as timestamp) * cast(1 as decimal(5, 0)) FRO -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP\"", @@ -3052,7 +3052,7 @@ SELECT cast('2017*12*11 09:30:00.0' as timestamp) * cast(1 as decimal(10, 0)) FR -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP\"", @@ -3074,7 +3074,7 @@ SELECT cast('2017*12*11 09:30:00.0' as timestamp) * cast(1 as decimal(20, 0)) FR -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP\"", @@ -3096,7 +3096,7 @@ SELECT cast('2017*12*11 09:30:00' as date) * cast(1 as decimal(3, 0)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DATE\"", @@ -3118,7 +3118,7 @@ SELECT cast('2017*12*11 09:30:00' as date) * cast(1 as decimal(5, 0)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DATE\"", @@ -3140,7 +3140,7 @@ SELECT cast('2017*12*11 09:30:00' as date) * cast(1 as decimal(10, 0)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DATE\"", @@ -3162,7 +3162,7 @@ SELECT cast('2017*12*11 09:30:00' as date) * cast(1 as decimal(20, 0)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DATE\"", @@ -3536,7 +3536,7 @@ SELECT cast(1 as decimal(3, 0)) * cast('1' as binary) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(3,0)\"", @@ -3558,7 +3558,7 @@ SELECT cast(1 as decimal(5, 0)) * cast('1' as binary) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(5,0)\"", @@ -3580,7 +3580,7 @@ SELECT cast(1 as decimal(10, 0)) * cast('1' as binary) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(10,0)\"", @@ -3602,7 +3602,7 @@ SELECT cast(1 as decimal(20, 0)) * cast('1' as binary) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(20,0)\"", @@ -3624,7 +3624,7 @@ SELECT cast(1 as decimal(3, 0)) * cast(1 as boolean) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(3,0)\"", @@ -3646,7 +3646,7 @@ SELECT cast(1 as decimal(5, 0)) * cast(1 as boolean) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(5,0)\"", @@ -3668,7 +3668,7 @@ SELECT cast(1 as decimal(10, 0)) * cast(1 as boolean) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(10,0)\"", @@ -3690,7 +3690,7 @@ SELECT cast(1 as decimal(20, 0)) * cast(1 as boolean) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(20,0)\"", @@ -3712,7 +3712,7 @@ SELECT cast(1 as decimal(3, 0)) * cast('2017*12*11 09:30:00.0' as timestamp) FR -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(3,0)\"", @@ -3734,7 +3734,7 @@ SELECT cast(1 as decimal(5, 0)) * cast('2017*12*11 09:30:00.0' as timestamp) FR -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(5,0)\"", @@ -3756,7 +3756,7 @@ SELECT cast(1 as decimal(10, 0)) * cast('2017*12*11 09:30:00.0' as timestamp) FR -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(10,0)\"", @@ -3778,7 +3778,7 @@ SELECT cast(1 as decimal(20, 0)) * cast('2017*12*11 09:30:00.0' as timestamp) FR -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(20,0)\"", @@ -3800,7 +3800,7 @@ SELECT cast(1 as decimal(3, 0)) * cast('2017*12*11 09:30:00' as date) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(3,0)\"", @@ -3822,7 +3822,7 @@ SELECT cast(1 as decimal(5, 0)) * cast('2017*12*11 09:30:00' as date) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(5,0)\"", @@ -3844,7 +3844,7 @@ SELECT cast(1 as decimal(10, 0)) * cast('2017*12*11 09:30:00' as date) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(10,0)\"", @@ -3866,7 +3866,7 @@ SELECT cast(1 as decimal(20, 0)) * cast('2017*12*11 09:30:00' as date) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(20,0)\"", @@ -4196,7 +4196,7 @@ SELECT cast('1' as binary) / cast(1 as decimal(3, 0)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BINARY\"", @@ -4218,7 +4218,7 @@ SELECT cast('1' as binary) / cast(1 as decimal(5, 0)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BINARY\"", @@ -4240,7 +4240,7 @@ SELECT cast('1' as binary) / cast(1 as decimal(10, 0)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BINARY\"", @@ -4262,7 +4262,7 @@ SELECT cast('1' as binary) / cast(1 as decimal(20, 0)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BINARY\"", @@ -4284,7 +4284,7 @@ SELECT cast('2017/12/11 09:30:00.0' as timestamp) / cast(1 as decimal(3, 0)) FRO -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP\"", @@ -4306,7 +4306,7 @@ SELECT cast('2017/12/11 09:30:00.0' as timestamp) / cast(1 as decimal(5, 0)) FRO -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP\"", @@ -4328,7 +4328,7 @@ SELECT cast('2017/12/11 09:30:00.0' as timestamp) / cast(1 as decimal(10, 0)) FR -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP\"", @@ -4350,7 +4350,7 @@ SELECT cast('2017/12/11 09:30:00.0' as timestamp) / cast(1 as decimal(20, 0)) FR -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP\"", @@ -4372,7 +4372,7 @@ SELECT cast('2017/12/11 09:30:00' as date) / cast(1 as decimal(3, 0)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DATE\"", @@ -4394,7 +4394,7 @@ SELECT cast('2017/12/11 09:30:00' as date) / cast(1 as decimal(5, 0)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DATE\"", @@ -4416,7 +4416,7 @@ SELECT cast('2017/12/11 09:30:00' as date) / cast(1 as decimal(10, 0)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DATE\"", @@ -4438,7 +4438,7 @@ SELECT cast('2017/12/11 09:30:00' as date) / cast(1 as decimal(20, 0)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DATE\"", @@ -4812,7 +4812,7 @@ SELECT cast(1 as decimal(3, 0)) / cast('1' as binary) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(3,0)\"", @@ -4834,7 +4834,7 @@ SELECT cast(1 as decimal(5, 0)) / cast('1' as binary) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(5,0)\"", @@ -4856,7 +4856,7 @@ SELECT cast(1 as decimal(10, 0)) / cast('1' as binary) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(10,0)\"", @@ -4878,7 +4878,7 @@ SELECT cast(1 as decimal(20, 0)) / cast('1' as binary) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(20,0)\"", @@ -4900,7 +4900,7 @@ SELECT cast(1 as decimal(3, 0)) / cast(1 as boolean) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(3,0)\"", @@ -4922,7 +4922,7 @@ SELECT cast(1 as decimal(5, 0)) / cast(1 as boolean) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(5,0)\"", @@ -4944,7 +4944,7 @@ SELECT cast(1 as decimal(10, 0)) / cast(1 as boolean) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(10,0)\"", @@ -4966,7 +4966,7 @@ SELECT cast(1 as decimal(20, 0)) / cast(1 as boolean) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(20,0)\"", @@ -4988,7 +4988,7 @@ SELECT cast(1 as decimal(3, 0)) / cast('2017/12/11 09:30:00.0' as timestamp) FR -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(3,0)\"", @@ -5010,7 +5010,7 @@ SELECT cast(1 as decimal(5, 0)) / cast('2017/12/11 09:30:00.0' as timestamp) FR -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(5,0)\"", @@ -5032,7 +5032,7 @@ SELECT cast(1 as decimal(10, 0)) / cast('2017/12/11 09:30:00.0' as timestamp) FR -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(10,0)\"", @@ -5054,7 +5054,7 @@ SELECT cast(1 as decimal(20, 0)) / cast('2017/12/11 09:30:00.0' as timestamp) FR -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(20,0)\"", @@ -5076,7 +5076,7 @@ SELECT cast(1 as decimal(3, 0)) / cast('2017/12/11 09:30:00' as date) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(3,0)\"", @@ -5098,7 +5098,7 @@ SELECT cast(1 as decimal(5, 0)) / cast('2017/12/11 09:30:00' as date) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(5,0)\"", @@ -5120,7 +5120,7 @@ SELECT cast(1 as decimal(10, 0)) / cast('2017/12/11 09:30:00' as date) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(10,0)\"", @@ -5142,7 +5142,7 @@ SELECT cast(1 as decimal(20, 0)) / cast('2017/12/11 09:30:00' as date) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(20,0)\"", @@ -5472,7 +5472,7 @@ SELECT cast('1' as binary) % cast(1 as decimal(3, 0)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BINARY\"", @@ -5494,7 +5494,7 @@ SELECT cast('1' as binary) % cast(1 as decimal(5, 0)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BINARY\"", @@ -5516,7 +5516,7 @@ SELECT cast('1' as binary) % cast(1 as decimal(10, 0)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BINARY\"", @@ -5538,7 +5538,7 @@ SELECT cast('1' as binary) % cast(1 as decimal(20, 0)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BINARY\"", @@ -5560,7 +5560,7 @@ SELECT cast('2017-12-11 09:30:00.0' as timestamp) % cast(1 as decimal(3, 0)) FRO -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP\"", @@ -5582,7 +5582,7 @@ SELECT cast('2017-12-11 09:30:00.0' as timestamp) % cast(1 as decimal(5, 0)) FRO -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP\"", @@ -5604,7 +5604,7 @@ SELECT cast('2017-12-11 09:30:00.0' as timestamp) % cast(1 as decimal(10, 0)) FR -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP\"", @@ -5626,7 +5626,7 @@ SELECT cast('2017-12-11 09:30:00.0' as timestamp) % cast(1 as decimal(20, 0)) FR -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP\"", @@ -5648,7 +5648,7 @@ SELECT cast('2017-12-11 09:30:00' as date) % cast(1 as decimal(3, 0)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DATE\"", @@ -5670,7 +5670,7 @@ SELECT cast('2017-12-11 09:30:00' as date) % cast(1 as decimal(5, 0)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DATE\"", @@ -5692,7 +5692,7 @@ SELECT cast('2017-12-11 09:30:00' as date) % cast(1 as decimal(10, 0)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DATE\"", @@ -5714,7 +5714,7 @@ SELECT cast('2017-12-11 09:30:00' as date) % cast(1 as decimal(20, 0)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DATE\"", @@ -6088,7 +6088,7 @@ SELECT cast(1 as decimal(3, 0)) % cast('1' as binary) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(3,0)\"", @@ -6110,7 +6110,7 @@ SELECT cast(1 as decimal(5, 0)) % cast('1' as binary) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(5,0)\"", @@ -6132,7 +6132,7 @@ SELECT cast(1 as decimal(10, 0)) % cast('1' as binary) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(10,0)\"", @@ -6154,7 +6154,7 @@ SELECT cast(1 as decimal(20, 0)) % cast('1' as binary) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(20,0)\"", @@ -6176,7 +6176,7 @@ SELECT cast(1 as decimal(3, 0)) % cast(1 as boolean) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(3,0)\"", @@ -6198,7 +6198,7 @@ SELECT cast(1 as decimal(5, 0)) % cast(1 as boolean) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(5,0)\"", @@ -6220,7 +6220,7 @@ SELECT cast(1 as decimal(10, 0)) % cast(1 as boolean) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(10,0)\"", @@ -6242,7 +6242,7 @@ SELECT cast(1 as decimal(20, 0)) % cast(1 as boolean) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(20,0)\"", @@ -6264,7 +6264,7 @@ SELECT cast(1 as decimal(3, 0)) % cast('2017-12-11 09:30:00.0' as timestamp) FR -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(3,0)\"", @@ -6286,7 +6286,7 @@ SELECT cast(1 as decimal(5, 0)) % cast('2017-12-11 09:30:00.0' as timestamp) FR -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(5,0)\"", @@ -6308,7 +6308,7 @@ SELECT cast(1 as decimal(10, 0)) % cast('2017-12-11 09:30:00.0' as timestamp) FR -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(10,0)\"", @@ -6330,7 +6330,7 @@ SELECT cast(1 as decimal(20, 0)) % cast('2017-12-11 09:30:00.0' as timestamp) FR -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(20,0)\"", @@ -6352,7 +6352,7 @@ SELECT cast(1 as decimal(3, 0)) % cast('2017-12-11 09:30:00' as date) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(3,0)\"", @@ -6374,7 +6374,7 @@ SELECT cast(1 as decimal(5, 0)) % cast('2017-12-11 09:30:00' as date) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(5,0)\"", @@ -6396,7 +6396,7 @@ SELECT cast(1 as decimal(10, 0)) % cast('2017-12-11 09:30:00' as date) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(10,0)\"", @@ -6418,7 +6418,7 @@ SELECT cast(1 as decimal(20, 0)) % cast('2017-12-11 09:30:00' as date) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(20,0)\"", @@ -6748,7 +6748,7 @@ SELECT pmod(cast('1' as binary), cast(1 as decimal(3, 0))) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BINARY\"", @@ -6770,7 +6770,7 @@ SELECT pmod(cast('1' as binary), cast(1 as decimal(5, 0))) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BINARY\"", @@ -6792,7 +6792,7 @@ SELECT pmod(cast('1' as binary), cast(1 as decimal(10, 0))) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BINARY\"", @@ -6814,7 +6814,7 @@ SELECT pmod(cast('1' as binary), cast(1 as decimal(20, 0))) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BINARY\"", @@ -6836,7 +6836,7 @@ SELECT pmod(cast('2017-12-11 09:30:00.0' as timestamp), cast(1 as decimal(3, 0)) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP\"", @@ -6858,7 +6858,7 @@ SELECT pmod(cast('2017-12-11 09:30:00.0' as timestamp), cast(1 as decimal(5, 0)) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP\"", @@ -6880,7 +6880,7 @@ SELECT pmod(cast('2017-12-11 09:30:00.0' as timestamp), cast(1 as decimal(10, 0) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP\"", @@ -6902,7 +6902,7 @@ SELECT pmod(cast('2017-12-11 09:30:00.0' as timestamp), cast(1 as decimal(20, 0) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP\"", @@ -6924,7 +6924,7 @@ SELECT pmod(cast('2017-12-11 09:30:00' as date), cast(1 as decimal(3, 0))) FROM -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DATE\"", @@ -6946,7 +6946,7 @@ SELECT pmod(cast('2017-12-11 09:30:00' as date), cast(1 as decimal(5, 0))) FROM -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DATE\"", @@ -6968,7 +6968,7 @@ SELECT pmod(cast('2017-12-11 09:30:00' as date), cast(1 as decimal(10, 0))) FROM -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DATE\"", @@ -6990,7 +6990,7 @@ SELECT pmod(cast('2017-12-11 09:30:00' as date), cast(1 as decimal(20, 0))) FROM -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DATE\"", @@ -7364,7 +7364,7 @@ SELECT pmod(cast(1 as decimal(3, 0)) , cast('1' as binary)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(3,0)\"", @@ -7386,7 +7386,7 @@ SELECT pmod(cast(1 as decimal(5, 0)) , cast('1' as binary)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(5,0)\"", @@ -7408,7 +7408,7 @@ SELECT pmod(cast(1 as decimal(10, 0)), cast('1' as binary)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(10,0)\"", @@ -7430,7 +7430,7 @@ SELECT pmod(cast(1 as decimal(20, 0)), cast('1' as binary)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(20,0)\"", @@ -7452,7 +7452,7 @@ SELECT pmod(cast(1 as decimal(3, 0)) , cast(1 as boolean)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(3,0)\"", @@ -7474,7 +7474,7 @@ SELECT pmod(cast(1 as decimal(5, 0)) , cast(1 as boolean)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(5,0)\"", @@ -7496,7 +7496,7 @@ SELECT pmod(cast(1 as decimal(10, 0)), cast(1 as boolean)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(10,0)\"", @@ -7518,7 +7518,7 @@ SELECT pmod(cast(1 as decimal(20, 0)), cast(1 as boolean)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(20,0)\"", @@ -7540,7 +7540,7 @@ SELECT pmod(cast(1 as decimal(3, 0)) , cast('2017-12-11 09:30:00.0' as timestamp -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(3,0)\"", @@ -7562,7 +7562,7 @@ SELECT pmod(cast(1 as decimal(5, 0)) , cast('2017-12-11 09:30:00.0' as timestamp -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(5,0)\"", @@ -7584,7 +7584,7 @@ SELECT pmod(cast(1 as decimal(10, 0)), cast('2017-12-11 09:30:00.0' as timestamp -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(10,0)\"", @@ -7606,7 +7606,7 @@ SELECT pmod(cast(1 as decimal(20, 0)), cast('2017-12-11 09:30:00.0' as timestamp -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(20,0)\"", @@ -7628,7 +7628,7 @@ SELECT pmod(cast(1 as decimal(3, 0)) , cast('2017-12-11 09:30:00' as date)) FROM -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(3,0)\"", @@ -7650,7 +7650,7 @@ SELECT pmod(cast(1 as decimal(5, 0)) , cast('2017-12-11 09:30:00' as date)) FROM -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(5,0)\"", @@ -7672,7 +7672,7 @@ SELECT pmod(cast(1 as decimal(10, 0)), cast('2017-12-11 09:30:00' as date)) FROM -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(10,0)\"", @@ -7694,7 +7694,7 @@ SELECT pmod(cast(1 as decimal(20, 0)), cast('2017-12-11 09:30:00' as date)) FROM -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(20,0)\"", @@ -8024,7 +8024,7 @@ SELECT cast('1' as binary) = cast(1 as decimal(3, 0)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BINARY\"", @@ -8046,7 +8046,7 @@ SELECT cast('1' as binary) = cast(1 as decimal(5, 0)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BINARY\"", @@ -8068,7 +8068,7 @@ SELECT cast('1' as binary) = cast(1 as decimal(10, 0)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BINARY\"", @@ -8090,7 +8090,7 @@ SELECT cast('1' as binary) = cast(1 as decimal(20, 0)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BINARY\"", @@ -8112,7 +8112,7 @@ SELECT cast('2017-12-11 09:30:00.0' as timestamp) = cast(1 as decimal(3, 0)) FRO -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP\"", @@ -8134,7 +8134,7 @@ SELECT cast('2017-12-11 09:30:00.0' as timestamp) = cast(1 as decimal(5, 0)) FRO -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP\"", @@ -8156,7 +8156,7 @@ SELECT cast('2017-12-11 09:30:00.0' as timestamp) = cast(1 as decimal(10, 0)) FR -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP\"", @@ -8178,7 +8178,7 @@ SELECT cast('2017-12-11 09:30:00.0' as timestamp) = cast(1 as decimal(20, 0)) FR -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP\"", @@ -8200,7 +8200,7 @@ SELECT cast('2017-12-11 09:30:00' as date) = cast(1 as decimal(3, 0)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DATE\"", @@ -8222,7 +8222,7 @@ SELECT cast('2017-12-11 09:30:00' as date) = cast(1 as decimal(5, 0)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DATE\"", @@ -8244,7 +8244,7 @@ SELECT cast('2017-12-11 09:30:00' as date) = cast(1 as decimal(10, 0)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DATE\"", @@ -8266,7 +8266,7 @@ SELECT cast('2017-12-11 09:30:00' as date) = cast(1 as decimal(20, 0)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DATE\"", @@ -8640,7 +8640,7 @@ SELECT cast(1 as decimal(3, 0)) = cast('1' as binary) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(3,0)\"", @@ -8662,7 +8662,7 @@ SELECT cast(1 as decimal(5, 0)) = cast('1' as binary) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(5,0)\"", @@ -8684,7 +8684,7 @@ SELECT cast(1 as decimal(10, 0)) = cast('1' as binary) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(10,0)\"", @@ -8706,7 +8706,7 @@ SELECT cast(1 as decimal(20, 0)) = cast('1' as binary) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(20,0)\"", @@ -8728,7 +8728,7 @@ SELECT cast(1 as decimal(3, 0)) = cast(1 as boolean) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(3,0)\"", @@ -8750,7 +8750,7 @@ SELECT cast(1 as decimal(5, 0)) = cast(1 as boolean) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(5,0)\"", @@ -8772,7 +8772,7 @@ SELECT cast(1 as decimal(10, 0)) = cast(1 as boolean) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(10,0)\"", @@ -8794,7 +8794,7 @@ SELECT cast(1 as decimal(20, 0)) = cast(1 as boolean) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(20,0)\"", @@ -8816,7 +8816,7 @@ SELECT cast(1 as decimal(3, 0)) = cast('2017-12-11 09:30:00.0' as timestamp) FR -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(3,0)\"", @@ -8838,7 +8838,7 @@ SELECT cast(1 as decimal(5, 0)) = cast('2017-12-11 09:30:00.0' as timestamp) FR -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(5,0)\"", @@ -8860,7 +8860,7 @@ SELECT cast(1 as decimal(10, 0)) = cast('2017-12-11 09:30:00.0' as timestamp) FR -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(10,0)\"", @@ -8882,7 +8882,7 @@ SELECT cast(1 as decimal(20, 0)) = cast('2017-12-11 09:30:00.0' as timestamp) FR -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(20,0)\"", @@ -8904,7 +8904,7 @@ SELECT cast(1 as decimal(3, 0)) = cast('2017-12-11 09:30:00' as date) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(3,0)\"", @@ -8926,7 +8926,7 @@ SELECT cast(1 as decimal(5, 0)) = cast('2017-12-11 09:30:00' as date) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(5,0)\"", @@ -8948,7 +8948,7 @@ SELECT cast(1 as decimal(10, 0)) = cast('2017-12-11 09:30:00' as date) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(10,0)\"", @@ -8970,7 +8970,7 @@ SELECT cast(1 as decimal(20, 0)) = cast('2017-12-11 09:30:00' as date) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(20,0)\"", @@ -9300,7 +9300,7 @@ SELECT cast('1' as binary) <=> cast(1 as decimal(3, 0)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BINARY\"", @@ -9322,7 +9322,7 @@ SELECT cast('1' as binary) <=> cast(1 as decimal(5, 0)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BINARY\"", @@ -9344,7 +9344,7 @@ SELECT cast('1' as binary) <=> cast(1 as decimal(10, 0)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BINARY\"", @@ -9366,7 +9366,7 @@ SELECT cast('1' as binary) <=> cast(1 as decimal(20, 0)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BINARY\"", @@ -9388,7 +9388,7 @@ SELECT cast('2017-12-11 09:30:00.0' as timestamp) <=> cast(1 as decimal(3, 0)) F -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP\"", @@ -9410,7 +9410,7 @@ SELECT cast('2017-12-11 09:30:00.0' as timestamp) <=> cast(1 as decimal(5, 0)) F -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP\"", @@ -9432,7 +9432,7 @@ SELECT cast('2017-12-11 09:30:00.0' as timestamp) <=> cast(1 as decimal(10, 0)) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP\"", @@ -9454,7 +9454,7 @@ SELECT cast('2017-12-11 09:30:00.0' as timestamp) <=> cast(1 as decimal(20, 0)) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP\"", @@ -9476,7 +9476,7 @@ SELECT cast('2017-12-11 09:30:00' as date) <=> cast(1 as decimal(3, 0)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DATE\"", @@ -9498,7 +9498,7 @@ SELECT cast('2017-12-11 09:30:00' as date) <=> cast(1 as decimal(5, 0)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DATE\"", @@ -9520,7 +9520,7 @@ SELECT cast('2017-12-11 09:30:00' as date) <=> cast(1 as decimal(10, 0)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DATE\"", @@ -9542,7 +9542,7 @@ SELECT cast('2017-12-11 09:30:00' as date) <=> cast(1 as decimal(20, 0)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DATE\"", @@ -9916,7 +9916,7 @@ SELECT cast(1 as decimal(3, 0)) <=> cast('1' as binary) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(3,0)\"", @@ -9938,7 +9938,7 @@ SELECT cast(1 as decimal(5, 0)) <=> cast('1' as binary) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(5,0)\"", @@ -9960,7 +9960,7 @@ SELECT cast(1 as decimal(10, 0)) <=> cast('1' as binary) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(10,0)\"", @@ -9982,7 +9982,7 @@ SELECT cast(1 as decimal(20, 0)) <=> cast('1' as binary) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(20,0)\"", @@ -10004,7 +10004,7 @@ SELECT cast(1 as decimal(3, 0)) <=> cast(1 as boolean) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(3,0)\"", @@ -10026,7 +10026,7 @@ SELECT cast(1 as decimal(5, 0)) <=> cast(1 as boolean) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(5,0)\"", @@ -10048,7 +10048,7 @@ SELECT cast(1 as decimal(10, 0)) <=> cast(1 as boolean) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(10,0)\"", @@ -10070,7 +10070,7 @@ SELECT cast(1 as decimal(20, 0)) <=> cast(1 as boolean) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(20,0)\"", @@ -10092,7 +10092,7 @@ SELECT cast(1 as decimal(3, 0)) <=> cast('2017-12-11 09:30:00.0' as timestamp) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(3,0)\"", @@ -10114,7 +10114,7 @@ SELECT cast(1 as decimal(5, 0)) <=> cast('2017-12-11 09:30:00.0' as timestamp) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(5,0)\"", @@ -10136,7 +10136,7 @@ SELECT cast(1 as decimal(10, 0)) <=> cast('2017-12-11 09:30:00.0' as timestamp) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(10,0)\"", @@ -10158,7 +10158,7 @@ SELECT cast(1 as decimal(20, 0)) <=> cast('2017-12-11 09:30:00.0' as timestamp) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(20,0)\"", @@ -10180,7 +10180,7 @@ SELECT cast(1 as decimal(3, 0)) <=> cast('2017-12-11 09:30:00' as date) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(3,0)\"", @@ -10202,7 +10202,7 @@ SELECT cast(1 as decimal(5, 0)) <=> cast('2017-12-11 09:30:00' as date) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(5,0)\"", @@ -10224,7 +10224,7 @@ SELECT cast(1 as decimal(10, 0)) <=> cast('2017-12-11 09:30:00' as date) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(10,0)\"", @@ -10246,7 +10246,7 @@ SELECT cast(1 as decimal(20, 0)) <=> cast('2017-12-11 09:30:00' as date) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(20,0)\"", @@ -10576,7 +10576,7 @@ SELECT cast('1' as binary) < cast(1 as decimal(3, 0)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BINARY\"", @@ -10598,7 +10598,7 @@ SELECT cast('1' as binary) < cast(1 as decimal(5, 0)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BINARY\"", @@ -10620,7 +10620,7 @@ SELECT cast('1' as binary) < cast(1 as decimal(10, 0)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BINARY\"", @@ -10642,7 +10642,7 @@ SELECT cast('1' as binary) < cast(1 as decimal(20, 0)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BINARY\"", @@ -10664,7 +10664,7 @@ SELECT cast('2017-12-11 09:30:00.0' as timestamp) < cast(1 as decimal(3, 0)) FRO -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP\"", @@ -10686,7 +10686,7 @@ SELECT cast('2017-12-11 09:30:00.0' as timestamp) < cast(1 as decimal(5, 0)) FRO -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP\"", @@ -10708,7 +10708,7 @@ SELECT cast('2017-12-11 09:30:00.0' as timestamp) < cast(1 as decimal(10, 0)) FR -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP\"", @@ -10730,7 +10730,7 @@ SELECT cast('2017-12-11 09:30:00.0' as timestamp) < cast(1 as decimal(20, 0)) FR -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP\"", @@ -10752,7 +10752,7 @@ SELECT cast('2017-12-11 09:30:00' as date) < cast(1 as decimal(3, 0)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DATE\"", @@ -10774,7 +10774,7 @@ SELECT cast('2017-12-11 09:30:00' as date) < cast(1 as decimal(5, 0)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DATE\"", @@ -10796,7 +10796,7 @@ SELECT cast('2017-12-11 09:30:00' as date) < cast(1 as decimal(10, 0)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DATE\"", @@ -10818,7 +10818,7 @@ SELECT cast('2017-12-11 09:30:00' as date) < cast(1 as decimal(20, 0)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DATE\"", @@ -11192,7 +11192,7 @@ SELECT cast(1 as decimal(3, 0)) < cast('1' as binary) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(3,0)\"", @@ -11214,7 +11214,7 @@ SELECT cast(1 as decimal(5, 0)) < cast('1' as binary) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(5,0)\"", @@ -11236,7 +11236,7 @@ SELECT cast(1 as decimal(10, 0)) < cast('1' as binary) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(10,0)\"", @@ -11258,7 +11258,7 @@ SELECT cast(1 as decimal(20, 0)) < cast('1' as binary) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(20,0)\"", @@ -11280,7 +11280,7 @@ SELECT cast(1 as decimal(3, 0)) < cast(1 as boolean) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(3,0)\"", @@ -11302,7 +11302,7 @@ SELECT cast(1 as decimal(5, 0)) < cast(1 as boolean) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(5,0)\"", @@ -11324,7 +11324,7 @@ SELECT cast(1 as decimal(10, 0)) < cast(1 as boolean) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(10,0)\"", @@ -11346,7 +11346,7 @@ SELECT cast(1 as decimal(20, 0)) < cast(1 as boolean) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(20,0)\"", @@ -11368,7 +11368,7 @@ SELECT cast(1 as decimal(3, 0)) < cast('2017-12-11 09:30:00.0' as timestamp) FR -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(3,0)\"", @@ -11390,7 +11390,7 @@ SELECT cast(1 as decimal(5, 0)) < cast('2017-12-11 09:30:00.0' as timestamp) FR -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(5,0)\"", @@ -11412,7 +11412,7 @@ SELECT cast(1 as decimal(10, 0)) < cast('2017-12-11 09:30:00.0' as timestamp) FR -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(10,0)\"", @@ -11434,7 +11434,7 @@ SELECT cast(1 as decimal(20, 0)) < cast('2017-12-11 09:30:00.0' as timestamp) FR -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(20,0)\"", @@ -11456,7 +11456,7 @@ SELECT cast(1 as decimal(3, 0)) < cast('2017-12-11 09:30:00' as date) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(3,0)\"", @@ -11478,7 +11478,7 @@ SELECT cast(1 as decimal(5, 0)) < cast('2017-12-11 09:30:00' as date) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(5,0)\"", @@ -11500,7 +11500,7 @@ SELECT cast(1 as decimal(10, 0)) < cast('2017-12-11 09:30:00' as date) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(10,0)\"", @@ -11522,7 +11522,7 @@ SELECT cast(1 as decimal(20, 0)) < cast('2017-12-11 09:30:00' as date) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(20,0)\"", @@ -11852,7 +11852,7 @@ SELECT cast('1' as binary) <= cast(1 as decimal(3, 0)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BINARY\"", @@ -11874,7 +11874,7 @@ SELECT cast('1' as binary) <= cast(1 as decimal(5, 0)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BINARY\"", @@ -11896,7 +11896,7 @@ SELECT cast('1' as binary) <= cast(1 as decimal(10, 0)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BINARY\"", @@ -11918,7 +11918,7 @@ SELECT cast('1' as binary) <= cast(1 as decimal(20, 0)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BINARY\"", @@ -11940,7 +11940,7 @@ SELECT cast('2017-12-11 09:30:00.0' as timestamp) <= cast(1 as decimal(3, 0)) FR -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP\"", @@ -11962,7 +11962,7 @@ SELECT cast('2017-12-11 09:30:00.0' as timestamp) <= cast(1 as decimal(5, 0)) FR -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP\"", @@ -11984,7 +11984,7 @@ SELECT cast('2017-12-11 09:30:00.0' as timestamp) <= cast(1 as decimal(10, 0)) F -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP\"", @@ -12006,7 +12006,7 @@ SELECT cast('2017-12-11 09:30:00.0' as timestamp) <= cast(1 as decimal(20, 0)) F -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP\"", @@ -12028,7 +12028,7 @@ SELECT cast('2017-12-11 09:30:00' as date) <= cast(1 as decimal(3, 0)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DATE\"", @@ -12050,7 +12050,7 @@ SELECT cast('2017-12-11 09:30:00' as date) <= cast(1 as decimal(5, 0)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DATE\"", @@ -12072,7 +12072,7 @@ SELECT cast('2017-12-11 09:30:00' as date) <= cast(1 as decimal(10, 0)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DATE\"", @@ -12094,7 +12094,7 @@ SELECT cast('2017-12-11 09:30:00' as date) <= cast(1 as decimal(20, 0)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DATE\"", @@ -12468,7 +12468,7 @@ SELECT cast(1 as decimal(3, 0)) <= cast('1' as binary) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(3,0)\"", @@ -12490,7 +12490,7 @@ SELECT cast(1 as decimal(5, 0)) <= cast('1' as binary) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(5,0)\"", @@ -12512,7 +12512,7 @@ SELECT cast(1 as decimal(10, 0)) <= cast('1' as binary) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(10,0)\"", @@ -12534,7 +12534,7 @@ SELECT cast(1 as decimal(20, 0)) <= cast('1' as binary) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(20,0)\"", @@ -12556,7 +12556,7 @@ SELECT cast(1 as decimal(3, 0)) <= cast(1 as boolean) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(3,0)\"", @@ -12578,7 +12578,7 @@ SELECT cast(1 as decimal(5, 0)) <= cast(1 as boolean) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(5,0)\"", @@ -12600,7 +12600,7 @@ SELECT cast(1 as decimal(10, 0)) <= cast(1 as boolean) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(10,0)\"", @@ -12622,7 +12622,7 @@ SELECT cast(1 as decimal(20, 0)) <= cast(1 as boolean) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(20,0)\"", @@ -12644,7 +12644,7 @@ SELECT cast(1 as decimal(3, 0)) <= cast('2017-12-11 09:30:00.0' as timestamp) F -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(3,0)\"", @@ -12666,7 +12666,7 @@ SELECT cast(1 as decimal(5, 0)) <= cast('2017-12-11 09:30:00.0' as timestamp) F -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(5,0)\"", @@ -12688,7 +12688,7 @@ SELECT cast(1 as decimal(10, 0)) <= cast('2017-12-11 09:30:00.0' as timestamp) F -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(10,0)\"", @@ -12710,7 +12710,7 @@ SELECT cast(1 as decimal(20, 0)) <= cast('2017-12-11 09:30:00.0' as timestamp) F -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(20,0)\"", @@ -12732,7 +12732,7 @@ SELECT cast(1 as decimal(3, 0)) <= cast('2017-12-11 09:30:00' as date) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(3,0)\"", @@ -12754,7 +12754,7 @@ SELECT cast(1 as decimal(5, 0)) <= cast('2017-12-11 09:30:00' as date) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(5,0)\"", @@ -12776,7 +12776,7 @@ SELECT cast(1 as decimal(10, 0)) <= cast('2017-12-11 09:30:00' as date) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(10,0)\"", @@ -12798,7 +12798,7 @@ SELECT cast(1 as decimal(20, 0)) <= cast('2017-12-11 09:30:00' as date) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(20,0)\"", @@ -13128,7 +13128,7 @@ SELECT cast('1' as binary) > cast(1 as decimal(3, 0)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BINARY\"", @@ -13150,7 +13150,7 @@ SELECT cast('1' as binary) > cast(1 as decimal(5, 0)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BINARY\"", @@ -13172,7 +13172,7 @@ SELECT cast('1' as binary) > cast(1 as decimal(10, 0)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BINARY\"", @@ -13194,7 +13194,7 @@ SELECT cast('1' as binary) > cast(1 as decimal(20, 0)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BINARY\"", @@ -13216,7 +13216,7 @@ SELECT cast('2017-12-11 09:30:00.0' as timestamp) > cast(1 as decimal(3, 0)) FRO -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP\"", @@ -13238,7 +13238,7 @@ SELECT cast('2017-12-11 09:30:00.0' as timestamp) > cast(1 as decimal(5, 0)) FRO -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP\"", @@ -13260,7 +13260,7 @@ SELECT cast('2017-12-11 09:30:00.0' as timestamp) > cast(1 as decimal(10, 0)) FR -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP\"", @@ -13282,7 +13282,7 @@ SELECT cast('2017-12-11 09:30:00.0' as timestamp) > cast(1 as decimal(20, 0)) FR -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP\"", @@ -13304,7 +13304,7 @@ SELECT cast('2017-12-11 09:30:00' as date) > cast(1 as decimal(3, 0)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DATE\"", @@ -13326,7 +13326,7 @@ SELECT cast('2017-12-11 09:30:00' as date) > cast(1 as decimal(5, 0)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DATE\"", @@ -13348,7 +13348,7 @@ SELECT cast('2017-12-11 09:30:00' as date) > cast(1 as decimal(10, 0)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DATE\"", @@ -13370,7 +13370,7 @@ SELECT cast('2017-12-11 09:30:00' as date) > cast(1 as decimal(20, 0)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DATE\"", @@ -13744,7 +13744,7 @@ SELECT cast(1 as decimal(3, 0)) > cast('1' as binary) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(3,0)\"", @@ -13766,7 +13766,7 @@ SELECT cast(1 as decimal(5, 0)) > cast('1' as binary) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(5,0)\"", @@ -13788,7 +13788,7 @@ SELECT cast(1 as decimal(10, 0)) > cast('1' as binary) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(10,0)\"", @@ -13810,7 +13810,7 @@ SELECT cast(1 as decimal(20, 0)) > cast('1' as binary) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(20,0)\"", @@ -13832,7 +13832,7 @@ SELECT cast(1 as decimal(3, 0)) > cast(1 as boolean) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(3,0)\"", @@ -13854,7 +13854,7 @@ SELECT cast(1 as decimal(5, 0)) > cast(1 as boolean) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(5,0)\"", @@ -13876,7 +13876,7 @@ SELECT cast(1 as decimal(10, 0)) > cast(1 as boolean) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(10,0)\"", @@ -13898,7 +13898,7 @@ SELECT cast(1 as decimal(20, 0)) > cast(1 as boolean) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(20,0)\"", @@ -13920,7 +13920,7 @@ SELECT cast(1 as decimal(3, 0)) > cast('2017-12-11 09:30:00.0' as timestamp) FR -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(3,0)\"", @@ -13942,7 +13942,7 @@ SELECT cast(1 as decimal(5, 0)) > cast('2017-12-11 09:30:00.0' as timestamp) FR -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(5,0)\"", @@ -13964,7 +13964,7 @@ SELECT cast(1 as decimal(10, 0)) > cast('2017-12-11 09:30:00.0' as timestamp) FR -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(10,0)\"", @@ -13986,7 +13986,7 @@ SELECT cast(1 as decimal(20, 0)) > cast('2017-12-11 09:30:00.0' as timestamp) FR -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(20,0)\"", @@ -14008,7 +14008,7 @@ SELECT cast(1 as decimal(3, 0)) > cast('2017-12-11 09:30:00' as date) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(3,0)\"", @@ -14030,7 +14030,7 @@ SELECT cast(1 as decimal(5, 0)) > cast('2017-12-11 09:30:00' as date) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(5,0)\"", @@ -14052,7 +14052,7 @@ SELECT cast(1 as decimal(10, 0)) > cast('2017-12-11 09:30:00' as date) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(10,0)\"", @@ -14074,7 +14074,7 @@ SELECT cast(1 as decimal(20, 0)) > cast('2017-12-11 09:30:00' as date) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(20,0)\"", @@ -14404,7 +14404,7 @@ SELECT cast('1' as binary) >= cast(1 as decimal(3, 0)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BINARY\"", @@ -14426,7 +14426,7 @@ SELECT cast('1' as binary) >= cast(1 as decimal(5, 0)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BINARY\"", @@ -14448,7 +14448,7 @@ SELECT cast('1' as binary) >= cast(1 as decimal(10, 0)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BINARY\"", @@ -14470,7 +14470,7 @@ SELECT cast('1' as binary) >= cast(1 as decimal(20, 0)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BINARY\"", @@ -14492,7 +14492,7 @@ SELECT cast('2017-12-11 09:30:00.0' as timestamp) >= cast(1 as decimal(3, 0)) FR -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP\"", @@ -14514,7 +14514,7 @@ SELECT cast('2017-12-11 09:30:00.0' as timestamp) >= cast(1 as decimal(5, 0)) FR -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP\"", @@ -14536,7 +14536,7 @@ SELECT cast('2017-12-11 09:30:00.0' as timestamp) >= cast(1 as decimal(10, 0)) F -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP\"", @@ -14558,7 +14558,7 @@ SELECT cast('2017-12-11 09:30:00.0' as timestamp) >= cast(1 as decimal(20, 0)) F -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP\"", @@ -14580,7 +14580,7 @@ SELECT cast('2017-12-11 09:30:00' as date) >= cast(1 as decimal(3, 0)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DATE\"", @@ -14602,7 +14602,7 @@ SELECT cast('2017-12-11 09:30:00' as date) >= cast(1 as decimal(5, 0)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DATE\"", @@ -14624,7 +14624,7 @@ SELECT cast('2017-12-11 09:30:00' as date) >= cast(1 as decimal(10, 0)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DATE\"", @@ -14646,7 +14646,7 @@ SELECT cast('2017-12-11 09:30:00' as date) >= cast(1 as decimal(20, 0)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DATE\"", @@ -15020,7 +15020,7 @@ SELECT cast(1 as decimal(3, 0)) >= cast('1' as binary) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(3,0)\"", @@ -15042,7 +15042,7 @@ SELECT cast(1 as decimal(5, 0)) >= cast('1' as binary) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(5,0)\"", @@ -15064,7 +15064,7 @@ SELECT cast(1 as decimal(10, 0)) >= cast('1' as binary) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(10,0)\"", @@ -15086,7 +15086,7 @@ SELECT cast(1 as decimal(20, 0)) >= cast('1' as binary) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(20,0)\"", @@ -15108,7 +15108,7 @@ SELECT cast(1 as decimal(3, 0)) >= cast(1 as boolean) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(3,0)\"", @@ -15130,7 +15130,7 @@ SELECT cast(1 as decimal(5, 0)) >= cast(1 as boolean) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(5,0)\"", @@ -15152,7 +15152,7 @@ SELECT cast(1 as decimal(10, 0)) >= cast(1 as boolean) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(10,0)\"", @@ -15174,7 +15174,7 @@ SELECT cast(1 as decimal(20, 0)) >= cast(1 as boolean) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(20,0)\"", @@ -15196,7 +15196,7 @@ SELECT cast(1 as decimal(3, 0)) >= cast('2017-12-11 09:30:00.0' as timestamp) F -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(3,0)\"", @@ -15218,7 +15218,7 @@ SELECT cast(1 as decimal(5, 0)) >= cast('2017-12-11 09:30:00.0' as timestamp) F -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(5,0)\"", @@ -15240,7 +15240,7 @@ SELECT cast(1 as decimal(10, 0)) >= cast('2017-12-11 09:30:00.0' as timestamp) F -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(10,0)\"", @@ -15262,7 +15262,7 @@ SELECT cast(1 as decimal(20, 0)) >= cast('2017-12-11 09:30:00.0' as timestamp) F -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(20,0)\"", @@ -15284,7 +15284,7 @@ SELECT cast(1 as decimal(3, 0)) >= cast('2017-12-11 09:30:00' as date) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(3,0)\"", @@ -15306,7 +15306,7 @@ SELECT cast(1 as decimal(5, 0)) >= cast('2017-12-11 09:30:00' as date) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(5,0)\"", @@ -15328,7 +15328,7 @@ SELECT cast(1 as decimal(10, 0)) >= cast('2017-12-11 09:30:00' as date) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(10,0)\"", @@ -15350,7 +15350,7 @@ SELECT cast(1 as decimal(20, 0)) >= cast('2017-12-11 09:30:00' as date) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(20,0)\"", @@ -15680,7 +15680,7 @@ SELECT cast('1' as binary) <> cast(1 as decimal(3, 0)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BINARY\"", @@ -15702,7 +15702,7 @@ SELECT cast('1' as binary) <> cast(1 as decimal(5, 0)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BINARY\"", @@ -15724,7 +15724,7 @@ SELECT cast('1' as binary) <> cast(1 as decimal(10, 0)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BINARY\"", @@ -15746,7 +15746,7 @@ SELECT cast('1' as binary) <> cast(1 as decimal(20, 0)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BINARY\"", @@ -15768,7 +15768,7 @@ SELECT cast('2017-12-11 09:30:00.0' as timestamp) <> cast(1 as decimal(3, 0)) FR -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP\"", @@ -15790,7 +15790,7 @@ SELECT cast('2017-12-11 09:30:00.0' as timestamp) <> cast(1 as decimal(5, 0)) FR -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP\"", @@ -15812,7 +15812,7 @@ SELECT cast('2017-12-11 09:30:00.0' as timestamp) <> cast(1 as decimal(10, 0)) F -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP\"", @@ -15834,7 +15834,7 @@ SELECT cast('2017-12-11 09:30:00.0' as timestamp) <> cast(1 as decimal(20, 0)) F -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP\"", @@ -15856,7 +15856,7 @@ SELECT cast('2017-12-11 09:30:00' as date) <> cast(1 as decimal(3, 0)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DATE\"", @@ -15878,7 +15878,7 @@ SELECT cast('2017-12-11 09:30:00' as date) <> cast(1 as decimal(5, 0)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DATE\"", @@ -15900,7 +15900,7 @@ SELECT cast('2017-12-11 09:30:00' as date) <> cast(1 as decimal(10, 0)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DATE\"", @@ -15922,7 +15922,7 @@ SELECT cast('2017-12-11 09:30:00' as date) <> cast(1 as decimal(20, 0)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DATE\"", @@ -16296,7 +16296,7 @@ SELECT cast(1 as decimal(3, 0)) <> cast('1' as binary) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(3,0)\"", @@ -16318,7 +16318,7 @@ SELECT cast(1 as decimal(5, 0)) <> cast('1' as binary) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(5,0)\"", @@ -16340,7 +16340,7 @@ SELECT cast(1 as decimal(10, 0)) <> cast('1' as binary) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(10,0)\"", @@ -16362,7 +16362,7 @@ SELECT cast(1 as decimal(20, 0)) <> cast('1' as binary) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(20,0)\"", @@ -16384,7 +16384,7 @@ SELECT cast(1 as decimal(3, 0)) <> cast(1 as boolean) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(3,0)\"", @@ -16406,7 +16406,7 @@ SELECT cast(1 as decimal(5, 0)) <> cast(1 as boolean) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(5,0)\"", @@ -16428,7 +16428,7 @@ SELECT cast(1 as decimal(10, 0)) <> cast(1 as boolean) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(10,0)\"", @@ -16450,7 +16450,7 @@ SELECT cast(1 as decimal(20, 0)) <> cast(1 as boolean) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(20,0)\"", @@ -16472,7 +16472,7 @@ SELECT cast(1 as decimal(3, 0)) <> cast('2017-12-11 09:30:00.0' as timestamp) F -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(3,0)\"", @@ -16494,7 +16494,7 @@ SELECT cast(1 as decimal(5, 0)) <> cast('2017-12-11 09:30:00.0' as timestamp) F -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(5,0)\"", @@ -16516,7 +16516,7 @@ SELECT cast(1 as decimal(10, 0)) <> cast('2017-12-11 09:30:00.0' as timestamp) F -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(10,0)\"", @@ -16538,7 +16538,7 @@ SELECT cast(1 as decimal(20, 0)) <> cast('2017-12-11 09:30:00.0' as timestamp) F -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(20,0)\"", @@ -16560,7 +16560,7 @@ SELECT cast(1 as decimal(3, 0)) <> cast('2017-12-11 09:30:00' as date) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(3,0)\"", @@ -16582,7 +16582,7 @@ SELECT cast(1 as decimal(5, 0)) <> cast('2017-12-11 09:30:00' as date) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(5,0)\"", @@ -16604,7 +16604,7 @@ SELECT cast(1 as decimal(10, 0)) <> cast('2017-12-11 09:30:00' as date) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(10,0)\"", @@ -16626,7 +16626,7 @@ SELECT cast(1 as decimal(20, 0)) <> cast('2017-12-11 09:30:00' as date) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(20,0)\"", diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/typeCoercion/native/division.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/typeCoercion/native/division.sql.out index f168b3221150..fedb6fe1d6c7 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/typeCoercion/native/division.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/typeCoercion/native/division.sql.out @@ -100,7 +100,7 @@ SELECT cast(1 as tinyint) / cast('1' as binary) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TINYINT\"", @@ -122,7 +122,7 @@ SELECT cast(1 as tinyint) / cast(1 as boolean) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TINYINT\"", @@ -144,7 +144,7 @@ SELECT cast(1 as tinyint) / cast('2017-12-11 09:30:00.0' as timestamp) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TINYINT\"", @@ -166,7 +166,7 @@ SELECT cast(1 as tinyint) / cast('2017-12-11 09:30:00' as date) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TINYINT\"", @@ -276,7 +276,7 @@ SELECT cast(1 as smallint) / cast('1' as binary) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"SMALLINT\"", @@ -298,7 +298,7 @@ SELECT cast(1 as smallint) / cast(1 as boolean) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"SMALLINT\"", @@ -320,7 +320,7 @@ SELECT cast(1 as smallint) / cast('2017-12-11 09:30:00.0' as timestamp) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"SMALLINT\"", @@ -342,7 +342,7 @@ SELECT cast(1 as smallint) / cast('2017-12-11 09:30:00' as date) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"SMALLINT\"", @@ -452,7 +452,7 @@ SELECT cast(1 as int) / cast('1' as binary) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"INT\"", @@ -474,7 +474,7 @@ SELECT cast(1 as int) / cast(1 as boolean) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"INT\"", @@ -496,7 +496,7 @@ SELECT cast(1 as int) / cast('2017-12-11 09:30:00.0' as timestamp) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"INT\"", @@ -518,7 +518,7 @@ SELECT cast(1 as int) / cast('2017-12-11 09:30:00' as date) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"INT\"", @@ -628,7 +628,7 @@ SELECT cast(1 as bigint) / cast('1' as binary) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BIGINT\"", @@ -650,7 +650,7 @@ SELECT cast(1 as bigint) / cast(1 as boolean) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BIGINT\"", @@ -672,7 +672,7 @@ SELECT cast(1 as bigint) / cast('2017-12-11 09:30:00.0' as timestamp) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BIGINT\"", @@ -694,7 +694,7 @@ SELECT cast(1 as bigint) / cast('2017-12-11 09:30:00' as date) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BIGINT\"", @@ -804,7 +804,7 @@ SELECT cast(1 as float) / cast('1' as binary) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"FLOAT\"", @@ -826,7 +826,7 @@ SELECT cast(1 as float) / cast(1 as boolean) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"FLOAT\"", @@ -848,7 +848,7 @@ SELECT cast(1 as float) / cast('2017-12-11 09:30:00.0' as timestamp) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"FLOAT\"", @@ -870,7 +870,7 @@ SELECT cast(1 as float) / cast('2017-12-11 09:30:00' as date) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"FLOAT\"", @@ -980,7 +980,7 @@ SELECT cast(1 as double) / cast('1' as binary) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DOUBLE\"", @@ -1002,7 +1002,7 @@ SELECT cast(1 as double) / cast(1 as boolean) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DOUBLE\"", @@ -1024,7 +1024,7 @@ SELECT cast(1 as double) / cast('2017-12-11 09:30:00.0' as timestamp) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DOUBLE\"", @@ -1046,7 +1046,7 @@ SELECT cast(1 as double) / cast('2017-12-11 09:30:00' as date) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DOUBLE\"", @@ -1156,7 +1156,7 @@ SELECT cast(1 as decimal(10, 0)) / cast('1' as binary) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(10,0)\"", @@ -1178,7 +1178,7 @@ SELECT cast(1 as decimal(10, 0)) / cast(1 as boolean) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(10,0)\"", @@ -1200,7 +1200,7 @@ SELECT cast(1 as decimal(10, 0)) / cast('2017-12-11 09:30:00.0' as timestamp) FR -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(10,0)\"", @@ -1222,7 +1222,7 @@ SELECT cast(1 as decimal(10, 0)) / cast('2017-12-11 09:30:00' as date) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(10,0)\"", @@ -1321,7 +1321,7 @@ SELECT cast(1 as string) / cast(1 as string) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"STRING\"", @@ -1343,7 +1343,7 @@ SELECT cast(1 as string) / cast('1' as binary) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"BINARY\"", @@ -1365,7 +1365,7 @@ SELECT cast(1 as string) / cast(1 as boolean) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"BOOLEAN\"", @@ -1387,7 +1387,7 @@ SELECT cast(1 as string) / cast('2017-12-11 09:30:00.0' as timestamp) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"TIMESTAMP\"", @@ -1409,7 +1409,7 @@ SELECT cast(1 as string) / cast('2017-12-11 09:30:00' as date) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"DATE\"", @@ -1431,7 +1431,7 @@ SELECT cast('1' as binary) / cast(1 as tinyint) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BINARY\"", @@ -1453,7 +1453,7 @@ SELECT cast('1' as binary) / cast(1 as smallint) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BINARY\"", @@ -1475,7 +1475,7 @@ SELECT cast('1' as binary) / cast(1 as int) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BINARY\"", @@ -1497,7 +1497,7 @@ SELECT cast('1' as binary) / cast(1 as bigint) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BINARY\"", @@ -1519,7 +1519,7 @@ SELECT cast('1' as binary) / cast(1 as float) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BINARY\"", @@ -1541,7 +1541,7 @@ SELECT cast('1' as binary) / cast(1 as double) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BINARY\"", @@ -1563,7 +1563,7 @@ SELECT cast('1' as binary) / cast(1 as decimal(10, 0)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BINARY\"", @@ -1585,7 +1585,7 @@ SELECT cast('1' as binary) / cast(1 as string) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"BINARY\"", @@ -1607,7 +1607,7 @@ SELECT cast('1' as binary) / cast('1' as binary) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"BINARY\"", @@ -1629,7 +1629,7 @@ SELECT cast('1' as binary) / cast(1 as boolean) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BINARY\"", @@ -1651,7 +1651,7 @@ SELECT cast('1' as binary) / cast('2017-12-11 09:30:00.0' as timestamp) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BINARY\"", @@ -1673,7 +1673,7 @@ SELECT cast('1' as binary) / cast('2017-12-11 09:30:00' as date) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BINARY\"", @@ -1695,7 +1695,7 @@ SELECT cast(1 as boolean) / cast(1 as tinyint) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BOOLEAN\"", @@ -1717,7 +1717,7 @@ SELECT cast(1 as boolean) / cast(1 as smallint) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BOOLEAN\"", @@ -1739,7 +1739,7 @@ SELECT cast(1 as boolean) / cast(1 as int) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BOOLEAN\"", @@ -1761,7 +1761,7 @@ SELECT cast(1 as boolean) / cast(1 as bigint) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BOOLEAN\"", @@ -1783,7 +1783,7 @@ SELECT cast(1 as boolean) / cast(1 as float) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BOOLEAN\"", @@ -1805,7 +1805,7 @@ SELECT cast(1 as boolean) / cast(1 as double) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BOOLEAN\"", @@ -1827,7 +1827,7 @@ SELECT cast(1 as boolean) / cast(1 as decimal(10, 0)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BOOLEAN\"", @@ -1849,7 +1849,7 @@ SELECT cast(1 as boolean) / cast(1 as string) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"BOOLEAN\"", @@ -1871,7 +1871,7 @@ SELECT cast(1 as boolean) / cast('1' as binary) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BOOLEAN\"", @@ -1893,7 +1893,7 @@ SELECT cast(1 as boolean) / cast(1 as boolean) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"BOOLEAN\"", @@ -1915,7 +1915,7 @@ SELECT cast(1 as boolean) / cast('2017-12-11 09:30:00.0' as timestamp) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BOOLEAN\"", @@ -1937,7 +1937,7 @@ SELECT cast(1 as boolean) / cast('2017-12-11 09:30:00' as date) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BOOLEAN\"", @@ -1959,7 +1959,7 @@ SELECT cast('2017-12-11 09:30:00.0' as timestamp) / cast(1 as tinyint) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP\"", @@ -1981,7 +1981,7 @@ SELECT cast('2017-12-11 09:30:00.0' as timestamp) / cast(1 as smallint) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP\"", @@ -2003,7 +2003,7 @@ SELECT cast('2017-12-11 09:30:00.0' as timestamp) / cast(1 as int) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP\"", @@ -2025,7 +2025,7 @@ SELECT cast('2017-12-11 09:30:00.0' as timestamp) / cast(1 as bigint) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP\"", @@ -2047,7 +2047,7 @@ SELECT cast('2017-12-11 09:30:00.0' as timestamp) / cast(1 as float) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP\"", @@ -2069,7 +2069,7 @@ SELECT cast('2017-12-11 09:30:00.0' as timestamp) / cast(1 as double) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP\"", @@ -2091,7 +2091,7 @@ SELECT cast('2017-12-11 09:30:00.0' as timestamp) / cast(1 as decimal(10, 0)) FR -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP\"", @@ -2113,7 +2113,7 @@ SELECT cast('2017-12-11 09:30:00.0' as timestamp) / cast(1 as string) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"TIMESTAMP\"", @@ -2135,7 +2135,7 @@ SELECT cast('2017-12-11 09:30:00.0' as timestamp) / cast('1' as binary) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP\"", @@ -2157,7 +2157,7 @@ SELECT cast('2017-12-11 09:30:00.0' as timestamp) / cast(1 as boolean) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP\"", @@ -2179,7 +2179,7 @@ SELECT cast('2017-12-11 09:30:00.0' as timestamp) / cast('2017-12-11 09:30:00.0' -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"TIMESTAMP\"", @@ -2201,7 +2201,7 @@ SELECT cast('2017-12-11 09:30:00.0' as timestamp) / cast('2017-12-11 09:30:00' a -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP\"", @@ -2223,7 +2223,7 @@ SELECT cast('2017-12-11 09:30:00' as date) / cast(1 as tinyint) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DATE\"", @@ -2245,7 +2245,7 @@ SELECT cast('2017-12-11 09:30:00' as date) / cast(1 as smallint) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DATE\"", @@ -2267,7 +2267,7 @@ SELECT cast('2017-12-11 09:30:00' as date) / cast(1 as int) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DATE\"", @@ -2289,7 +2289,7 @@ SELECT cast('2017-12-11 09:30:00' as date) / cast(1 as bigint) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DATE\"", @@ -2311,7 +2311,7 @@ SELECT cast('2017-12-11 09:30:00' as date) / cast(1 as float) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DATE\"", @@ -2333,7 +2333,7 @@ SELECT cast('2017-12-11 09:30:00' as date) / cast(1 as double) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DATE\"", @@ -2355,7 +2355,7 @@ SELECT cast('2017-12-11 09:30:00' as date) / cast(1 as decimal(10, 0)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DATE\"", @@ -2377,7 +2377,7 @@ SELECT cast('2017-12-11 09:30:00' as date) / cast(1 as string) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"DATE\"", @@ -2399,7 +2399,7 @@ SELECT cast('2017-12-11 09:30:00' as date) / cast('1' as binary) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DATE\"", @@ -2421,7 +2421,7 @@ SELECT cast('2017-12-11 09:30:00' as date) / cast(1 as boolean) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DATE\"", @@ -2443,7 +2443,7 @@ SELECT cast('2017-12-11 09:30:00' as date) / cast('2017-12-11 09:30:00.0' as tim -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DATE\"", @@ -2465,7 +2465,7 @@ SELECT cast('2017-12-11 09:30:00' as date) / cast('2017-12-11 09:30:00' as date) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"DATE\"", diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/typeCoercion/native/ifCoercion.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/typeCoercion/native/ifCoercion.sql.out index 791b75a86971..8332de096dc4 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/typeCoercion/native/ifCoercion.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/typeCoercion/native/ifCoercion.sql.out @@ -100,7 +100,7 @@ SELECT IF(true, cast(1 as tinyint), cast('2' as binary)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"TINYINT\", \"BINARY\"]", @@ -122,7 +122,7 @@ SELECT IF(true, cast(1 as tinyint), cast(2 as boolean)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"TINYINT\", \"BOOLEAN\"]", @@ -144,7 +144,7 @@ SELECT IF(true, cast(1 as tinyint), cast('2017-12-11 09:30:00.0' as timestamp)) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"TINYINT\", \"TIMESTAMP\"]", @@ -166,7 +166,7 @@ SELECT IF(true, cast(1 as tinyint), cast('2017-12-11 09:30:00' as date)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"TINYINT\", \"DATE\"]", @@ -276,7 +276,7 @@ SELECT IF(true, cast(1 as smallint), cast('2' as binary)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"SMALLINT\", \"BINARY\"]", @@ -298,7 +298,7 @@ SELECT IF(true, cast(1 as smallint), cast(2 as boolean)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"SMALLINT\", \"BOOLEAN\"]", @@ -320,7 +320,7 @@ SELECT IF(true, cast(1 as smallint), cast('2017-12-11 09:30:00.0' as timestamp)) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"SMALLINT\", \"TIMESTAMP\"]", @@ -342,7 +342,7 @@ SELECT IF(true, cast(1 as smallint), cast('2017-12-11 09:30:00' as date)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"SMALLINT\", \"DATE\"]", @@ -452,7 +452,7 @@ SELECT IF(true, cast(1 as int), cast('2' as binary)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"INT\", \"BINARY\"]", @@ -474,7 +474,7 @@ SELECT IF(true, cast(1 as int), cast(2 as boolean)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"INT\", \"BOOLEAN\"]", @@ -496,7 +496,7 @@ SELECT IF(true, cast(1 as int), cast('2017-12-11 09:30:00.0' as timestamp)) FROM -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"INT\", \"TIMESTAMP\"]", @@ -518,7 +518,7 @@ SELECT IF(true, cast(1 as int), cast('2017-12-11 09:30:00' as date)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"INT\", \"DATE\"]", @@ -628,7 +628,7 @@ SELECT IF(true, cast(1 as bigint), cast('2' as binary)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BIGINT\", \"BINARY\"]", @@ -650,7 +650,7 @@ SELECT IF(true, cast(1 as bigint), cast(2 as boolean)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BIGINT\", \"BOOLEAN\"]", @@ -672,7 +672,7 @@ SELECT IF(true, cast(1 as bigint), cast('2017-12-11 09:30:00.0' as timestamp)) F -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BIGINT\", \"TIMESTAMP\"]", @@ -694,7 +694,7 @@ SELECT IF(true, cast(1 as bigint), cast('2017-12-11 09:30:00' as date)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BIGINT\", \"DATE\"]", @@ -804,7 +804,7 @@ SELECT IF(true, cast(1 as float), cast('2' as binary)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"FLOAT\", \"BINARY\"]", @@ -826,7 +826,7 @@ SELECT IF(true, cast(1 as float), cast(2 as boolean)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"FLOAT\", \"BOOLEAN\"]", @@ -848,7 +848,7 @@ SELECT IF(true, cast(1 as float), cast('2017-12-11 09:30:00.0' as timestamp)) FR -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"FLOAT\", \"TIMESTAMP\"]", @@ -870,7 +870,7 @@ SELECT IF(true, cast(1 as float), cast('2017-12-11 09:30:00' as date)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"FLOAT\", \"DATE\"]", @@ -980,7 +980,7 @@ SELECT IF(true, cast(1 as double), cast('2' as binary)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DOUBLE\", \"BINARY\"]", @@ -1002,7 +1002,7 @@ SELECT IF(true, cast(1 as double), cast(2 as boolean)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DOUBLE\", \"BOOLEAN\"]", @@ -1024,7 +1024,7 @@ SELECT IF(true, cast(1 as double), cast('2017-12-11 09:30:00.0' as timestamp)) F -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DOUBLE\", \"TIMESTAMP\"]", @@ -1046,7 +1046,7 @@ SELECT IF(true, cast(1 as double), cast('2017-12-11 09:30:00' as date)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DOUBLE\", \"DATE\"]", @@ -1156,7 +1156,7 @@ SELECT IF(true, cast(1 as decimal(10, 0)), cast('2' as binary)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DECIMAL(10,0)\", \"BINARY\"]", @@ -1178,7 +1178,7 @@ SELECT IF(true, cast(1 as decimal(10, 0)), cast(2 as boolean)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DECIMAL(10,0)\", \"BOOLEAN\"]", @@ -1200,7 +1200,7 @@ SELECT IF(true, cast(1 as decimal(10, 0)), cast('2017-12-11 09:30:00.0' as times -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DECIMAL(10,0)\", \"TIMESTAMP\"]", @@ -1222,7 +1222,7 @@ SELECT IF(true, cast(1 as decimal(10, 0)), cast('2017-12-11 09:30:00' as date)) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DECIMAL(10,0)\", \"DATE\"]", @@ -1376,7 +1376,7 @@ SELECT IF(true, cast('1' as binary), cast(2 as tinyint)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BINARY\", \"TINYINT\"]", @@ -1398,7 +1398,7 @@ SELECT IF(true, cast('1' as binary), cast(2 as smallint)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BINARY\", \"SMALLINT\"]", @@ -1420,7 +1420,7 @@ SELECT IF(true, cast('1' as binary), cast(2 as int)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BINARY\", \"INT\"]", @@ -1442,7 +1442,7 @@ SELECT IF(true, cast('1' as binary), cast(2 as bigint)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BINARY\", \"BIGINT\"]", @@ -1464,7 +1464,7 @@ SELECT IF(true, cast('1' as binary), cast(2 as float)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BINARY\", \"FLOAT\"]", @@ -1486,7 +1486,7 @@ SELECT IF(true, cast('1' as binary), cast(2 as double)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BINARY\", \"DOUBLE\"]", @@ -1508,7 +1508,7 @@ SELECT IF(true, cast('1' as binary), cast(2 as decimal(10, 0))) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BINARY\", \"DECIMAL(10,0)\"]", @@ -1552,7 +1552,7 @@ SELECT IF(true, cast('1' as binary), cast(2 as boolean)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BINARY\", \"BOOLEAN\"]", @@ -1574,7 +1574,7 @@ SELECT IF(true, cast('1' as binary), cast('2017-12-11 09:30:00.0' as timestamp)) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BINARY\", \"TIMESTAMP\"]", @@ -1596,7 +1596,7 @@ SELECT IF(true, cast('1' as binary), cast('2017-12-11 09:30:00' as date)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BINARY\", \"DATE\"]", @@ -1618,7 +1618,7 @@ SELECT IF(true, cast(1 as boolean), cast(2 as tinyint)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BOOLEAN\", \"TINYINT\"]", @@ -1640,7 +1640,7 @@ SELECT IF(true, cast(1 as boolean), cast(2 as smallint)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BOOLEAN\", \"SMALLINT\"]", @@ -1662,7 +1662,7 @@ SELECT IF(true, cast(1 as boolean), cast(2 as int)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BOOLEAN\", \"INT\"]", @@ -1684,7 +1684,7 @@ SELECT IF(true, cast(1 as boolean), cast(2 as bigint)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BOOLEAN\", \"BIGINT\"]", @@ -1706,7 +1706,7 @@ SELECT IF(true, cast(1 as boolean), cast(2 as float)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BOOLEAN\", \"FLOAT\"]", @@ -1728,7 +1728,7 @@ SELECT IF(true, cast(1 as boolean), cast(2 as double)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BOOLEAN\", \"DOUBLE\"]", @@ -1750,7 +1750,7 @@ SELECT IF(true, cast(1 as boolean), cast(2 as decimal(10, 0))) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BOOLEAN\", \"DECIMAL(10,0)\"]", @@ -1783,7 +1783,7 @@ SELECT IF(true, cast(1 as boolean), cast('2' as binary)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BOOLEAN\", \"BINARY\"]", @@ -1816,7 +1816,7 @@ SELECT IF(true, cast(1 as boolean), cast('2017-12-11 09:30:00.0' as timestamp)) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BOOLEAN\", \"TIMESTAMP\"]", @@ -1838,7 +1838,7 @@ SELECT IF(true, cast(1 as boolean), cast('2017-12-11 09:30:00' as date)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BOOLEAN\", \"DATE\"]", @@ -1860,7 +1860,7 @@ SELECT IF(true, cast('2017-12-12 09:30:00.0' as timestamp), cast(2 as tinyint)) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"TIMESTAMP\", \"TINYINT\"]", @@ -1882,7 +1882,7 @@ SELECT IF(true, cast('2017-12-12 09:30:00.0' as timestamp), cast(2 as smallint)) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"TIMESTAMP\", \"SMALLINT\"]", @@ -1904,7 +1904,7 @@ SELECT IF(true, cast('2017-12-12 09:30:00.0' as timestamp), cast(2 as int)) FROM -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"TIMESTAMP\", \"INT\"]", @@ -1926,7 +1926,7 @@ SELECT IF(true, cast('2017-12-12 09:30:00.0' as timestamp), cast(2 as bigint)) F -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"TIMESTAMP\", \"BIGINT\"]", @@ -1948,7 +1948,7 @@ SELECT IF(true, cast('2017-12-12 09:30:00.0' as timestamp), cast(2 as float)) FR -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"TIMESTAMP\", \"FLOAT\"]", @@ -1970,7 +1970,7 @@ SELECT IF(true, cast('2017-12-12 09:30:00.0' as timestamp), cast(2 as double)) F -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"TIMESTAMP\", \"DOUBLE\"]", @@ -1992,7 +1992,7 @@ SELECT IF(true, cast('2017-12-12 09:30:00.0' as timestamp), cast(2 as decimal(10 -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"TIMESTAMP\", \"DECIMAL(10,0)\"]", @@ -2025,7 +2025,7 @@ SELECT IF(true, cast('2017-12-12 09:30:00.0' as timestamp), cast('2' as binary)) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"TIMESTAMP\", \"BINARY\"]", @@ -2047,7 +2047,7 @@ SELECT IF(true, cast('2017-12-12 09:30:00.0' as timestamp), cast(2 as boolean)) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"TIMESTAMP\", \"BOOLEAN\"]", @@ -2091,7 +2091,7 @@ SELECT IF(true, cast('2017-12-12 09:30:00' as date), cast(2 as tinyint)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DATE\", \"TINYINT\"]", @@ -2113,7 +2113,7 @@ SELECT IF(true, cast('2017-12-12 09:30:00' as date), cast(2 as smallint)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DATE\", \"SMALLINT\"]", @@ -2135,7 +2135,7 @@ SELECT IF(true, cast('2017-12-12 09:30:00' as date), cast(2 as int)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DATE\", \"INT\"]", @@ -2157,7 +2157,7 @@ SELECT IF(true, cast('2017-12-12 09:30:00' as date), cast(2 as bigint)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DATE\", \"BIGINT\"]", @@ -2179,7 +2179,7 @@ SELECT IF(true, cast('2017-12-12 09:30:00' as date), cast(2 as float)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DATE\", \"FLOAT\"]", @@ -2201,7 +2201,7 @@ SELECT IF(true, cast('2017-12-12 09:30:00' as date), cast(2 as double)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DATE\", \"DOUBLE\"]", @@ -2223,7 +2223,7 @@ SELECT IF(true, cast('2017-12-12 09:30:00' as date), cast(2 as decimal(10, 0))) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DATE\", \"DECIMAL(10,0)\"]", @@ -2256,7 +2256,7 @@ SELECT IF(true, cast('2017-12-12 09:30:00' as date), cast('2' as binary)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DATE\", \"BINARY\"]", @@ -2278,7 +2278,7 @@ SELECT IF(true, cast('2017-12-12 09:30:00' as date), cast(2 as boolean)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DATE\", \"BOOLEAN\"]", diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/typeCoercion/native/inConversion.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/typeCoercion/native/inConversion.sql.out index 71bc2fef3ab8..55172ac7545b 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/typeCoercion/native/inConversion.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/typeCoercion/native/inConversion.sql.out @@ -100,7 +100,7 @@ SELECT cast(1 as tinyint) in (cast('1' as binary)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"TINYINT\", \"BINARY\"]", @@ -122,7 +122,7 @@ SELECT cast(1 as tinyint) in (cast(1 as boolean)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"TINYINT\", \"BOOLEAN\"]", @@ -144,7 +144,7 @@ SELECT cast(1 as tinyint) in (cast('2017-12-11 09:30:00.0' as timestamp)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"TINYINT\", \"TIMESTAMP\"]", @@ -166,7 +166,7 @@ SELECT cast(1 as tinyint) in (cast('2017-12-11 09:30:00' as date)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"TINYINT\", \"DATE\"]", @@ -276,7 +276,7 @@ SELECT cast(1 as smallint) in (cast('1' as binary)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"SMALLINT\", \"BINARY\"]", @@ -298,7 +298,7 @@ SELECT cast(1 as smallint) in (cast(1 as boolean)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"SMALLINT\", \"BOOLEAN\"]", @@ -320,7 +320,7 @@ SELECT cast(1 as smallint) in (cast('2017-12-11 09:30:00.0' as timestamp)) FROM -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"SMALLINT\", \"TIMESTAMP\"]", @@ -342,7 +342,7 @@ SELECT cast(1 as smallint) in (cast('2017-12-11 09:30:00' as date)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"SMALLINT\", \"DATE\"]", @@ -452,7 +452,7 @@ SELECT cast(1 as int) in (cast('1' as binary)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"INT\", \"BINARY\"]", @@ -474,7 +474,7 @@ SELECT cast(1 as int) in (cast(1 as boolean)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"INT\", \"BOOLEAN\"]", @@ -496,7 +496,7 @@ SELECT cast(1 as int) in (cast('2017-12-11 09:30:00.0' as timestamp)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"INT\", \"TIMESTAMP\"]", @@ -518,7 +518,7 @@ SELECT cast(1 as int) in (cast('2017-12-11 09:30:00' as date)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"INT\", \"DATE\"]", @@ -628,7 +628,7 @@ SELECT cast(1 as bigint) in (cast('1' as binary)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BIGINT\", \"BINARY\"]", @@ -650,7 +650,7 @@ SELECT cast(1 as bigint) in (cast(1 as boolean)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BIGINT\", \"BOOLEAN\"]", @@ -672,7 +672,7 @@ SELECT cast(1 as bigint) in (cast('2017-12-11 09:30:00.0' as timestamp)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BIGINT\", \"TIMESTAMP\"]", @@ -694,7 +694,7 @@ SELECT cast(1 as bigint) in (cast('2017-12-11 09:30:00' as date)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BIGINT\", \"DATE\"]", @@ -804,7 +804,7 @@ SELECT cast(1 as float) in (cast('1' as binary)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"FLOAT\", \"BINARY\"]", @@ -826,7 +826,7 @@ SELECT cast(1 as float) in (cast(1 as boolean)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"FLOAT\", \"BOOLEAN\"]", @@ -848,7 +848,7 @@ SELECT cast(1 as float) in (cast('2017-12-11 09:30:00.0' as timestamp)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"FLOAT\", \"TIMESTAMP\"]", @@ -870,7 +870,7 @@ SELECT cast(1 as float) in (cast('2017-12-11 09:30:00' as date)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"FLOAT\", \"DATE\"]", @@ -980,7 +980,7 @@ SELECT cast(1 as double) in (cast('1' as binary)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DOUBLE\", \"BINARY\"]", @@ -1002,7 +1002,7 @@ SELECT cast(1 as double) in (cast(1 as boolean)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DOUBLE\", \"BOOLEAN\"]", @@ -1024,7 +1024,7 @@ SELECT cast(1 as double) in (cast('2017-12-11 09:30:00.0' as timestamp)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DOUBLE\", \"TIMESTAMP\"]", @@ -1046,7 +1046,7 @@ SELECT cast(1 as double) in (cast('2017-12-11 09:30:00' as date)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DOUBLE\", \"DATE\"]", @@ -1156,7 +1156,7 @@ SELECT cast(1 as decimal(10, 0)) in (cast('1' as binary)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DECIMAL(10,0)\", \"BINARY\"]", @@ -1178,7 +1178,7 @@ SELECT cast(1 as decimal(10, 0)) in (cast(1 as boolean)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DECIMAL(10,0)\", \"BOOLEAN\"]", @@ -1200,7 +1200,7 @@ SELECT cast(1 as decimal(10, 0)) in (cast('2017-12-11 09:30:00.0' as timestamp)) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DECIMAL(10,0)\", \"TIMESTAMP\"]", @@ -1222,7 +1222,7 @@ SELECT cast(1 as decimal(10, 0)) in (cast('2017-12-11 09:30:00' as date)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DECIMAL(10,0)\", \"DATE\"]", @@ -1376,7 +1376,7 @@ SELECT cast('1' as binary) in (cast(1 as tinyint)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BINARY\", \"TINYINT\"]", @@ -1398,7 +1398,7 @@ SELECT cast('1' as binary) in (cast(1 as smallint)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BINARY\", \"SMALLINT\"]", @@ -1420,7 +1420,7 @@ SELECT cast('1' as binary) in (cast(1 as int)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BINARY\", \"INT\"]", @@ -1442,7 +1442,7 @@ SELECT cast('1' as binary) in (cast(1 as bigint)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BINARY\", \"BIGINT\"]", @@ -1464,7 +1464,7 @@ SELECT cast('1' as binary) in (cast(1 as float)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BINARY\", \"FLOAT\"]", @@ -1486,7 +1486,7 @@ SELECT cast('1' as binary) in (cast(1 as double)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BINARY\", \"DOUBLE\"]", @@ -1508,7 +1508,7 @@ SELECT cast('1' as binary) in (cast(1 as decimal(10, 0))) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BINARY\", \"DECIMAL(10,0)\"]", @@ -1552,7 +1552,7 @@ SELECT cast('1' as binary) in (cast(1 as boolean)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BINARY\", \"BOOLEAN\"]", @@ -1574,7 +1574,7 @@ SELECT cast('1' as binary) in (cast('2017-12-11 09:30:00.0' as timestamp)) FROM -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BINARY\", \"TIMESTAMP\"]", @@ -1596,7 +1596,7 @@ SELECT cast('1' as binary) in (cast('2017-12-11 09:30:00' as date)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BINARY\", \"DATE\"]", @@ -1618,7 +1618,7 @@ SELECT true in (cast(1 as tinyint)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BOOLEAN\", \"TINYINT\"]", @@ -1640,7 +1640,7 @@ SELECT true in (cast(1 as smallint)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BOOLEAN\", \"SMALLINT\"]", @@ -1662,7 +1662,7 @@ SELECT true in (cast(1 as int)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BOOLEAN\", \"INT\"]", @@ -1684,7 +1684,7 @@ SELECT true in (cast(1 as bigint)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BOOLEAN\", \"BIGINT\"]", @@ -1706,7 +1706,7 @@ SELECT true in (cast(1 as float)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BOOLEAN\", \"FLOAT\"]", @@ -1728,7 +1728,7 @@ SELECT true in (cast(1 as double)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BOOLEAN\", \"DOUBLE\"]", @@ -1750,7 +1750,7 @@ SELECT true in (cast(1 as decimal(10, 0))) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BOOLEAN\", \"DECIMAL(10,0)\"]", @@ -1783,7 +1783,7 @@ SELECT true in (cast('1' as binary)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BOOLEAN\", \"BINARY\"]", @@ -1816,7 +1816,7 @@ SELECT true in (cast('2017-12-11 09:30:00.0' as timestamp)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BOOLEAN\", \"TIMESTAMP\"]", @@ -1838,7 +1838,7 @@ SELECT true in (cast('2017-12-11 09:30:00' as date)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BOOLEAN\", \"DATE\"]", @@ -1860,7 +1860,7 @@ SELECT cast('2017-12-12 09:30:00.0' as timestamp) in (cast(2 as tinyint)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"TIMESTAMP\", \"TINYINT\"]", @@ -1882,7 +1882,7 @@ SELECT cast('2017-12-12 09:30:00.0' as timestamp) in (cast(2 as smallint)) FROM -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"TIMESTAMP\", \"SMALLINT\"]", @@ -1904,7 +1904,7 @@ SELECT cast('2017-12-12 09:30:00.0' as timestamp) in (cast(2 as int)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"TIMESTAMP\", \"INT\"]", @@ -1926,7 +1926,7 @@ SELECT cast('2017-12-12 09:30:00.0' as timestamp) in (cast(2 as bigint)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"TIMESTAMP\", \"BIGINT\"]", @@ -1948,7 +1948,7 @@ SELECT cast('2017-12-12 09:30:00.0' as timestamp) in (cast(2 as float)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"TIMESTAMP\", \"FLOAT\"]", @@ -1970,7 +1970,7 @@ SELECT cast('2017-12-12 09:30:00.0' as timestamp) in (cast(2 as double)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"TIMESTAMP\", \"DOUBLE\"]", @@ -1992,7 +1992,7 @@ SELECT cast('2017-12-12 09:30:00.0' as timestamp) in (cast(2 as decimal(10, 0))) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"TIMESTAMP\", \"DECIMAL(10,0)\"]", @@ -2025,7 +2025,7 @@ SELECT cast('2017-12-12 09:30:00.0' as timestamp) in (cast('2' as binary)) FROM -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"TIMESTAMP\", \"BINARY\"]", @@ -2047,7 +2047,7 @@ SELECT cast('2017-12-12 09:30:00.0' as timestamp) in (cast(2 as boolean)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"TIMESTAMP\", \"BOOLEAN\"]", @@ -2091,7 +2091,7 @@ SELECT cast('2017-12-12 09:30:00' as date) in (cast(2 as tinyint)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DATE\", \"TINYINT\"]", @@ -2113,7 +2113,7 @@ SELECT cast('2017-12-12 09:30:00' as date) in (cast(2 as smallint)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DATE\", \"SMALLINT\"]", @@ -2135,7 +2135,7 @@ SELECT cast('2017-12-12 09:30:00' as date) in (cast(2 as int)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DATE\", \"INT\"]", @@ -2157,7 +2157,7 @@ SELECT cast('2017-12-12 09:30:00' as date) in (cast(2 as bigint)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DATE\", \"BIGINT\"]", @@ -2179,7 +2179,7 @@ SELECT cast('2017-12-12 09:30:00' as date) in (cast(2 as float)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DATE\", \"FLOAT\"]", @@ -2201,7 +2201,7 @@ SELECT cast('2017-12-12 09:30:00' as date) in (cast(2 as double)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DATE\", \"DOUBLE\"]", @@ -2223,7 +2223,7 @@ SELECT cast('2017-12-12 09:30:00' as date) in (cast(2 as decimal(10, 0))) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DATE\", \"DECIMAL(10,0)\"]", @@ -2256,7 +2256,7 @@ SELECT cast('2017-12-12 09:30:00' as date) in (cast('2' as binary)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DATE\", \"BINARY\"]", @@ -2278,7 +2278,7 @@ SELECT cast('2017-12-12 09:30:00' as date) in (cast(2 as boolean)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DATE\", \"BOOLEAN\"]", @@ -2410,7 +2410,7 @@ SELECT cast(1 as tinyint) in (cast(1 as tinyint), cast('1' as binary)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"TINYINT\", \"TINYINT\", \"BINARY\"]", @@ -2432,7 +2432,7 @@ SELECT cast(1 as tinyint) in (cast(1 as tinyint), cast(1 as boolean)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"TINYINT\", \"TINYINT\", \"BOOLEAN\"]", @@ -2454,7 +2454,7 @@ SELECT cast(1 as tinyint) in (cast(1 as tinyint), cast('2017-12-11 09:30:00.0' a -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"TINYINT\", \"TINYINT\", \"TIMESTAMP\"]", @@ -2476,7 +2476,7 @@ SELECT cast(1 as tinyint) in (cast(1 as tinyint), cast('2017-12-11 09:30:00' as -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"TINYINT\", \"TINYINT\", \"DATE\"]", @@ -2586,7 +2586,7 @@ SELECT cast(1 as smallint) in (cast(1 as smallint), cast('1' as binary)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"SMALLINT\", \"SMALLINT\", \"BINARY\"]", @@ -2608,7 +2608,7 @@ SELECT cast(1 as smallint) in (cast(1 as smallint), cast(1 as boolean)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"SMALLINT\", \"SMALLINT\", \"BOOLEAN\"]", @@ -2630,7 +2630,7 @@ SELECT cast(1 as smallint) in (cast(1 as smallint), cast('2017-12-11 09:30:00.0' -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"SMALLINT\", \"SMALLINT\", \"TIMESTAMP\"]", @@ -2652,7 +2652,7 @@ SELECT cast(1 as smallint) in (cast(1 as smallint), cast('2017-12-11 09:30:00' a -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"SMALLINT\", \"SMALLINT\", \"DATE\"]", @@ -2762,7 +2762,7 @@ SELECT cast(1 as int) in (cast(1 as int), cast('1' as binary)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"INT\", \"INT\", \"BINARY\"]", @@ -2784,7 +2784,7 @@ SELECT cast(1 as int) in (cast(1 as int), cast(1 as boolean)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"INT\", \"INT\", \"BOOLEAN\"]", @@ -2806,7 +2806,7 @@ SELECT cast(1 as int) in (cast(1 as int), cast('2017-12-11 09:30:00.0' as timest -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"INT\", \"INT\", \"TIMESTAMP\"]", @@ -2828,7 +2828,7 @@ SELECT cast(1 as int) in (cast(1 as int), cast('2017-12-11 09:30:00' as date)) F -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"INT\", \"INT\", \"DATE\"]", @@ -2938,7 +2938,7 @@ SELECT cast(1 as bigint) in (cast(1 as bigint), cast('1' as binary)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BIGINT\", \"BIGINT\", \"BINARY\"]", @@ -2960,7 +2960,7 @@ SELECT cast(1 as bigint) in (cast(1 as bigint), cast(1 as boolean)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BIGINT\", \"BIGINT\", \"BOOLEAN\"]", @@ -2982,7 +2982,7 @@ SELECT cast(1 as bigint) in (cast(1 as bigint), cast('2017-12-11 09:30:00.0' as -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BIGINT\", \"BIGINT\", \"TIMESTAMP\"]", @@ -3004,7 +3004,7 @@ SELECT cast(1 as bigint) in (cast(1 as bigint), cast('2017-12-11 09:30:00' as da -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BIGINT\", \"BIGINT\", \"DATE\"]", @@ -3114,7 +3114,7 @@ SELECT cast(1 as float) in (cast(1 as float), cast('1' as binary)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"FLOAT\", \"FLOAT\", \"BINARY\"]", @@ -3136,7 +3136,7 @@ SELECT cast(1 as float) in (cast(1 as float), cast(1 as boolean)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"FLOAT\", \"FLOAT\", \"BOOLEAN\"]", @@ -3158,7 +3158,7 @@ SELECT cast(1 as float) in (cast(1 as float), cast('2017-12-11 09:30:00.0' as ti -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"FLOAT\", \"FLOAT\", \"TIMESTAMP\"]", @@ -3180,7 +3180,7 @@ SELECT cast(1 as float) in (cast(1 as float), cast('2017-12-11 09:30:00' as date -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"FLOAT\", \"FLOAT\", \"DATE\"]", @@ -3290,7 +3290,7 @@ SELECT cast(1 as double) in (cast(1 as double), cast('1' as binary)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DOUBLE\", \"DOUBLE\", \"BINARY\"]", @@ -3312,7 +3312,7 @@ SELECT cast(1 as double) in (cast(1 as double), cast(1 as boolean)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DOUBLE\", \"DOUBLE\", \"BOOLEAN\"]", @@ -3334,7 +3334,7 @@ SELECT cast(1 as double) in (cast(1 as double), cast('2017-12-11 09:30:00.0' as -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DOUBLE\", \"DOUBLE\", \"TIMESTAMP\"]", @@ -3356,7 +3356,7 @@ SELECT cast(1 as double) in (cast(1 as double), cast('2017-12-11 09:30:00' as da -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DOUBLE\", \"DOUBLE\", \"DATE\"]", @@ -3466,7 +3466,7 @@ SELECT cast(1 as decimal(10, 0)) in (cast(1 as decimal(10, 0)), cast('1' as bina -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DECIMAL(10,0)\", \"DECIMAL(10,0)\", \"BINARY\"]", @@ -3488,7 +3488,7 @@ SELECT cast(1 as decimal(10, 0)) in (cast(1 as decimal(10, 0)), cast(1 as boolea -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DECIMAL(10,0)\", \"DECIMAL(10,0)\", \"BOOLEAN\"]", @@ -3510,7 +3510,7 @@ SELECT cast(1 as decimal(10, 0)) in (cast(1 as decimal(10, 0)), cast('2017-12-11 -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DECIMAL(10,0)\", \"DECIMAL(10,0)\", \"TIMESTAMP\"]", @@ -3532,7 +3532,7 @@ SELECT cast(1 as decimal(10, 0)) in (cast(1 as decimal(10, 0)), cast('2017-12-11 -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DECIMAL(10,0)\", \"DECIMAL(10,0)\", \"DATE\"]", @@ -3686,7 +3686,7 @@ SELECT cast('1' as binary) in (cast('1' as binary), cast(1 as tinyint)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BINARY\", \"BINARY\", \"TINYINT\"]", @@ -3708,7 +3708,7 @@ SELECT cast('1' as binary) in (cast('1' as binary), cast(1 as smallint)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BINARY\", \"BINARY\", \"SMALLINT\"]", @@ -3730,7 +3730,7 @@ SELECT cast('1' as binary) in (cast('1' as binary), cast(1 as int)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BINARY\", \"BINARY\", \"INT\"]", @@ -3752,7 +3752,7 @@ SELECT cast('1' as binary) in (cast('1' as binary), cast(1 as bigint)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BINARY\", \"BINARY\", \"BIGINT\"]", @@ -3774,7 +3774,7 @@ SELECT cast('1' as binary) in (cast('1' as binary), cast(1 as float)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BINARY\", \"BINARY\", \"FLOAT\"]", @@ -3796,7 +3796,7 @@ SELECT cast('1' as binary) in (cast('1' as binary), cast(1 as double)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BINARY\", \"BINARY\", \"DOUBLE\"]", @@ -3818,7 +3818,7 @@ SELECT cast('1' as binary) in (cast('1' as binary), cast(1 as decimal(10, 0))) F -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BINARY\", \"BINARY\", \"DECIMAL(10,0)\"]", @@ -3862,7 +3862,7 @@ SELECT cast('1' as binary) in (cast('1' as binary), cast(1 as boolean)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BINARY\", \"BINARY\", \"BOOLEAN\"]", @@ -3884,7 +3884,7 @@ SELECT cast('1' as binary) in (cast('1' as binary), cast('2017-12-11 09:30:00.0' -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BINARY\", \"BINARY\", \"TIMESTAMP\"]", @@ -3906,7 +3906,7 @@ SELECT cast('1' as binary) in (cast('1' as binary), cast('2017-12-11 09:30:00' a -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BINARY\", \"BINARY\", \"DATE\"]", @@ -3928,7 +3928,7 @@ SELECT cast('1' as boolean) in (cast('1' as boolean), cast(1 as tinyint)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BOOLEAN\", \"BOOLEAN\", \"TINYINT\"]", @@ -3950,7 +3950,7 @@ SELECT cast('1' as boolean) in (cast('1' as boolean), cast(1 as smallint)) FROM -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BOOLEAN\", \"BOOLEAN\", \"SMALLINT\"]", @@ -3972,7 +3972,7 @@ SELECT cast('1' as boolean) in (cast('1' as boolean), cast(1 as int)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BOOLEAN\", \"BOOLEAN\", \"INT\"]", @@ -3994,7 +3994,7 @@ SELECT cast('1' as boolean) in (cast('1' as boolean), cast(1 as bigint)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BOOLEAN\", \"BOOLEAN\", \"BIGINT\"]", @@ -4016,7 +4016,7 @@ SELECT cast('1' as boolean) in (cast('1' as boolean), cast(1 as float)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BOOLEAN\", \"BOOLEAN\", \"FLOAT\"]", @@ -4038,7 +4038,7 @@ SELECT cast('1' as boolean) in (cast('1' as boolean), cast(1 as double)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BOOLEAN\", \"BOOLEAN\", \"DOUBLE\"]", @@ -4060,7 +4060,7 @@ SELECT cast('1' as boolean) in (cast('1' as boolean), cast(1 as decimal(10, 0))) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BOOLEAN\", \"BOOLEAN\", \"DECIMAL(10,0)\"]", @@ -4093,7 +4093,7 @@ SELECT cast('1' as boolean) in (cast('1' as boolean), cast('1' as binary)) FROM -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BOOLEAN\", \"BOOLEAN\", \"BINARY\"]", @@ -4126,7 +4126,7 @@ SELECT cast('1' as boolean) in (cast('1' as boolean), cast('2017-12-11 09:30:00. -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BOOLEAN\", \"BOOLEAN\", \"TIMESTAMP\"]", @@ -4148,7 +4148,7 @@ SELECT cast('1' as boolean) in (cast('1' as boolean), cast('2017-12-11 09:30:00' -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BOOLEAN\", \"BOOLEAN\", \"DATE\"]", @@ -4170,7 +4170,7 @@ SELECT cast('2017-12-12 09:30:00.0' as timestamp) in (cast('2017-12-12 09:30:00. -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"TIMESTAMP\", \"TIMESTAMP\", \"TINYINT\"]", @@ -4192,7 +4192,7 @@ SELECT cast('2017-12-12 09:30:00.0' as timestamp) in (cast('2017-12-12 09:30:00. -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"TIMESTAMP\", \"TIMESTAMP\", \"SMALLINT\"]", @@ -4214,7 +4214,7 @@ SELECT cast('2017-12-12 09:30:00.0' as timestamp) in (cast('2017-12-12 09:30:00. -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"TIMESTAMP\", \"TIMESTAMP\", \"INT\"]", @@ -4236,7 +4236,7 @@ SELECT cast('2017-12-12 09:30:00.0' as timestamp) in (cast('2017-12-12 09:30:00. -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"TIMESTAMP\", \"TIMESTAMP\", \"BIGINT\"]", @@ -4258,7 +4258,7 @@ SELECT cast('2017-12-12 09:30:00.0' as timestamp) in (cast('2017-12-12 09:30:00. -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"TIMESTAMP\", \"TIMESTAMP\", \"FLOAT\"]", @@ -4280,7 +4280,7 @@ SELECT cast('2017-12-12 09:30:00.0' as timestamp) in (cast('2017-12-12 09:30:00. -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"TIMESTAMP\", \"TIMESTAMP\", \"DOUBLE\"]", @@ -4302,7 +4302,7 @@ SELECT cast('2017-12-12 09:30:00.0' as timestamp) in (cast('2017-12-12 09:30:00. -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"TIMESTAMP\", \"TIMESTAMP\", \"DECIMAL(10,0)\"]", @@ -4335,7 +4335,7 @@ SELECT cast('2017-12-12 09:30:00.0' as timestamp) in (cast('2017-12-12 09:30:00. -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"TIMESTAMP\", \"TIMESTAMP\", \"BINARY\"]", @@ -4357,7 +4357,7 @@ SELECT cast('2017-12-12 09:30:00.0' as timestamp) in (cast('2017-12-12 09:30:00. -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"TIMESTAMP\", \"TIMESTAMP\", \"BOOLEAN\"]", @@ -4401,7 +4401,7 @@ SELECT cast('2017-12-12 09:30:00' as date) in (cast('2017-12-12 09:30:00' as dat -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DATE\", \"DATE\", \"TINYINT\"]", @@ -4423,7 +4423,7 @@ SELECT cast('2017-12-12 09:30:00' as date) in (cast('2017-12-12 09:30:00' as dat -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DATE\", \"DATE\", \"SMALLINT\"]", @@ -4445,7 +4445,7 @@ SELECT cast('2017-12-12 09:30:00' as date) in (cast('2017-12-12 09:30:00' as dat -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DATE\", \"DATE\", \"INT\"]", @@ -4467,7 +4467,7 @@ SELECT cast('2017-12-12 09:30:00' as date) in (cast('2017-12-12 09:30:00' as dat -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DATE\", \"DATE\", \"BIGINT\"]", @@ -4489,7 +4489,7 @@ SELECT cast('2017-12-12 09:30:00' as date) in (cast('2017-12-12 09:30:00' as dat -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DATE\", \"DATE\", \"FLOAT\"]", @@ -4511,7 +4511,7 @@ SELECT cast('2017-12-12 09:30:00' as date) in (cast('2017-12-12 09:30:00' as dat -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DATE\", \"DATE\", \"DOUBLE\"]", @@ -4533,7 +4533,7 @@ SELECT cast('2017-12-12 09:30:00' as date) in (cast('2017-12-12 09:30:00' as dat -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DATE\", \"DATE\", \"DECIMAL(10,0)\"]", @@ -4566,7 +4566,7 @@ SELECT cast('2017-12-12 09:30:00' as date) in (cast('2017-12-12 09:30:00' as dat -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DATE\", \"DATE\", \"BINARY\"]", @@ -4588,7 +4588,7 @@ SELECT cast('2017-12-12 09:30:00' as date) in (cast('2017-12-12 09:30:00' as dat -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DATE\", \"DATE\", \"BOOLEAN\"]", diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/typeCoercion/native/mapZipWith.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/typeCoercion/native/mapZipWith.sql.out index 13cca708b8cb..551f9ff59957 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/typeCoercion/native/mapZipWith.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/typeCoercion/native/mapZipWith.sql.out @@ -195,7 +195,7 @@ FROM various_maps -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.MAP_ZIP_WITH_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.MAP_ZIP_WITH_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "functionName" : "`map_zip_with`", @@ -219,7 +219,7 @@ FROM various_maps -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.MAP_ZIP_WITH_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.MAP_ZIP_WITH_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "functionName" : "`map_zip_with`", @@ -243,7 +243,7 @@ FROM various_maps -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.MAP_ZIP_WITH_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.MAP_ZIP_WITH_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "functionName" : "`map_zip_with`", @@ -267,7 +267,7 @@ FROM various_maps -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.MAP_ZIP_WITH_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.MAP_ZIP_WITH_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "functionName" : "`map_zip_with`", diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/typeCoercion/native/mapconcat.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/typeCoercion/native/mapconcat.sql.out index 6fee49a07667..7ae45bc0f241 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/typeCoercion/native/mapconcat.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/typeCoercion/native/mapconcat.sql.out @@ -119,7 +119,7 @@ FROM various_maps -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "(\"MAP\" or \"MAP\")", @@ -143,7 +143,7 @@ FROM various_maps -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "(\"MAP\" or \"MAP, ARRAY>\")", @@ -167,7 +167,7 @@ FROM various_maps -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "(\"MAP\" or \"MAP\")", @@ -191,7 +191,7 @@ FROM various_maps -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "(\"MAP\" or \"MAP, STRUCT>\")", @@ -215,7 +215,7 @@ FROM various_maps -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "(\"MAP, STRUCT>\" or \"MAP, ARRAY>\")", @@ -239,7 +239,7 @@ FROM various_maps -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "(\"MAP\" or \"MAP, ARRAY>\")", diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/typeCoercion/native/promoteStrings.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/typeCoercion/native/promoteStrings.sql.out index a3420c3cb063..ed449f1a7fe3 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/typeCoercion/native/promoteStrings.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/typeCoercion/native/promoteStrings.sql.out @@ -89,7 +89,7 @@ SELECT '1' + '1' FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"STRING\"", @@ -111,7 +111,7 @@ SELECT '1' + cast('1' as binary) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"BINARY\"", @@ -133,7 +133,7 @@ SELECT '1' + cast(1 as boolean) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"BOOLEAN\"", @@ -155,7 +155,7 @@ SELECT '1' + cast('2017-12-11 09:30:00.0' as timestamp) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"TIMESTAMP\"", @@ -177,7 +177,7 @@ SELECT '1' + cast('2017-12-11 09:30:00' as date) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"CAST(2017-12-11 09:30:00 AS DATE)\"", @@ -278,7 +278,7 @@ SELECT '1' - '1' FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"STRING\"", @@ -300,7 +300,7 @@ SELECT '1' - cast('1' as binary) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"BINARY\"", @@ -322,7 +322,7 @@ SELECT '1' - cast(1 as boolean) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"BOOLEAN\"", @@ -443,7 +443,7 @@ SELECT '1' * '1' FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"STRING\"", @@ -465,7 +465,7 @@ SELECT '1' * cast('1' as binary) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"BINARY\"", @@ -487,7 +487,7 @@ SELECT '1' * cast(1 as boolean) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"BOOLEAN\"", @@ -509,7 +509,7 @@ SELECT '1' * cast('2017-12-11 09:30:00.0' as timestamp) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"TIMESTAMP\"", @@ -531,7 +531,7 @@ SELECT '1' * cast('2017-12-11 09:30:00' as date) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"DATE\"", @@ -630,7 +630,7 @@ SELECT '1' / '1' FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"STRING\"", @@ -652,7 +652,7 @@ SELECT '1' / cast('1' as binary) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"BINARY\"", @@ -674,7 +674,7 @@ SELECT '1' / cast(1 as boolean) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"BOOLEAN\"", @@ -696,7 +696,7 @@ SELECT '1' / cast('2017-12-11 09:30:00.0' as timestamp) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"TIMESTAMP\"", @@ -718,7 +718,7 @@ SELECT '1' / cast('2017-12-11 09:30:00' as date) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"DATE\"", @@ -817,7 +817,7 @@ SELECT '1' % '1' FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"STRING\"", @@ -839,7 +839,7 @@ SELECT '1' % cast('1' as binary) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"BINARY\"", @@ -861,7 +861,7 @@ SELECT '1' % cast(1 as boolean) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"BOOLEAN\"", @@ -883,7 +883,7 @@ SELECT '1' % cast('2017-12-11 09:30:00.0' as timestamp) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"TIMESTAMP\"", @@ -905,7 +905,7 @@ SELECT '1' % cast('2017-12-11 09:30:00' as date) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"DATE\"", @@ -1004,7 +1004,7 @@ SELECT pmod('1', '1') FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"STRING\"", @@ -1026,7 +1026,7 @@ SELECT pmod('1', cast('1' as binary)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"BINARY\"", @@ -1048,7 +1048,7 @@ SELECT pmod('1', cast(1 as boolean)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"BOOLEAN\"", @@ -1070,7 +1070,7 @@ SELECT pmod('1', cast('2017-12-11 09:30:00.0' as timestamp)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"TIMESTAMP\"", @@ -1092,7 +1092,7 @@ SELECT pmod('1', cast('2017-12-11 09:30:00' as date)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"DATE\"", @@ -1191,7 +1191,7 @@ SELECT cast('1' as binary) + '1' FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"BINARY\"", @@ -1213,7 +1213,7 @@ SELECT cast(1 as boolean) + '1' FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"BOOLEAN\"", @@ -1235,7 +1235,7 @@ SELECT cast('2017-12-11 09:30:00.0' as timestamp) + '1' FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"TIMESTAMP\"", @@ -1257,7 +1257,7 @@ SELECT cast('2017-12-11 09:30:00' as date) + '1' FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"1\"", @@ -1358,7 +1358,7 @@ SELECT cast('1' as binary) - '1' FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"BINARY\"", @@ -1380,7 +1380,7 @@ SELECT cast(1 as boolean) - '1' FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"BOOLEAN\"", @@ -1501,7 +1501,7 @@ SELECT cast('1' as binary) * '1' FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"BINARY\"", @@ -1523,7 +1523,7 @@ SELECT cast(1 as boolean) * '1' FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"BOOLEAN\"", @@ -1545,7 +1545,7 @@ SELECT cast('2017-12-11 09:30:00.0' as timestamp) * '1' FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"TIMESTAMP\"", @@ -1567,7 +1567,7 @@ SELECT cast('2017-12-11 09:30:00' as date) * '1' FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"DATE\"", @@ -1666,7 +1666,7 @@ SELECT cast('1' as binary) / '1' FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"BINARY\"", @@ -1688,7 +1688,7 @@ SELECT cast(1 as boolean) / '1' FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"BOOLEAN\"", @@ -1710,7 +1710,7 @@ SELECT cast('2017-12-11 09:30:00.0' as timestamp) / '1' FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"TIMESTAMP\"", @@ -1732,7 +1732,7 @@ SELECT cast('2017-12-11 09:30:00' as date) / '1' FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"DATE\"", @@ -1831,7 +1831,7 @@ SELECT cast('1' as binary) % '1' FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"BINARY\"", @@ -1853,7 +1853,7 @@ SELECT cast(1 as boolean) % '1' FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"BOOLEAN\"", @@ -1875,7 +1875,7 @@ SELECT cast('2017-12-11 09:30:00.0' as timestamp) % '1' FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"TIMESTAMP\"", @@ -1897,7 +1897,7 @@ SELECT cast('2017-12-11 09:30:00' as date) % '1' FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"DATE\"", @@ -1996,7 +1996,7 @@ SELECT pmod(cast('1' as binary), '1') FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"BINARY\"", @@ -2018,7 +2018,7 @@ SELECT pmod(cast(1 as boolean), '1') FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"BOOLEAN\"", @@ -2040,7 +2040,7 @@ SELECT pmod(cast('2017-12-11 09:30:00.0' as timestamp), '1') FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"TIMESTAMP\"", @@ -2062,7 +2062,7 @@ SELECT pmod(cast('2017-12-11 09:30:00' as date), '1') FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"DATE\"", diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/typeCoercion/native/stringCastAndExpressions.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/typeCoercion/native/stringCastAndExpressions.sql.out index e57f803124ee..0498680255d1 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/typeCoercion/native/stringCastAndExpressions.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/typeCoercion/native/stringCastAndExpressions.sql.out @@ -133,7 +133,7 @@ select cast(a as array) from t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.CAST_WITHOUT_SUGGESTION", + "condition" : "DATATYPE_MISMATCH.CAST_WITHOUT_SUGGESTION", "sqlState" : "42K09", "messageParameters" : { "sqlExpr" : "\"a\"", @@ -155,7 +155,7 @@ select cast(a as struct) from t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.CAST_WITHOUT_SUGGESTION", + "condition" : "DATATYPE_MISMATCH.CAST_WITHOUT_SUGGESTION", "sqlState" : "42K09", "messageParameters" : { "sqlExpr" : "\"a\"", @@ -177,7 +177,7 @@ select cast(a as map) from t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.CAST_WITHOUT_SUGGESTION", + "condition" : "DATATYPE_MISMATCH.CAST_WITHOUT_SUGGESTION", "sqlState" : "42K09", "messageParameters" : { "sqlExpr" : "\"a\"", diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/typeCoercion/native/widenSetOperationTypes.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/typeCoercion/native/widenSetOperationTypes.sql.out index e855cdc14a92..c8694e2683a4 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/typeCoercion/native/widenSetOperationTypes.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/typeCoercion/native/widenSetOperationTypes.sql.out @@ -173,7 +173,7 @@ SELECT cast(1 as tinyint) FROM t UNION SELECT cast('2' as binary) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -198,7 +198,7 @@ SELECT cast(1 as tinyint) FROM t UNION SELECT cast(2 as boolean) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -223,7 +223,7 @@ SELECT cast(1 as tinyint) FROM t UNION SELECT cast('2017-12-11 09:30:00.0' as ti -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -248,7 +248,7 @@ SELECT cast(1 as tinyint) FROM t UNION SELECT cast('2017-12-11 09:30:00' as date -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -434,7 +434,7 @@ SELECT cast(1 as smallint) FROM t UNION SELECT cast('2' as binary) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -459,7 +459,7 @@ SELECT cast(1 as smallint) FROM t UNION SELECT cast(2 as boolean) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -484,7 +484,7 @@ SELECT cast(1 as smallint) FROM t UNION SELECT cast('2017-12-11 09:30:00.0' as t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -509,7 +509,7 @@ SELECT cast(1 as smallint) FROM t UNION SELECT cast('2017-12-11 09:30:00' as dat -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -695,7 +695,7 @@ SELECT cast(1 as int) FROM t UNION SELECT cast('2' as binary) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -720,7 +720,7 @@ SELECT cast(1 as int) FROM t UNION SELECT cast(2 as boolean) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -745,7 +745,7 @@ SELECT cast(1 as int) FROM t UNION SELECT cast('2017-12-11 09:30:00.0' as timest -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -770,7 +770,7 @@ SELECT cast(1 as int) FROM t UNION SELECT cast('2017-12-11 09:30:00' as date) FR -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -956,7 +956,7 @@ SELECT cast(1 as bigint) FROM t UNION SELECT cast('2' as binary) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -981,7 +981,7 @@ SELECT cast(1 as bigint) FROM t UNION SELECT cast(2 as boolean) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -1006,7 +1006,7 @@ SELECT cast(1 as bigint) FROM t UNION SELECT cast('2017-12-11 09:30:00.0' as tim -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -1031,7 +1031,7 @@ SELECT cast(1 as bigint) FROM t UNION SELECT cast('2017-12-11 09:30:00' as date) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -1221,7 +1221,7 @@ SELECT cast(1 as float) FROM t UNION SELECT cast('2' as binary) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -1246,7 +1246,7 @@ SELECT cast(1 as float) FROM t UNION SELECT cast(2 as boolean) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -1271,7 +1271,7 @@ SELECT cast(1 as float) FROM t UNION SELECT cast('2017-12-11 09:30:00.0' as time -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -1296,7 +1296,7 @@ SELECT cast(1 as float) FROM t UNION SELECT cast('2017-12-11 09:30:00' as date) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -1480,7 +1480,7 @@ SELECT cast(1 as double) FROM t UNION SELECT cast('2' as binary) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -1505,7 +1505,7 @@ SELECT cast(1 as double) FROM t UNION SELECT cast(2 as boolean) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -1530,7 +1530,7 @@ SELECT cast(1 as double) FROM t UNION SELECT cast('2017-12-11 09:30:00.0' as tim -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -1555,7 +1555,7 @@ SELECT cast(1 as double) FROM t UNION SELECT cast('2017-12-11 09:30:00' as date) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -1742,7 +1742,7 @@ SELECT cast(1 as decimal(10, 0)) FROM t UNION SELECT cast('2' as binary) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -1767,7 +1767,7 @@ SELECT cast(1 as decimal(10, 0)) FROM t UNION SELECT cast(2 as boolean) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -1792,7 +1792,7 @@ SELECT cast(1 as decimal(10, 0)) FROM t UNION SELECT cast('2017-12-11 09:30:00.0 -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -1817,7 +1817,7 @@ SELECT cast(1 as decimal(10, 0)) FROM t UNION SELECT cast('2017-12-11 09:30:00' -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -2086,7 +2086,7 @@ SELECT cast('1' as binary) FROM t UNION SELECT cast(2 as tinyint) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -2111,7 +2111,7 @@ SELECT cast('1' as binary) FROM t UNION SELECT cast(2 as smallint) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -2136,7 +2136,7 @@ SELECT cast('1' as binary) FROM t UNION SELECT cast(2 as int) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -2161,7 +2161,7 @@ SELECT cast('1' as binary) FROM t UNION SELECT cast(2 as bigint) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -2186,7 +2186,7 @@ SELECT cast('1' as binary) FROM t UNION SELECT cast(2 as float) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -2211,7 +2211,7 @@ SELECT cast('1' as binary) FROM t UNION SELECT cast(2 as double) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -2236,7 +2236,7 @@ SELECT cast('1' as binary) FROM t UNION SELECT cast(2 as decimal(10, 0)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -2300,7 +2300,7 @@ SELECT cast('1' as binary) FROM t UNION SELECT cast(2 as boolean) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -2325,7 +2325,7 @@ SELECT cast('1' as binary) FROM t UNION SELECT cast('2017-12-11 09:30:00.0' as t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -2350,7 +2350,7 @@ SELECT cast('1' as binary) FROM t UNION SELECT cast('2017-12-11 09:30:00' as dat -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -2375,7 +2375,7 @@ SELECT cast(1 as boolean) FROM t UNION SELECT cast(2 as tinyint) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -2400,7 +2400,7 @@ SELECT cast(1 as boolean) FROM t UNION SELECT cast(2 as smallint) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -2425,7 +2425,7 @@ SELECT cast(1 as boolean) FROM t UNION SELECT cast(2 as int) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -2450,7 +2450,7 @@ SELECT cast(1 as boolean) FROM t UNION SELECT cast(2 as bigint) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -2475,7 +2475,7 @@ SELECT cast(1 as boolean) FROM t UNION SELECT cast(2 as float) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -2500,7 +2500,7 @@ SELECT cast(1 as boolean) FROM t UNION SELECT cast(2 as double) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -2525,7 +2525,7 @@ SELECT cast(1 as boolean) FROM t UNION SELECT cast(2 as decimal(10, 0)) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -2570,7 +2570,7 @@ SELECT cast(1 as boolean) FROM t UNION SELECT cast('2' as binary) FROM t -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -2614,7 +2614,7 @@ SELECT cast(1 as boolean) FROM t UNION SELECT cast('2017-12-11 09:30:00.0' as ti -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -2639,7 +2639,7 @@ SELECT cast(1 as boolean) FROM t UNION SELECT cast('2017-12-11 09:30:00' as date -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -2664,7 +2664,7 @@ SELECT cast('2017-12-12 09:30:00.0' as timestamp) FROM t UNION SELECT cast(2 as -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -2689,7 +2689,7 @@ SELECT cast('2017-12-12 09:30:00.0' as timestamp) FROM t UNION SELECT cast(2 as -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -2714,7 +2714,7 @@ SELECT cast('2017-12-12 09:30:00.0' as timestamp) FROM t UNION SELECT cast(2 as -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -2739,7 +2739,7 @@ SELECT cast('2017-12-12 09:30:00.0' as timestamp) FROM t UNION SELECT cast(2 as -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -2764,7 +2764,7 @@ SELECT cast('2017-12-12 09:30:00.0' as timestamp) FROM t UNION SELECT cast(2 as -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -2789,7 +2789,7 @@ SELECT cast('2017-12-12 09:30:00.0' as timestamp) FROM t UNION SELECT cast(2 as -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -2814,7 +2814,7 @@ SELECT cast('2017-12-12 09:30:00.0' as timestamp) FROM t UNION SELECT cast(2 as -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -2859,7 +2859,7 @@ SELECT cast('2017-12-12 09:30:00.0' as timestamp) FROM t UNION SELECT cast('2' a -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -2884,7 +2884,7 @@ SELECT cast('2017-12-12 09:30:00.0' as timestamp) FROM t UNION SELECT cast(2 as -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -2948,7 +2948,7 @@ SELECT cast('2017-12-12 09:30:00' as date) FROM t UNION SELECT cast(2 as tinyint -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -2973,7 +2973,7 @@ SELECT cast('2017-12-12 09:30:00' as date) FROM t UNION SELECT cast(2 as smallin -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -2998,7 +2998,7 @@ SELECT cast('2017-12-12 09:30:00' as date) FROM t UNION SELECT cast(2 as int) FR -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -3023,7 +3023,7 @@ SELECT cast('2017-12-12 09:30:00' as date) FROM t UNION SELECT cast(2 as bigint) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -3048,7 +3048,7 @@ SELECT cast('2017-12-12 09:30:00' as date) FROM t UNION SELECT cast(2 as float) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -3073,7 +3073,7 @@ SELECT cast('2017-12-12 09:30:00' as date) FROM t UNION SELECT cast(2 as double) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -3098,7 +3098,7 @@ SELECT cast('2017-12-12 09:30:00' as date) FROM t UNION SELECT cast(2 as decimal -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -3143,7 +3143,7 @@ SELECT cast('2017-12-12 09:30:00' as date) FROM t UNION SELECT cast('2' as binar -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -3168,7 +3168,7 @@ SELECT cast('2017-12-12 09:30:00' as date) FROM t UNION SELECT cast(2 as boolean -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/typeCoercion/native/windowFrameCoercion.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/typeCoercion/native/windowFrameCoercion.sql.out index d516b82508de..a800e31a302b 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/typeCoercion/native/windowFrameCoercion.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/typeCoercion/native/windowFrameCoercion.sql.out @@ -278,7 +278,7 @@ SELECT COUNT(*) OVER (PARTITION BY 1 ORDER BY cast(1 as string) DESC RANGE BETWE -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.SPECIFIED_WINDOW_FRAME_UNACCEPTED_TYPE", + "condition" : "DATATYPE_MISMATCH.SPECIFIED_WINDOW_FRAME_UNACCEPTED_TYPE", "sqlState" : "42K09", "messageParameters" : { "expectedType" : "(\"NUMERIC\" or \"INTERVAL DAY TO SECOND\" or \"INTERVAL YEAR TO MONTH\" or \"INTERVAL\")", @@ -301,7 +301,7 @@ SELECT COUNT(*) OVER (PARTITION BY 1 ORDER BY cast('1' as binary) DESC RANGE BET -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.RANGE_FRAME_INVALID_TYPE", + "condition" : "DATATYPE_MISMATCH.RANGE_FRAME_INVALID_TYPE", "sqlState" : "42K09", "messageParameters" : { "orderSpecType" : "\"BINARY\"", @@ -323,7 +323,7 @@ SELECT COUNT(*) OVER (PARTITION BY 1 ORDER BY cast(1 as boolean) DESC RANGE BETW -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.SPECIFIED_WINDOW_FRAME_UNACCEPTED_TYPE", + "condition" : "DATATYPE_MISMATCH.SPECIFIED_WINDOW_FRAME_UNACCEPTED_TYPE", "sqlState" : "42K09", "messageParameters" : { "expectedType" : "(\"NUMERIC\" or \"INTERVAL DAY TO SECOND\" or \"INTERVAL YEAR TO MONTH\" or \"INTERVAL\")", @@ -346,7 +346,7 @@ SELECT COUNT(*) OVER (PARTITION BY 1 ORDER BY cast('2017-12-11 09:30:00.0' as ti -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.RANGE_FRAME_INVALID_TYPE", + "condition" : "DATATYPE_MISMATCH.RANGE_FRAME_INVALID_TYPE", "sqlState" : "42K09", "messageParameters" : { "orderSpecType" : "\"TIMESTAMP\"", diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/udf/postgreSQL/udf-aggregates_part1.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/udf/postgreSQL/udf-aggregates_part1.sql.out index ddfe742f7ea0..f25d5a473131 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/udf/postgreSQL/udf-aggregates_part1.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/udf/postgreSQL/udf-aggregates_part1.sql.out @@ -474,7 +474,7 @@ having exists (select 1 from onek b -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "UNSUPPORTED_SUBQUERY_EXPRESSION_CATEGORY.AGGREGATE_FUNCTION_MIXED_OUTER_LOCAL_REFERENCES", + "condition" : "UNSUPPORTED_SUBQUERY_EXPRESSION_CATEGORY.AGGREGATE_FUNCTION_MIXED_OUTER_LOCAL_REFERENCES", "sqlState" : "0A000", "messageParameters" : { "function" : "sum(DISTINCT (outer(a.four) + b.four))" @@ -496,7 +496,7 @@ from tenk1 o -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`o`.`unique1`", diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/udf/postgreSQL/udf-aggregates_part3.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/udf/postgreSQL/udf-aggregates_part3.sql.out index 14f9c4f9d87a..b0e382c984d0 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/udf/postgreSQL/udf-aggregates_part3.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/udf/postgreSQL/udf-aggregates_part3.sql.out @@ -4,7 +4,7 @@ select udf(max(min(unique1))) from tenk1 -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "NESTED_AGGREGATE_FUNCTION", + "condition" : "NESTED_AGGREGATE_FUNCTION", "sqlState" : "42607", "queryContext" : [ { "objectType" : "", diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/udf/postgreSQL/udf-join.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/udf/postgreSQL/udf-join.sql.out index e7bd8a9535eb..27f02a3dcdb6 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/udf/postgreSQL/udf-join.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/udf/postgreSQL/udf-join.sql.out @@ -348,7 +348,7 @@ SELECT udf('') AS `xxx`, udf(i) AS i, udf(k), udf(t) AS t -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "AMBIGUOUS_REFERENCE", + "condition" : "AMBIGUOUS_REFERENCE", "sqlState" : "42704", "messageParameters" : { "name" : "`i`", @@ -2988,7 +2988,7 @@ select * from -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "AMBIGUOUS_REFERENCE", + "condition" : "AMBIGUOUS_REFERENCE", "sqlState" : "42704", "messageParameters" : { "name" : "`f1`", @@ -3010,7 +3010,7 @@ select * from -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`y`.`f1`", @@ -3067,7 +3067,7 @@ select udf(t1.uunique1) from -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`t1`.`uunique1`", @@ -3089,7 +3089,7 @@ select udf(udf(t2.uunique1)) from -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`t2`.`uunique1`", @@ -3111,7 +3111,7 @@ select udf(uunique1) from -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`uunique1`", @@ -3435,7 +3435,7 @@ select udf(udf(f1,g)) from int4_tbl a, (select udf(udf(f1)) as g) ss -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITHOUT_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITHOUT_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`f1`" @@ -3455,7 +3455,7 @@ select udf(f1,g) from int4_tbl a, (select a.f1 as g) ss -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITHOUT_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITHOUT_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`a`.`f1`" @@ -3475,7 +3475,7 @@ select udf(udf(f1,g)) from int4_tbl a cross join (select udf(f1) as g) ss -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITHOUT_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITHOUT_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`f1`" @@ -3495,7 +3495,7 @@ select udf(f1,g) from int4_tbl a cross join (select udf(udf(a.f1)) as g) ss -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITHOUT_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITHOUT_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`a`.`f1`" diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/udf/postgreSQL/udf-select_having.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/udf/postgreSQL/udf-select_having.sql.out index ea6b1716869f..4b964130ff55 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/udf/postgreSQL/udf-select_having.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/udf/postgreSQL/udf-select_having.sql.out @@ -147,7 +147,7 @@ SELECT udf(a) FROM test_having HAVING udf(min(a)) < udf(max(a)) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "MISSING_GROUP_BY", + "condition" : "MISSING_GROUP_BY", "sqlState" : "42803", "queryContext" : [ { "objectType" : "", @@ -164,7 +164,7 @@ SELECT 1 AS one FROM test_having HAVING udf(a) > 1 -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`a`", diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/udf/postgreSQL/udf-select_implicit.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/udf/postgreSQL/udf-select_implicit.sql.out index 2dac7538cb44..53d8471af447 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/udf/postgreSQL/udf-select_implicit.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/udf/postgreSQL/udf-select_implicit.sql.out @@ -112,7 +112,7 @@ SELECT udf(count(*)) FROM test_missing_target GROUP BY udf(a) ORDER BY udf(b) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`b`", @@ -191,7 +191,7 @@ SELECT udf(c), udf(count(*)) FROM test_missing_target GROUP BY 3 -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "GROUP_BY_POS_OUT_OF_RANGE", + "condition" : "GROUP_BY_POS_OUT_OF_RANGE", "sqlState" : "42805", "messageParameters" : { "index" : "3", @@ -214,7 +214,7 @@ SELECT udf(count(*)) FROM test_missing_target x, test_missing_target y -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "AMBIGUOUS_REFERENCE", + "condition" : "AMBIGUOUS_REFERENCE", "sqlState" : "42704", "messageParameters" : { "name" : "`b`", @@ -325,7 +325,7 @@ SELECT udf(count(udf(a))) FROM test_missing_target GROUP BY udf(a) ORDER BY udf( -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`b`", @@ -389,7 +389,7 @@ SELECT udf(count(udf(x.a))) FROM test_missing_target x, test_missing_target y -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "AMBIGUOUS_REFERENCE", + "condition" : "AMBIGUOUS_REFERENCE", "sqlState" : "42704", "messageParameters" : { "name" : "`b`", @@ -430,7 +430,7 @@ SELECT udf(count(udf(b))) FROM test_missing_target x, test_missing_target y -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "AMBIGUOUS_REFERENCE", + "condition" : "AMBIGUOUS_REFERENCE", "sqlState" : "42704", "messageParameters" : { "name" : "`b`", diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/udf/udf-except-all.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/udf/udf-except-all.sql.out index 37c10b6fcd03..520ade2408b3 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/udf/udf-except-all.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/udf/udf-except-all.sql.out @@ -201,7 +201,7 @@ SELECT array(1) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -378,7 +378,7 @@ SELECT k, v FROM tab4 -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "NUM_COLUMNS_MISMATCH", + "condition" : "NUM_COLUMNS_MISMATCH", "sqlState" : "42826", "messageParameters" : { "firstNumColumns" : "1", diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/udf/udf-group-analytics.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/udf/udf-group-analytics.sql.out index fbee3e2c8c89..2d0bcea2d436 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/udf/udf-group-analytics.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/udf/udf-group-analytics.sql.out @@ -205,7 +205,7 @@ SELECT course, udf(year), GROUPING(course) FROM courseSales GROUP BY course, udf -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNSUPPORTED_GROUPING_EXPRESSION", + "condition" : "UNSUPPORTED_GROUPING_EXPRESSION", "sqlState" : "42K0E", "queryContext" : [ { "objectType" : "", @@ -222,7 +222,7 @@ SELECT course, udf(year), GROUPING_ID(course, year) FROM courseSales GROUP BY ud -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNSUPPORTED_GROUPING_EXPRESSION", + "condition" : "UNSUPPORTED_GROUPING_EXPRESSION", "sqlState" : "42K0E", "queryContext" : [ { "objectType" : "", @@ -272,7 +272,7 @@ SELECT course, udf(year) FROM courseSales GROUP BY udf(course), year HAVING GROU -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "UNSUPPORTED_GROUPING_EXPRESSION", + "condition" : "UNSUPPORTED_GROUPING_EXPRESSION", "sqlState" : "42K0E" } @@ -282,7 +282,7 @@ SELECT course, udf(udf(year)) FROM courseSales GROUP BY course, year HAVING GROU -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "UNSUPPORTED_GROUPING_EXPRESSION", + "condition" : "UNSUPPORTED_GROUPING_EXPRESSION", "sqlState" : "42K0E" } @@ -342,7 +342,7 @@ SELECT course, udf(year) FROM courseSales GROUP BY course, udf(year) ORDER BY GR -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "UNSUPPORTED_GROUPING_EXPRESSION", + "condition" : "UNSUPPORTED_GROUPING_EXPRESSION", "sqlState" : "42K0E" } @@ -352,7 +352,7 @@ SELECT course, udf(year) FROM courseSales GROUP BY course, udf(year) ORDER BY GR -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "UNSUPPORTED_GROUPING_EXPRESSION", + "condition" : "UNSUPPORTED_GROUPING_EXPRESSION", "sqlState" : "42K0E" } diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/udf/udf-group-by.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/udf/udf-group-by.sql.out index e016a8e4fab1..8ae1dccd9afb 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/udf/udf-group-by.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/udf/udf-group-by.sql.out @@ -17,7 +17,7 @@ SELECT udf(a), udf(COUNT(b)) FROM testData -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "MISSING_GROUP_BY", + "condition" : "MISSING_GROUP_BY", "sqlState" : "42803", "queryContext" : [ { "objectType" : "", @@ -58,7 +58,7 @@ SELECT udf(a), udf(COUNT(udf(b))) FROM testData GROUP BY b -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "MISSING_AGGREGATION", + "condition" : "MISSING_AGGREGATION", "sqlState" : "42803", "messageParameters" : { "expression" : "\"a\"", @@ -147,7 +147,7 @@ SELECT udf(a + 2), udf(COUNT(b)) FROM testData GROUP BY a + 1 -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "MISSING_AGGREGATION", + "condition" : "MISSING_AGGREGATION", "sqlState" : "42803", "messageParameters" : { "expression" : "\"a\"", @@ -220,7 +220,7 @@ SELECT udf(COUNT(b)) AS k FROM testData GROUP BY k -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "GROUP_BY_AGGREGATE", + "condition" : "GROUP_BY_AGGREGATE", "sqlState" : "42903", "messageParameters" : { "sqlExpr" : "CAST(udf(cast(count(b) as string)) AS BIGINT)" @@ -251,7 +251,7 @@ SELECT k AS a, udf(COUNT(udf(v))) FROM testDataHasSameNameWithAlias GROUP BY udf -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "MISSING_AGGREGATION", + "condition" : "MISSING_AGGREGATION", "sqlState" : "42803", "messageParameters" : { "expression" : "\"k\"", @@ -271,7 +271,7 @@ SELECT a AS k, udf(COUNT(udf(b))) FROM testData GROUP BY k -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`k`", @@ -379,7 +379,7 @@ SELECT udf(id) FROM range(10) HAVING id > 0 -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "MISSING_GROUP_BY", + "condition" : "MISSING_GROUP_BY", "sqlState" : "42803", "queryContext" : [ { "objectType" : "", @@ -549,7 +549,7 @@ SELECT every(udf(1)) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"udf(1)\"", @@ -573,7 +573,7 @@ SELECT some(udf(1S)) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"udf(1)\"", @@ -597,7 +597,7 @@ SELECT any(udf(1L)) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"udf(1)\"", @@ -716,7 +716,7 @@ SELECT udf(count(*)) FROM test_agg WHERE count(*) > 1L -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INVALID_WHERE_CONDITION", + "condition" : "INVALID_WHERE_CONDITION", "sqlState" : "42903", "messageParameters" : { "condition" : "\"(count(1) > 1)\"", @@ -737,7 +737,7 @@ SELECT udf(count(*)) FROM test_agg WHERE count(*) + 1L > 1L -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INVALID_WHERE_CONDITION", + "condition" : "INVALID_WHERE_CONDITION", "sqlState" : "42903", "messageParameters" : { "condition" : "\"((count(1) + 1) > 1)\"", @@ -758,7 +758,7 @@ SELECT udf(count(*)) FROM test_agg WHERE k = 1 or k = 2 or count(*) + 1L > 1L or -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INVALID_WHERE_CONDITION", + "condition" : "INVALID_WHERE_CONDITION", "sqlState" : "42903", "messageParameters" : { "condition" : "\"(((k = 1) OR (k = 2)) OR (((count(1) + 1) > 1) OR (max(k) > 1)))\"", diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/udf/udf-inline-table.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/udf/udf-inline-table.sql.out index 786b5ac49b12..d0421c9bd780 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/udf/udf-inline-table.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/udf/udf-inline-table.sql.out @@ -83,7 +83,7 @@ select udf(a), b from values ("one", rand(5)), ("two", 3.0D) as data(a, b) -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_INLINE_TABLE.CANNOT_EVALUATE_EXPRESSION_IN_INLINE_TABLE", + "condition" : "INVALID_INLINE_TABLE.CANNOT_EVALUATE_EXPRESSION_IN_INLINE_TABLE", "sqlState" : "42000", "messageParameters" : { "expr" : "\"rand(5)\"" @@ -103,7 +103,7 @@ select udf(a), udf(b) from values ("one", 2.0), ("two") as data(a, b) -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_INLINE_TABLE.NUM_COLUMNS_MISMATCH", + "condition" : "INVALID_INLINE_TABLE.NUM_COLUMNS_MISMATCH", "sqlState" : "42000", "messageParameters" : { "actualNumCols" : "1", @@ -125,7 +125,7 @@ select udf(a), udf(b) from values ("one", array(0, 1)), ("two", struct(1, 2)) as -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_INLINE_TABLE.INCOMPATIBLE_TYPES_IN_INLINE_TABLE", + "condition" : "INVALID_INLINE_TABLE.INCOMPATIBLE_TYPES_IN_INLINE_TABLE", "sqlState" : "42000", "messageParameters" : { "colName" : "`b`" @@ -145,7 +145,7 @@ select udf(a), udf(b) from values ("one"), ("two") as data(a, b) -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_INLINE_TABLE.NUM_COLUMNS_MISMATCH", + "condition" : "INVALID_INLINE_TABLE.NUM_COLUMNS_MISMATCH", "sqlState" : "42000", "messageParameters" : { "actualNumCols" : "1", @@ -167,7 +167,7 @@ select udf(a), udf(b) from values ("one", random_not_exist_func(1)), ("two", 2) -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "UNRESOLVED_ROUTINE", + "condition" : "UNRESOLVED_ROUTINE", "sqlState" : "42883", "messageParameters" : { "routineName" : "`random_not_exist_func`", @@ -188,7 +188,7 @@ select udf(a), udf(b) from values ("one", count(1)), ("two", 2) as data(a, b) -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_INLINE_TABLE.CANNOT_EVALUATE_EXPRESSION_IN_INLINE_TABLE", + "condition" : "INVALID_INLINE_TABLE.CANNOT_EVALUATE_EXPRESSION_IN_INLINE_TABLE", "sqlState" : "42000", "messageParameters" : { "expr" : "\"count(1)\"" diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/udf/udf-intersect-all.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/udf/udf-intersect-all.sql.out index 35306746932e..f3ab77da77e2 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/udf/udf-intersect-all.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/udf/udf-intersect-all.sql.out @@ -163,7 +163,7 @@ SELECT array(1), udf(2) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -190,7 +190,7 @@ SELECT udf(k), udf(v) FROM tab2 -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "NUM_COLUMNS_MISMATCH", + "condition" : "NUM_COLUMNS_MISMATCH", "sqlState" : "42826", "messageParameters" : { "firstNumColumns" : "1", diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/udf/udf-pivot.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/udf/udf-pivot.sql.out index 5cfa86309f6d..97682e31a2dd 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/udf/udf-pivot.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/udf/udf-pivot.sql.out @@ -312,7 +312,7 @@ PIVOT ( -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "_LEGACY_ERROR_TEMP_1006", + "condition" : "_LEGACY_ERROR_TEMP_1006", "messageParameters" : { "sql" : "coursesales.earnings" } @@ -330,7 +330,7 @@ PIVOT ( -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "_LEGACY_ERROR_TEMP_1006", + "condition" : "_LEGACY_ERROR_TEMP_1006", "messageParameters" : { "sql" : "__auto_generated_subquery_name.year" } @@ -348,7 +348,7 @@ PIVOT ( -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`year`", @@ -398,7 +398,7 @@ PIVOT ( -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "NESTED_AGGREGATE_FUNCTION", + "condition" : "NESTED_AGGREGATE_FUNCTION", "sqlState" : "42607", "queryContext" : [ { "objectType" : "", @@ -487,7 +487,7 @@ PIVOT ( -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "PIVOT_VALUE_DATA_TYPE_MISMATCH", + "condition" : "PIVOT_VALUE_DATA_TYPE_MISMATCH", "sqlState" : "42K09", "messageParameters" : { "pivotType" : "struct", @@ -506,7 +506,7 @@ PIVOT ( -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`s`", @@ -531,7 +531,7 @@ PIVOT ( -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "NON_LITERAL_PIVOT_VALUES", + "condition" : "NON_LITERAL_PIVOT_VALUES", "sqlState" : "42K08", "messageParameters" : { "expression" : "\"course\"" @@ -680,7 +680,7 @@ PIVOT ( -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "INCOMPARABLE_PIVOT_COLUMN", + "condition" : "INCOMPARABLE_PIVOT_COLUMN", "sqlState" : "42818", "messageParameters" : { "columnName" : "`m`" @@ -701,7 +701,7 @@ PIVOT ( -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "INCOMPARABLE_PIVOT_COLUMN", + "condition" : "INCOMPARABLE_PIVOT_COLUMN", "sqlState" : "42818", "messageParameters" : { "columnName" : "`named_struct('course', __auto_generated_subquery_name`.`course, 'm', __auto_generated_subquery_name`.`m)`" diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/udf/udf-udaf.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/udf/udf-udaf.sql.out index 248ed95df9de..9c92dadfa066 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/udf/udf-udaf.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/udf/udf-udaf.sql.out @@ -35,7 +35,7 @@ SELECT default.myDoubleAvg(udf(int_col1), udf(3)) as my_avg from t1 -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", + "condition" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", "sqlState" : "42605", "messageParameters" : { "actualNum" : "2", @@ -64,7 +64,7 @@ SELECT default.udaf1(udf(int_col1)) as udaf1, udf(default.udaf1(udf(int_col1))) -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "CANNOT_LOAD_FUNCTION_CLASS", + "condition" : "CANNOT_LOAD_FUNCTION_CLASS", "sqlState" : "46103", "messageParameters" : { "className" : "test.non.existent.udaf", diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/udf/udf-window.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/udf/udf-window.sql.out index 1d0d8974e97c..f9ff41cc81a6 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/udf/udf-window.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/udf/udf-window.sql.out @@ -70,7 +70,7 @@ ROWS BETWEEN CURRENT ROW AND 2147483648 FOLLOWING) FROM testData ORDER BY udf(ca -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.SPECIFIED_WINDOW_FRAME_UNACCEPTED_TYPE", + "condition" : "DATATYPE_MISMATCH.SPECIFIED_WINDOW_FRAME_UNACCEPTED_TYPE", "sqlState" : "42K09", "messageParameters" : { "expectedType" : "\"INT\"", @@ -221,7 +221,7 @@ ROWS BETWEEN UNBOUNDED FOLLOWING AND 1 FOLLOWING) FROM testData ORDER BY cate, u -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.SPECIFIED_WINDOW_FRAME_INVALID_BOUND", + "condition" : "DATATYPE_MISMATCH.SPECIFIED_WINDOW_FRAME_INVALID_BOUND", "sqlState" : "42K09", "messageParameters" : { "lower" : "\"UNBOUNDED FOLLOWING\"", @@ -244,7 +244,7 @@ RANGE BETWEEN CURRENT ROW AND 1 FOLLOWING) FROM testData ORDER BY cate, udf(val) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.RANGE_FRAME_WITHOUT_ORDER", + "condition" : "DATATYPE_MISMATCH.RANGE_FRAME_WITHOUT_ORDER", "sqlState" : "42K09", "messageParameters" : { "sqlExpr" : "\"(PARTITION BY udf(cate) RANGE BETWEEN CURRENT ROW AND 1 FOLLOWING)\"" @@ -265,7 +265,7 @@ RANGE BETWEEN CURRENT ROW AND 1 FOLLOWING) FROM testData ORDER BY cate, udf(val) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.RANGE_FRAME_MULTI_ORDER", + "condition" : "DATATYPE_MISMATCH.RANGE_FRAME_MULTI_ORDER", "sqlState" : "42K09", "messageParameters" : { "orderSpec" : "cast(udf(cast(val#x as string)) as int) ASC NULLS FIRST,cate#x ASC NULLS FIRST", @@ -287,7 +287,7 @@ RANGE BETWEEN CURRENT ROW AND 1 FOLLOWING) FROM testData ORDER BY cate, udf(val) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.RANGE_FRAME_INVALID_TYPE", + "condition" : "DATATYPE_MISMATCH.RANGE_FRAME_INVALID_TYPE", "sqlState" : "42K09", "messageParameters" : { "orderSpecType" : "\"TIMESTAMP\"", @@ -310,7 +310,7 @@ RANGE BETWEEN 1 FOLLOWING AND 1 PRECEDING) FROM testData ORDER BY udf(cate), val -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.SPECIFIED_WINDOW_FRAME_WRONG_COMPARISON", + "condition" : "DATATYPE_MISMATCH.SPECIFIED_WINDOW_FRAME_WRONG_COMPARISON", "sqlState" : "42K09", "messageParameters" : { "comparison" : "less than or equal", @@ -332,7 +332,7 @@ RANGE BETWEEN CURRENT ROW AND current_date PRECEDING) FROM testData ORDER BY cat -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0064", + "condition" : "_LEGACY_ERROR_TEMP_0064", "messageParameters" : { "msg" : "Frame bound value must be a literal." }, @@ -421,7 +421,7 @@ SELECT udf(val), cate, row_number() OVER(PARTITION BY cate) FROM testData ORDER -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "_LEGACY_ERROR_TEMP_1037", + "condition" : "_LEGACY_ERROR_TEMP_1037", "messageParameters" : { "wf" : "row_number()" } diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/udtf/udtf.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/udtf/udtf.sql.out index 8f42ba038876..8a26a27e0cf3 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/udtf/udtf.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/udtf/udtf.sql.out @@ -107,7 +107,7 @@ SELECT * FROM -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNSUPPORTED_SUBQUERY_EXPRESSION_CATEGORY.NON_DETERMINISTIC_LATERAL_SUBQUERIES", + "condition" : "UNSUPPORTED_SUBQUERY_EXPRESSION_CATEGORY.NON_DETERMINISTIC_LATERAL_SUBQUERIES", "sqlState" : "0A000", "messageParameters" : { "treeNode" : "LateralJoin lateral-subquery#x [], Inner\n: +- Project [count#x, total#x, last#x]\n: +- LateralJoin lateral-subquery#x [c#x], Inner\n: : +- SubqueryAlias __auto_generated_subquery_name_1\n: : +- Generate UDTFCountSumLast(outer(c#x))#x, false, [count#x, total#x, last#x]\n: : +- OneRowRelation\n: +- SubqueryAlias __auto_generated_subquery_name_0\n: +- Project [named_struct(partition_col, partition_col#x, input, input#x, partition_by_0, partition_by_0#x) AS c#x]\n: +- Sort [partition_by_0#x ASC NULLS FIRST, input#x DESC NULLS LAST], false\n: +- RepartitionByExpression [partition_by_0#x]\n: +- Project [partition_col#x, input#x, partition_col#x AS partition_by_0#x]\n: +- SubqueryAlias t2\n: +- View (`t2`, [partition_col#x, input#x])\n: +- Project [cast(partition_col#x as int) AS partition_col#x, cast(input#x as int) AS input#x]\n: +- SubqueryAlias t\n: +- LocalRelation [partition_col#x, input#x]\n+- SubqueryAlias t\n +- LocalRelation [col#x]\n" @@ -139,7 +139,7 @@ SELECT * FROM UDTFWithSinglePartition(0, TABLE(t2) WITH SINGLE PARTITION) -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "TABLE_VALUED_FUNCTION_REQUIRED_METADATA_INCOMPATIBLE_WITH_CALL", + "condition" : "TABLE_VALUED_FUNCTION_REQUIRED_METADATA_INCOMPATIBLE_WITH_CALL", "sqlState" : "22023", "messageParameters" : { "functionName" : "UDTFWithSinglePartition", @@ -161,7 +161,7 @@ SELECT * FROM UDTFWithSinglePartition(0, TABLE(t2) PARTITION BY partition_col) -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "TABLE_VALUED_FUNCTION_REQUIRED_METADATA_INCOMPATIBLE_WITH_CALL", + "condition" : "TABLE_VALUED_FUNCTION_REQUIRED_METADATA_INCOMPATIBLE_WITH_CALL", "sqlState" : "22023", "messageParameters" : { "functionName" : "UDTFWithSinglePartition", @@ -186,7 +186,7 @@ SELECT * FROM -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "TABLE_VALUED_FUNCTION_REQUIRED_METADATA_INCOMPATIBLE_WITH_CALL", + "condition" : "TABLE_VALUED_FUNCTION_REQUIRED_METADATA_INCOMPATIBLE_WITH_CALL", "sqlState" : "22023", "messageParameters" : { "functionName" : "UDTFWithSinglePartition", @@ -214,7 +214,7 @@ SELECT * FROM UDTFPartitionByOrderBy(TABLE(t2) WITH SINGLE PARTITION) -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "TABLE_VALUED_FUNCTION_REQUIRED_METADATA_INCOMPATIBLE_WITH_CALL", + "condition" : "TABLE_VALUED_FUNCTION_REQUIRED_METADATA_INCOMPATIBLE_WITH_CALL", "sqlState" : "22023", "messageParameters" : { "functionName" : "UDTFPartitionByOrderBy", @@ -236,7 +236,7 @@ SELECT * FROM UDTFPartitionByOrderBy(TABLE(t2) PARTITION BY partition_col) -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "TABLE_VALUED_FUNCTION_REQUIRED_METADATA_INCOMPATIBLE_WITH_CALL", + "condition" : "TABLE_VALUED_FUNCTION_REQUIRED_METADATA_INCOMPATIBLE_WITH_CALL", "sqlState" : "22023", "messageParameters" : { "functionName" : "UDTFPartitionByOrderBy", @@ -261,7 +261,7 @@ SELECT * FROM -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "TABLE_VALUED_FUNCTION_REQUIRED_METADATA_INCOMPATIBLE_WITH_CALL", + "condition" : "TABLE_VALUED_FUNCTION_REQUIRED_METADATA_INCOMPATIBLE_WITH_CALL", "sqlState" : "22023", "messageParameters" : { "functionName" : "UDTFPartitionByOrderBy", @@ -307,7 +307,7 @@ SELECT * FROM UDTFInvalidSelectExprParseError(TABLE(t2)) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`unparsable`", @@ -328,7 +328,7 @@ SELECT * FROM UDTFInvalidSelectExprStringValue(TABLE(t2)) -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "TABLE_VALUED_FUNCTION_FAILED_TO_ANALYZE_IN_PYTHON", + "condition" : "TABLE_VALUED_FUNCTION_FAILED_TO_ANALYZE_IN_PYTHON", "sqlState" : "38000", "messageParameters" : { "msg" : "Failed to evaluate the user-defined table function 'UDTFInvalidSelectExprStringValue' because the static 'analyze' method returned an 'AnalyzeResult' object with the 'select' field set to a value besides a list or tuple of 'SelectedColumn' objects. Please update the table function and then try the query again." @@ -348,7 +348,7 @@ SELECT * FROM UDTFInvalidComplexSelectExprMissingAlias(TABLE(t2)) -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "UDTF_INVALID_REQUESTED_SELECTED_EXPRESSION_FROM_ANALYZE_METHOD_REQUIRES_ALIAS", + "condition" : "UDTF_INVALID_REQUESTED_SELECTED_EXPRESSION_FROM_ANALYZE_METHOD_REQUIRES_ALIAS", "sqlState" : "42802", "messageParameters" : { "expression" : "(input + 1)" @@ -368,7 +368,7 @@ SELECT * FROM UDTFInvalidOrderByAscKeyword(TABLE(t2)) -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "UDTF_INVALID_ALIAS_IN_REQUESTED_ORDERING_STRING_FROM_ANALYZE_METHOD", + "condition" : "UDTF_INVALID_ALIAS_IN_REQUESTED_ORDERING_STRING_FROM_ANALYZE_METHOD", "sqlState" : "42802", "messageParameters" : { "aliasName" : "ASC" @@ -388,7 +388,7 @@ SELECT * FROM UDTFInvalidOrderByStringList(TABLE(t2)) -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "TABLE_VALUED_FUNCTION_FAILED_TO_ANALYZE_IN_PYTHON", + "condition" : "TABLE_VALUED_FUNCTION_FAILED_TO_ANALYZE_IN_PYTHON", "sqlState" : "38000", "messageParameters" : { "msg" : "Failed to evaluate the user-defined table function 'UDTFInvalidOrderByStringList' because the static 'analyze' method returned an 'AnalyzeResult' object with the 'orderBy' field set to a value besides a list or tuple of 'OrderingColumn' objects. Please update the table function and then try the query again." @@ -408,7 +408,7 @@ SELECT * FROM UDTFInvalidPartitionByAndWithSinglePartition(TABLE(t2)) -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "TABLE_VALUED_FUNCTION_REQUIRED_METADATA_INVALID", + "condition" : "TABLE_VALUED_FUNCTION_REQUIRED_METADATA_INVALID", "sqlState" : "22023", "messageParameters" : { "functionName" : "UDTFInvalidPartitionByAndWithSinglePartition", @@ -429,7 +429,7 @@ SELECT * FROM UDTFInvalidPartitionByAndWithSinglePartition(TABLE(t2) WITH SINGLE -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "TABLE_VALUED_FUNCTION_REQUIRED_METADATA_INVALID", + "condition" : "TABLE_VALUED_FUNCTION_REQUIRED_METADATA_INVALID", "sqlState" : "22023", "messageParameters" : { "functionName" : "UDTFInvalidPartitionByAndWithSinglePartition", @@ -450,7 +450,7 @@ SELECT * FROM UDTFInvalidPartitionByAndWithSinglePartition(TABLE(t2) PARTITION B -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "TABLE_VALUED_FUNCTION_REQUIRED_METADATA_INVALID", + "condition" : "TABLE_VALUED_FUNCTION_REQUIRED_METADATA_INVALID", "sqlState" : "22023", "messageParameters" : { "functionName" : "UDTFInvalidPartitionByAndWithSinglePartition", @@ -474,7 +474,7 @@ SELECT * FROM -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "TABLE_VALUED_FUNCTION_REQUIRED_METADATA_INVALID", + "condition" : "TABLE_VALUED_FUNCTION_REQUIRED_METADATA_INVALID", "sqlState" : "22023", "messageParameters" : { "functionName" : "UDTFInvalidPartitionByAndWithSinglePartition", @@ -495,7 +495,7 @@ SELECT * FROM UDTFInvalidOrderByWithoutPartitionBy(TABLE(t2)) -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "TABLE_VALUED_FUNCTION_REQUIRED_METADATA_INVALID", + "condition" : "TABLE_VALUED_FUNCTION_REQUIRED_METADATA_INVALID", "sqlState" : "22023", "messageParameters" : { "functionName" : "UDTFInvalidOrderByWithoutPartitionBy", @@ -516,7 +516,7 @@ SELECT * FROM UDTFInvalidOrderByWithoutPartitionBy(TABLE(t2) WITH SINGLE PARTITI -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "TABLE_VALUED_FUNCTION_REQUIRED_METADATA_INVALID", + "condition" : "TABLE_VALUED_FUNCTION_REQUIRED_METADATA_INVALID", "sqlState" : "22023", "messageParameters" : { "functionName" : "UDTFInvalidOrderByWithoutPartitionBy", @@ -537,7 +537,7 @@ SELECT * FROM UDTFInvalidOrderByWithoutPartitionBy(TABLE(t2) PARTITION BY partit -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "TABLE_VALUED_FUNCTION_REQUIRED_METADATA_INVALID", + "condition" : "TABLE_VALUED_FUNCTION_REQUIRED_METADATA_INVALID", "sqlState" : "22023", "messageParameters" : { "functionName" : "UDTFInvalidOrderByWithoutPartitionBy", @@ -561,7 +561,7 @@ SELECT * FROM -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "TABLE_VALUED_FUNCTION_REQUIRED_METADATA_INVALID", + "condition" : "TABLE_VALUED_FUNCTION_REQUIRED_METADATA_INVALID", "sqlState" : "22023", "messageParameters" : { "functionName" : "UDTFInvalidOrderByWithoutPartitionBy", @@ -648,7 +648,7 @@ SELECT * FROM UDTFForwardStateFromAnalyzeWithKwargs(1, 2) -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "TABLE_VALUED_FUNCTION_FAILED_TO_ANALYZE_IN_PYTHON", + "condition" : "TABLE_VALUED_FUNCTION_FAILED_TO_ANALYZE_IN_PYTHON", "sqlState" : "38000", "messageParameters" : { "msg" : "Failed to evaluate the user-defined table function 'UDTFForwardStateFromAnalyzeWithKwargs' because the function arguments did not match the expected signature of the static 'analyze' method (too many positional arguments). Please update the query so that this table function call provides arguments matching the expected signature, or else update the table function so that its static 'analyze' method accepts the provided arguments, and then try the query again." @@ -674,7 +674,7 @@ SELECT * FROM UDTFForwardStateFromAnalyzeWithKwargs(argument => 1, argument => 2 -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "DUPLICATE_ROUTINE_PARAMETER_ASSIGNMENT.DOUBLE_NAMED_ARGUMENT_REFERENCE", + "condition" : "DUPLICATE_ROUTINE_PARAMETER_ASSIGNMENT.DOUBLE_NAMED_ARGUMENT_REFERENCE", "sqlState" : "4274K", "messageParameters" : { "parameterName" : "`argument`", @@ -695,7 +695,7 @@ SELECT * FROM InvalidAnalyzeMethodWithSinglePartitionNoInputTable(argument => 1) -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "TABLE_VALUED_FUNCTION_FAILED_TO_ANALYZE_IN_PYTHON", + "condition" : "TABLE_VALUED_FUNCTION_FAILED_TO_ANALYZE_IN_PYTHON", "sqlState" : "38000", "messageParameters" : { "msg" : "Failed to evaluate the user-defined table function 'InvalidAnalyzeMethodWithSinglePartitionNoInputTable' because the static 'analyze' method returned an 'AnalyzeResult' object with the 'withSinglePartition' field set to 'true', but the function call did not provide any table argument. Please update the query so that it provides a table argument, or else update the table function so that its 'analyze' method returns an 'AnalyzeResult' object with the 'withSinglePartition' field set to 'false', and then try the query again." @@ -715,7 +715,7 @@ SELECT * FROM InvalidAnalyzeMethodWithPartitionByNoInputTable(argument => 1) -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "TABLE_VALUED_FUNCTION_FAILED_TO_ANALYZE_IN_PYTHON", + "condition" : "TABLE_VALUED_FUNCTION_FAILED_TO_ANALYZE_IN_PYTHON", "sqlState" : "38000", "messageParameters" : { "msg" : "Failed to evaluate the user-defined table function 'InvalidAnalyzeMethodWithPartitionByNoInputTable' because the static 'analyze' method returned an 'AnalyzeResult' object with the 'partitionBy' list set to non-empty, but the function call did not provide any table argument. Please update the query so that it provides a table argument, or else update the table function so that its 'analyze' method returns an 'AnalyzeResult' object with the 'partitionBy' list set to empty, and then try the query again." @@ -735,7 +735,7 @@ SELECT * FROM InvalidAnalyzeMethodReturnsNonStructTypeSchema(TABLE(t2)) -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "TABLE_VALUED_FUNCTION_FAILED_TO_ANALYZE_IN_PYTHON", + "condition" : "TABLE_VALUED_FUNCTION_FAILED_TO_ANALYZE_IN_PYTHON", "sqlState" : "38000", "messageParameters" : { "msg" : "Failed to evaluate the user-defined table function 'InvalidAnalyzeMethodReturnsNonStructTypeSchema' because the static 'analyze' method expects a result of type pyspark.sql.udtf.AnalyzeResult with a 'schema' field comprising a StructType, but the 'schema' field had the wrong type: " @@ -755,7 +755,7 @@ SELECT * FROM InvalidAnalyzeMethodWithPartitionByListOfStrings(argument => TABLE -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "TABLE_VALUED_FUNCTION_FAILED_TO_ANALYZE_IN_PYTHON", + "condition" : "TABLE_VALUED_FUNCTION_FAILED_TO_ANALYZE_IN_PYTHON", "sqlState" : "38000", "messageParameters" : { "msg" : "Failed to evaluate the user-defined table function 'InvalidAnalyzeMethodWithPartitionByListOfStrings' because the static 'analyze' method returned an 'AnalyzeResult' object with the 'partitionBy' field set to a value besides a list or tuple of 'PartitioningColumn' objects. Please update the table function and then try the query again." @@ -787,7 +787,7 @@ SELECT * FROM UDTFWithSinglePartition(1) -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "TABLE_VALUED_FUNCTION_FAILED_TO_ANALYZE_IN_PYTHON", + "condition" : "TABLE_VALUED_FUNCTION_FAILED_TO_ANALYZE_IN_PYTHON", "sqlState" : "38000", "messageParameters" : { "msg" : "Failed to evaluate the user-defined table function 'UDTFWithSinglePartition' because the function arguments did not match the expected signature of the static 'analyze' method (missing a required argument: 'input_table'). Please update the query so that this table function call provides arguments matching the expected signature, or else update the table function so that its static 'analyze' method accepts the provided arguments, and then try the query again." @@ -807,7 +807,7 @@ SELECT * FROM UDTFWithSinglePartition(1, 2, 3) -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "TABLE_VALUED_FUNCTION_FAILED_TO_ANALYZE_IN_PYTHON", + "condition" : "TABLE_VALUED_FUNCTION_FAILED_TO_ANALYZE_IN_PYTHON", "sqlState" : "38000", "messageParameters" : { "msg" : "Failed to evaluate the user-defined table function 'UDTFWithSinglePartition' because the function arguments did not match the expected signature of the static 'analyze' method (too many positional arguments). Please update the query so that this table function call provides arguments matching the expected signature, or else update the table function so that its static 'analyze' method accepts the provided arguments, and then try the query again." @@ -827,7 +827,7 @@ SELECT * FROM UDTFWithSinglePartition(1, invalid_arg_name => 2) -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "TABLE_VALUED_FUNCTION_FAILED_TO_ANALYZE_IN_PYTHON", + "condition" : "TABLE_VALUED_FUNCTION_FAILED_TO_ANALYZE_IN_PYTHON", "sqlState" : "38000", "messageParameters" : { "msg" : "Failed to evaluate the user-defined table function 'UDTFWithSinglePartition' because the function arguments did not match the expected signature of the static 'analyze' method (missing a required argument: 'input_table'). Please update the query so that this table function call provides arguments matching the expected signature, or else update the table function so that its static 'analyze' method accepts the provided arguments, and then try the query again." @@ -847,7 +847,7 @@ SELECT * FROM UDTFWithSinglePartition(1, initial_count => 2) -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "TABLE_VALUED_FUNCTION_FAILED_TO_ANALYZE_IN_PYTHON", + "condition" : "TABLE_VALUED_FUNCTION_FAILED_TO_ANALYZE_IN_PYTHON", "sqlState" : "38000", "messageParameters" : { "msg" : "Failed to evaluate the user-defined table function 'UDTFWithSinglePartition' because the function arguments did not match the expected signature of the static 'analyze' method (multiple values for argument 'initial_count'). Please update the query so that this table function call provides arguments matching the expected signature, or else update the table function so that its static 'analyze' method accepts the provided arguments, and then try the query again." @@ -867,7 +867,7 @@ SELECT * FROM UDTFWithSinglePartition(initial_count => 1, initial_count => 2) -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "DUPLICATE_ROUTINE_PARAMETER_ASSIGNMENT.DOUBLE_NAMED_ARGUMENT_REFERENCE", + "condition" : "DUPLICATE_ROUTINE_PARAMETER_ASSIGNMENT.DOUBLE_NAMED_ARGUMENT_REFERENCE", "sqlState" : "4274K", "messageParameters" : { "parameterName" : "`initial_count`", @@ -888,7 +888,7 @@ SELECT * FROM UDTFInvalidPartitionByOrderByParseError(TABLE(t2)) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`unparsable`", @@ -930,7 +930,7 @@ SELECT * FROM -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITHOUT_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITHOUT_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`unresolved_column`" diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/view-schema-binding-config.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/view-schema-binding-config.sql.out index efa221400b0b..6c308b8e1a32 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/view-schema-binding-config.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/view-schema-binding-config.sql.out @@ -16,7 +16,7 @@ CREATE OR REPLACE VIEW v WITH SCHEMA BINDING AS SELECT 1 -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "FEATURE_NOT_ENABLED", + "condition" : "FEATURE_NOT_ENABLED", "sqlState" : "56038", "messageParameters" : { "configKey" : "spark.sql.legacy.viewSchemaBindingMode", @@ -38,7 +38,7 @@ CREATE OR REPLACE VIEW v WITH SCHEMA COMPENSATION AS SELECT 1 -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "FEATURE_NOT_ENABLED", + "condition" : "FEATURE_NOT_ENABLED", "sqlState" : "56038", "messageParameters" : { "configKey" : "spark.sql.legacy.viewSchemaBindingMode", @@ -60,7 +60,7 @@ CREATE OR REPLACE VIEW v WITH SCHEMA TYPE EVOLUTION AS SELECT 1 -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "FEATURE_NOT_ENABLED", + "condition" : "FEATURE_NOT_ENABLED", "sqlState" : "56038", "messageParameters" : { "configKey" : "spark.sql.legacy.viewSchemaBindingMode", @@ -82,7 +82,7 @@ CREATE OR REPLACE VIEW v WITH SCHEMA EVOLUTION AS SELECT 1 -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "FEATURE_NOT_ENABLED", + "condition" : "FEATURE_NOT_ENABLED", "sqlState" : "56038", "messageParameters" : { "configKey" : "spark.sql.legacy.viewSchemaBindingMode", @@ -324,7 +324,7 @@ SELECT * FROM v -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "CANNOT_UP_CAST_DATATYPE", + "condition" : "CANNOT_UP_CAST_DATATYPE", "sqlState" : "42846", "messageParameters" : { "details" : "The type path of the target object is:\n\nYou can either add an explicit cast to the input data or choose a higher precision type of the field in the target object", @@ -521,7 +521,7 @@ SELECT * FROM v -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.CAST_WITHOUT_SUGGESTION", + "condition" : "DATATYPE_MISMATCH.CAST_WITHOUT_SUGGESTION", "sqlState" : "42K09", "messageParameters" : { "sqlExpr" : "\"c1\"", @@ -610,7 +610,7 @@ SELECT * FROM v -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "INCOMPATIBLE_VIEW_SCHEMA_CHANGE", + "condition" : "INCOMPATIBLE_VIEW_SCHEMA_CHANGE", "sqlState" : "51024", "messageParameters" : { "actualCols" : "[]", @@ -646,7 +646,7 @@ SELECT * FROM v -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "INCOMPATIBLE_VIEW_SCHEMA_CHANGE", + "condition" : "INCOMPATIBLE_VIEW_SCHEMA_CHANGE", "sqlState" : "51024", "messageParameters" : { "actualCols" : "[]", diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/view-schema-binding.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/view-schema-binding.sql.out index 75cae1f19d46..15fad1bcb1d3 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/view-schema-binding.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/view-schema-binding.sql.out @@ -57,7 +57,7 @@ SELECT * FROM v -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "CANNOT_UP_CAST_DATATYPE", + "condition" : "CANNOT_UP_CAST_DATATYPE", "sqlState" : "42846", "messageParameters" : { "details" : "The type path of the target object is:\n\nYou can either add an explicit cast to the input data or choose a higher precision type of the field in the target object", @@ -132,7 +132,7 @@ SELECT * FROM v -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "INCOMPATIBLE_VIEW_SCHEMA_CHANGE", + "condition" : "INCOMPATIBLE_VIEW_SCHEMA_CHANGE", "sqlState" : "51024", "messageParameters" : { "actualCols" : "[]", @@ -226,7 +226,7 @@ SELECT * FROM v -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "CANNOT_UP_CAST_DATATYPE", + "condition" : "CANNOT_UP_CAST_DATATYPE", "sqlState" : "42846", "messageParameters" : { "details" : "The type path of the target object is:\n\nYou can either add an explicit cast to the input data or choose a higher precision type of the field in the target object", diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/view-schema-compensation.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/view-schema-compensation.sql.out index 64295a6f9bc0..a897319a8d59 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/view-schema-compensation.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/view-schema-compensation.sql.out @@ -161,7 +161,7 @@ SELECT * FROM v -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.CAST_WITHOUT_SUGGESTION", + "condition" : "DATATYPE_MISMATCH.CAST_WITHOUT_SUGGESTION", "sqlState" : "42K09", "messageParameters" : { "sqlExpr" : "\"c1\"", @@ -250,7 +250,7 @@ SELECT * FROM v -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "INCOMPATIBLE_VIEW_SCHEMA_CHANGE", + "condition" : "INCOMPATIBLE_VIEW_SCHEMA_CHANGE", "sqlState" : "51024", "messageParameters" : { "actualCols" : "[]", @@ -286,7 +286,7 @@ SELECT * FROM v -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "INCOMPATIBLE_VIEW_SCHEMA_CHANGE", + "condition" : "INCOMPATIBLE_VIEW_SCHEMA_CHANGE", "sqlState" : "51024", "messageParameters" : { "actualCols" : "[]", @@ -366,7 +366,7 @@ SELECT * FROM v -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "CANNOT_UP_CAST_DATATYPE", + "condition" : "CANNOT_UP_CAST_DATATYPE", "sqlState" : "42846", "messageParameters" : { "details" : "The type path of the target object is:\n\nYou can either add an explicit cast to the input data or choose a higher precision type of the field in the target object", diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/view-schema-evolution.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/view-schema-evolution.sql.out index 258edf31d4c1..3ba5ce61099c 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/view-schema-evolution.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/view-schema-evolution.sql.out @@ -391,7 +391,7 @@ SELECT * FROM v -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "INCOMPATIBLE_VIEW_SCHEMA_CHANGE", + "condition" : "INCOMPATIBLE_VIEW_SCHEMA_CHANGE", "sqlState" : "51024", "messageParameters" : { "actualCols" : "[]", @@ -427,7 +427,7 @@ SELECT * FROM v -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "INCOMPATIBLE_VIEW_SCHEMA_CHANGE", + "condition" : "INCOMPATIBLE_VIEW_SCHEMA_CHANGE", "sqlState" : "51024", "messageParameters" : { "actualCols" : "[]", @@ -680,7 +680,7 @@ SELECT * FROM v -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "COLUMN_ALREADY_EXISTS", + "condition" : "COLUMN_ALREADY_EXISTS", "sqlState" : "42711", "messageParameters" : { "columnName" : "`c1`" diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/view-schema-type-evolution.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/view-schema-type-evolution.sql.out index 95aa35d59fdc..ac15d21a06b6 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/view-schema-type-evolution.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/view-schema-type-evolution.sql.out @@ -191,7 +191,7 @@ SELECT * FROM v -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "INCOMPATIBLE_VIEW_SCHEMA_CHANGE", + "condition" : "INCOMPATIBLE_VIEW_SCHEMA_CHANGE", "sqlState" : "51024", "messageParameters" : { "actualCols" : "[]", @@ -227,7 +227,7 @@ SELECT * FROM v -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "INCOMPATIBLE_VIEW_SCHEMA_CHANGE", + "condition" : "INCOMPATIBLE_VIEW_SCHEMA_CHANGE", "sqlState" : "51024", "messageParameters" : { "actualCols" : "[]", diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/window.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/window.sql.out index 77dc2f3ba273..f8a03652c02b 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/window.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/window.sql.out @@ -142,7 +142,7 @@ ROWS BETWEEN CURRENT ROW AND 2147483648 FOLLOWING) FROM testData ORDER BY cate, -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.SPECIFIED_WINDOW_FRAME_UNACCEPTED_TYPE", + "condition" : "DATATYPE_MISMATCH.SPECIFIED_WINDOW_FRAME_UNACCEPTED_TYPE", "sqlState" : "42K09", "messageParameters" : { "expectedType" : "\"INT\"", @@ -378,7 +378,7 @@ ORDER BY cate, val_date -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.RANGE_FRAME_INVALID_TYPE", + "condition" : "DATATYPE_MISMATCH.RANGE_FRAME_INVALID_TYPE", "sqlState" : "42K09", "messageParameters" : { "orderSpecType" : "\"DATE\"", @@ -418,7 +418,7 @@ ROWS BETWEEN UNBOUNDED FOLLOWING AND 1 FOLLOWING) FROM testData ORDER BY cate, v -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.SPECIFIED_WINDOW_FRAME_INVALID_BOUND", + "condition" : "DATATYPE_MISMATCH.SPECIFIED_WINDOW_FRAME_INVALID_BOUND", "sqlState" : "42K09", "messageParameters" : { "lower" : "\"UNBOUNDED FOLLOWING\"", @@ -441,7 +441,7 @@ RANGE BETWEEN CURRENT ROW AND 1 FOLLOWING) FROM testData ORDER BY cate, val -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.RANGE_FRAME_WITHOUT_ORDER", + "condition" : "DATATYPE_MISMATCH.RANGE_FRAME_WITHOUT_ORDER", "sqlState" : "42K09", "messageParameters" : { "sqlExpr" : "\"(PARTITION BY cate RANGE BETWEEN CURRENT ROW AND 1 FOLLOWING)\"" @@ -462,7 +462,7 @@ RANGE BETWEEN CURRENT ROW AND 1 FOLLOWING) FROM testData ORDER BY cate, val -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.RANGE_FRAME_MULTI_ORDER", + "condition" : "DATATYPE_MISMATCH.RANGE_FRAME_MULTI_ORDER", "sqlState" : "42K09", "messageParameters" : { "orderSpec" : "val#x ASC NULLS FIRST,cate#x ASC NULLS FIRST", @@ -484,7 +484,7 @@ RANGE BETWEEN CURRENT ROW AND 1 FOLLOWING) FROM testData ORDER BY cate, val -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.RANGE_FRAME_INVALID_TYPE", + "condition" : "DATATYPE_MISMATCH.RANGE_FRAME_INVALID_TYPE", "sqlState" : "42K09", "messageParameters" : { "orderSpecType" : "\"TIMESTAMP\"", @@ -507,7 +507,7 @@ RANGE BETWEEN 1 FOLLOWING AND 1 PRECEDING) FROM testData ORDER BY cate, val -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.SPECIFIED_WINDOW_FRAME_WRONG_COMPARISON", + "condition" : "DATATYPE_MISMATCH.SPECIFIED_WINDOW_FRAME_WRONG_COMPARISON", "sqlState" : "42K09", "messageParameters" : { "comparison" : "less than or equal", @@ -529,7 +529,7 @@ RANGE BETWEEN CURRENT ROW AND current_date PRECEDING) FROM testData ORDER BY cat -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0064", + "condition" : "_LEGACY_ERROR_TEMP_0064", "messageParameters" : { "msg" : "Frame bound value must be a literal." }, @@ -616,7 +616,7 @@ SELECT val, cate, row_number() OVER(PARTITION BY cate) FROM testData ORDER BY ca -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "_LEGACY_ERROR_TEMP_1037", + "condition" : "_LEGACY_ERROR_TEMP_1037", "messageParameters" : { "wf" : "row_number()" } @@ -693,7 +693,7 @@ FROM testData ORDER BY cate, val -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "_LEGACY_ERROR_TEMP_1030" + "condition" : "_LEGACY_ERROR_TEMP_1030" } @@ -967,7 +967,7 @@ ORDER BY salary DESC -- !query analysis org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_SQL_SYNTAX.REPETITIVE_WINDOW_DEFINITION", + "condition" : "INVALID_SQL_SYNTAX.REPETITIVE_WINDOW_DEFINITION", "sqlState" : "42000", "messageParameters" : { "windowName" : "`w`" @@ -1260,7 +1260,7 @@ FROM -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "MISSING_WINDOW_SPECIFICATION", + "condition" : "MISSING_WINDOW_SPECIFICATION", "sqlState" : "42P20", "messageParameters" : { "docroot" : "https://spark.apache.org/docs/latest", @@ -1277,7 +1277,7 @@ FROM -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "MISSING_WINDOW_SPECIFICATION", + "condition" : "MISSING_WINDOW_SPECIFICATION", "sqlState" : "42P20", "messageParameters" : { "docroot" : "https://spark.apache.org/docs/latest", @@ -1437,7 +1437,7 @@ SELECT *, mean(val_double) over (partition BY val ORDER BY val_date RANGE INTERV -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.RANGE_FRAME_INVALID_TYPE", + "condition" : "DATATYPE_MISMATCH.RANGE_FRAME_INVALID_TYPE", "sqlState" : "42K09", "messageParameters" : { "orderSpecType" : "\"DATE\"", @@ -1459,7 +1459,7 @@ SELECT *, mean(val_double) over (partition BY val ORDER BY val_date RANGE DATE ' -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.SPECIFIED_WINDOW_FRAME_UNACCEPTED_TYPE", + "condition" : "DATATYPE_MISMATCH.SPECIFIED_WINDOW_FRAME_UNACCEPTED_TYPE", "sqlState" : "42K09", "messageParameters" : { "expectedType" : "(\"NUMERIC\" or \"INTERVAL DAY TO SECOND\" or \"INTERVAL YEAR TO MONTH\" or \"INTERVAL\")", diff --git a/sql/core/src/test/resources/sql-tests/analyzer-results/xml-functions.sql.out b/sql/core/src/test/resources/sql-tests/analyzer-results/xml-functions.sql.out index e0a249e4cf3f..0536c25e7277 100644 --- a/sql/core/src/test/resources/sql-tests/analyzer-results/xml-functions.sql.out +++ b/sql/core/src/test/resources/sql-tests/analyzer-results/xml-functions.sql.out @@ -18,7 +18,7 @@ select to_xml(array(named_struct('a', 1, 'b', 2))) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"array(named_struct(a, 1, b, 2))\"", @@ -42,7 +42,7 @@ select to_xml(map('a', 1)) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"map(a, 1)\"", @@ -66,7 +66,7 @@ select to_xml(named_struct('a', 1, 'b', 2), named_struct('mode', 'PERMISSIVE')) -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_OPTIONS.NON_MAP_FUNCTION", + "condition" : "INVALID_OPTIONS.NON_MAP_FUNCTION", "sqlState" : "42K06", "queryContext" : [ { "objectType" : "", @@ -83,7 +83,7 @@ select to_xml(named_struct('a', 1, 'b', 2), map('mode', 1)) -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_OPTIONS.NON_STRING_TYPE", + "condition" : "INVALID_OPTIONS.NON_STRING_TYPE", "sqlState" : "42K06", "messageParameters" : { "mapType" : "\"MAP\"" @@ -103,7 +103,7 @@ select to_xml() -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", + "condition" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", "sqlState" : "42605", "messageParameters" : { "actualNum" : "0", @@ -140,7 +140,7 @@ select from_xml('

1

', 1) -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_SCHEMA.NON_STRING_LITERAL", + "condition" : "INVALID_SCHEMA.NON_STRING_LITERAL", "sqlState" : "42K07", "messageParameters" : { "inputSchema" : "\"1\"" @@ -160,7 +160,7 @@ select from_xml('

1

', 'a InvalidType') -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'InvalidType'", @@ -181,7 +181,7 @@ select from_xml('

1

', 'a INT', named_struct('mode', 'PERMISSIVE')) -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_OPTIONS.NON_MAP_FUNCTION", + "condition" : "INVALID_OPTIONS.NON_MAP_FUNCTION", "sqlState" : "42K06", "queryContext" : [ { "objectType" : "", @@ -198,7 +198,7 @@ select from_xml('

1

', 'a INT', map('mode', 1)) -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_OPTIONS.NON_STRING_TYPE", + "condition" : "INVALID_OPTIONS.NON_STRING_TYPE", "sqlState" : "42K06", "messageParameters" : { "mapType" : "\"MAP\"" @@ -218,7 +218,7 @@ select from_xml() -- !query analysis org.apache.spark.sql.AnalysisException { - "errorClass" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", + "condition" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", "sqlState" : "42605", "messageParameters" : { "actualNum" : "0", @@ -340,7 +340,7 @@ select schema_of_xml(null) -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_NULL", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_NULL", "sqlState" : "42K09", "messageParameters" : { "exprName" : "xml", @@ -369,7 +369,7 @@ SELECT schema_of_xml(xmlField) FROM xmlTable -- !query analysis org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.NON_FOLDABLE_INPUT", + "condition" : "DATATYPE_MISMATCH.NON_FOLDABLE_INPUT", "sqlState" : "42K09", "messageParameters" : { "inputExpr" : "\"xmlField\"", diff --git a/sql/core/src/test/resources/sql-tests/results/ansi/cast.sql.out b/sql/core/src/test/resources/sql-tests/results/ansi/cast.sql.out new file mode 100644 index 000000000000..4399570b3076 --- /dev/null +++ b/sql/core/src/test/resources/sql-tests/results/ansi/cast.sql.out @@ -0,0 +1,1999 @@ +-- Automatically generated by SQLQueryTestSuite +-- !query +SELECT CAST('1.23' AS int) +-- !query schema +struct<> +-- !query output +org.apache.spark.SparkNumberFormatException +{ + "condition" : "CAST_INVALID_INPUT", + "sqlState" : "22018", + "messageParameters" : { + "expression" : "'1.23'", + "sourceType" : "\"STRING\"", + "targetType" : "\"INT\"" + }, + "queryContext" : [ { + "objectType" : "", + "objectName" : "", + "startIndex" : 8, + "stopIndex" : 26, + "fragment" : "CAST('1.23' AS int)" + } ] +} + + +-- !query +SELECT CAST('1.23' AS long) +-- !query schema +struct<> +-- !query output +org.apache.spark.SparkNumberFormatException +{ + "condition" : "CAST_INVALID_INPUT", + "sqlState" : "22018", + "messageParameters" : { + "expression" : "'1.23'", + "sourceType" : "\"STRING\"", + "targetType" : "\"BIGINT\"" + }, + "queryContext" : [ { + "objectType" : "", + "objectName" : "", + "startIndex" : 8, + "stopIndex" : 27, + "fragment" : "CAST('1.23' AS long)" + } ] +} + + +-- !query +SELECT CAST('-4.56' AS int) +-- !query schema +struct<> +-- !query output +org.apache.spark.SparkNumberFormatException +{ + "condition" : "CAST_INVALID_INPUT", + "sqlState" : "22018", + "messageParameters" : { + "expression" : "'-4.56'", + "sourceType" : "\"STRING\"", + "targetType" : "\"INT\"" + }, + "queryContext" : [ { + "objectType" : "", + "objectName" : "", + "startIndex" : 8, + "stopIndex" : 27, + "fragment" : "CAST('-4.56' AS int)" + } ] +} + + +-- !query +SELECT CAST('-4.56' AS long) +-- !query schema +struct<> +-- !query output +org.apache.spark.SparkNumberFormatException +{ + "condition" : "CAST_INVALID_INPUT", + "sqlState" : "22018", + "messageParameters" : { + "expression" : "'-4.56'", + "sourceType" : "\"STRING\"", + "targetType" : "\"BIGINT\"" + }, + "queryContext" : [ { + "objectType" : "", + "objectName" : "", + "startIndex" : 8, + "stopIndex" : 28, + "fragment" : "CAST('-4.56' AS long)" + } ] +} + + +-- !query +SELECT CAST('abc' AS int) +-- !query schema +struct<> +-- !query output +org.apache.spark.SparkNumberFormatException +{ + "condition" : "CAST_INVALID_INPUT", + "sqlState" : "22018", + "messageParameters" : { + "expression" : "'abc'", + "sourceType" : "\"STRING\"", + "targetType" : "\"INT\"" + }, + "queryContext" : [ { + "objectType" : "", + "objectName" : "", + "startIndex" : 8, + "stopIndex" : 25, + "fragment" : "CAST('abc' AS int)" + } ] +} + + +-- !query +SELECT CAST('abc' AS long) +-- !query schema +struct<> +-- !query output +org.apache.spark.SparkNumberFormatException +{ + "condition" : "CAST_INVALID_INPUT", + "sqlState" : "22018", + "messageParameters" : { + "expression" : "'abc'", + "sourceType" : "\"STRING\"", + "targetType" : "\"BIGINT\"" + }, + "queryContext" : [ { + "objectType" : "", + "objectName" : "", + "startIndex" : 8, + "stopIndex" : 26, + "fragment" : "CAST('abc' AS long)" + } ] +} + + +-- !query +SELECT CAST('abc' AS float) +-- !query schema +struct<> +-- !query output +org.apache.spark.SparkNumberFormatException +{ + "condition" : "CAST_INVALID_INPUT", + "sqlState" : "22018", + "messageParameters" : { + "expression" : "'abc'", + "sourceType" : "\"STRING\"", + "targetType" : "\"FLOAT\"" + }, + "queryContext" : [ { + "objectType" : "", + "objectName" : "", + "startIndex" : 8, + "stopIndex" : 27, + "fragment" : "CAST('abc' AS float)" + } ] +} + + +-- !query +SELECT CAST('abc' AS double) +-- !query schema +struct<> +-- !query output +org.apache.spark.SparkNumberFormatException +{ + "condition" : "CAST_INVALID_INPUT", + "sqlState" : "22018", + "messageParameters" : { + "expression" : "'abc'", + "sourceType" : "\"STRING\"", + "targetType" : "\"DOUBLE\"" + }, + "queryContext" : [ { + "objectType" : "", + "objectName" : "", + "startIndex" : 8, + "stopIndex" : 28, + "fragment" : "CAST('abc' AS double)" + } ] +} + + +-- !query +SELECT CAST('1234567890123' AS int) +-- !query schema +struct<> +-- !query output +org.apache.spark.SparkNumberFormatException +{ + "condition" : "CAST_INVALID_INPUT", + "sqlState" : "22018", + "messageParameters" : { + "expression" : "'1234567890123'", + "sourceType" : "\"STRING\"", + "targetType" : "\"INT\"" + }, + "queryContext" : [ { + "objectType" : "", + "objectName" : "", + "startIndex" : 8, + "stopIndex" : 35, + "fragment" : "CAST('1234567890123' AS int)" + } ] +} + + +-- !query +SELECT CAST('12345678901234567890123' AS long) +-- !query schema +struct<> +-- !query output +org.apache.spark.SparkNumberFormatException +{ + "condition" : "CAST_INVALID_INPUT", + "sqlState" : "22018", + "messageParameters" : { + "expression" : "'12345678901234567890123'", + "sourceType" : "\"STRING\"", + "targetType" : "\"BIGINT\"" + }, + "queryContext" : [ { + "objectType" : "", + "objectName" : "", + "startIndex" : 8, + "stopIndex" : 46, + "fragment" : "CAST('12345678901234567890123' AS long)" + } ] +} + + +-- !query +SELECT CAST('' AS int) +-- !query schema +struct<> +-- !query output +org.apache.spark.SparkNumberFormatException +{ + "condition" : "CAST_INVALID_INPUT", + "sqlState" : "22018", + "messageParameters" : { + "expression" : "''", + "sourceType" : "\"STRING\"", + "targetType" : "\"INT\"" + }, + "queryContext" : [ { + "objectType" : "", + "objectName" : "", + "startIndex" : 8, + "stopIndex" : 22, + "fragment" : "CAST('' AS int)" + } ] +} + + +-- !query +SELECT CAST('' AS long) +-- !query schema +struct<> +-- !query output +org.apache.spark.SparkNumberFormatException +{ + "condition" : "CAST_INVALID_INPUT", + "sqlState" : "22018", + "messageParameters" : { + "expression" : "''", + "sourceType" : "\"STRING\"", + "targetType" : "\"BIGINT\"" + }, + "queryContext" : [ { + "objectType" : "", + "objectName" : "", + "startIndex" : 8, + "stopIndex" : 23, + "fragment" : "CAST('' AS long)" + } ] +} + + +-- !query +SELECT CAST('' AS float) +-- !query schema +struct<> +-- !query output +org.apache.spark.SparkNumberFormatException +{ + "condition" : "CAST_INVALID_INPUT", + "sqlState" : "22018", + "messageParameters" : { + "expression" : "''", + "sourceType" : "\"STRING\"", + "targetType" : "\"FLOAT\"" + }, + "queryContext" : [ { + "objectType" : "", + "objectName" : "", + "startIndex" : 8, + "stopIndex" : 24, + "fragment" : "CAST('' AS float)" + } ] +} + + +-- !query +SELECT CAST('' AS double) +-- !query schema +struct<> +-- !query output +org.apache.spark.SparkNumberFormatException +{ + "condition" : "CAST_INVALID_INPUT", + "sqlState" : "22018", + "messageParameters" : { + "expression" : "''", + "sourceType" : "\"STRING\"", + "targetType" : "\"DOUBLE\"" + }, + "queryContext" : [ { + "objectType" : "", + "objectName" : "", + "startIndex" : 8, + "stopIndex" : 25, + "fragment" : "CAST('' AS double)" + } ] +} + + +-- !query +SELECT CAST(NULL AS int) +-- !query schema +struct +-- !query output +NULL + + +-- !query +SELECT CAST(NULL AS long) +-- !query schema +struct +-- !query output +NULL + + +-- !query +SELECT CAST('123.a' AS int) +-- !query schema +struct<> +-- !query output +org.apache.spark.SparkNumberFormatException +{ + "condition" : "CAST_INVALID_INPUT", + "sqlState" : "22018", + "messageParameters" : { + "expression" : "'123.a'", + "sourceType" : "\"STRING\"", + "targetType" : "\"INT\"" + }, + "queryContext" : [ { + "objectType" : "", + "objectName" : "", + "startIndex" : 8, + "stopIndex" : 27, + "fragment" : "CAST('123.a' AS int)" + } ] +} + + +-- !query +SELECT CAST('123.a' AS long) +-- !query schema +struct<> +-- !query output +org.apache.spark.SparkNumberFormatException +{ + "condition" : "CAST_INVALID_INPUT", + "sqlState" : "22018", + "messageParameters" : { + "expression" : "'123.a'", + "sourceType" : "\"STRING\"", + "targetType" : "\"BIGINT\"" + }, + "queryContext" : [ { + "objectType" : "", + "objectName" : "", + "startIndex" : 8, + "stopIndex" : 28, + "fragment" : "CAST('123.a' AS long)" + } ] +} + + +-- !query +SELECT CAST('123.a' AS float) +-- !query schema +struct<> +-- !query output +org.apache.spark.SparkNumberFormatException +{ + "condition" : "CAST_INVALID_INPUT", + "sqlState" : "22018", + "messageParameters" : { + "expression" : "'123.a'", + "sourceType" : "\"STRING\"", + "targetType" : "\"FLOAT\"" + }, + "queryContext" : [ { + "objectType" : "", + "objectName" : "", + "startIndex" : 8, + "stopIndex" : 29, + "fragment" : "CAST('123.a' AS float)" + } ] +} + + +-- !query +SELECT CAST('123.a' AS double) +-- !query schema +struct<> +-- !query output +org.apache.spark.SparkNumberFormatException +{ + "condition" : "CAST_INVALID_INPUT", + "sqlState" : "22018", + "messageParameters" : { + "expression" : "'123.a'", + "sourceType" : "\"STRING\"", + "targetType" : "\"DOUBLE\"" + }, + "queryContext" : [ { + "objectType" : "", + "objectName" : "", + "startIndex" : 8, + "stopIndex" : 30, + "fragment" : "CAST('123.a' AS double)" + } ] +} + + +-- !query +SELECT CAST('-2147483648' AS int) +-- !query schema +struct +-- !query output +-2147483648 + + +-- !query +SELECT CAST('-2147483649' AS int) +-- !query schema +struct<> +-- !query output +org.apache.spark.SparkNumberFormatException +{ + "condition" : "CAST_INVALID_INPUT", + "sqlState" : "22018", + "messageParameters" : { + "expression" : "'-2147483649'", + "sourceType" : "\"STRING\"", + "targetType" : "\"INT\"" + }, + "queryContext" : [ { + "objectType" : "", + "objectName" : "", + "startIndex" : 8, + "stopIndex" : 33, + "fragment" : "CAST('-2147483649' AS int)" + } ] +} + + +-- !query +SELECT CAST('2147483647' AS int) +-- !query schema +struct +-- !query output +2147483647 + + +-- !query +SELECT CAST('2147483648' AS int) +-- !query schema +struct<> +-- !query output +org.apache.spark.SparkNumberFormatException +{ + "condition" : "CAST_INVALID_INPUT", + "sqlState" : "22018", + "messageParameters" : { + "expression" : "'2147483648'", + "sourceType" : "\"STRING\"", + "targetType" : "\"INT\"" + }, + "queryContext" : [ { + "objectType" : "", + "objectName" : "", + "startIndex" : 8, + "stopIndex" : 32, + "fragment" : "CAST('2147483648' AS int)" + } ] +} + + +-- !query +SELECT CAST('-9223372036854775808' AS long) +-- !query schema +struct +-- !query output +-9223372036854775808 + + +-- !query +SELECT CAST('-9223372036854775809' AS long) +-- !query schema +struct<> +-- !query output +org.apache.spark.SparkNumberFormatException +{ + "condition" : "CAST_INVALID_INPUT", + "sqlState" : "22018", + "messageParameters" : { + "expression" : "'-9223372036854775809'", + "sourceType" : "\"STRING\"", + "targetType" : "\"BIGINT\"" + }, + "queryContext" : [ { + "objectType" : "", + "objectName" : "", + "startIndex" : 8, + "stopIndex" : 43, + "fragment" : "CAST('-9223372036854775809' AS long)" + } ] +} + + +-- !query +SELECT CAST('9223372036854775807' AS long) +-- !query schema +struct +-- !query output +9223372036854775807 + + +-- !query +SELECT CAST('9223372036854775808' AS long) +-- !query schema +struct<> +-- !query output +org.apache.spark.SparkNumberFormatException +{ + "condition" : "CAST_INVALID_INPUT", + "sqlState" : "22018", + "messageParameters" : { + "expression" : "'9223372036854775808'", + "sourceType" : "\"STRING\"", + "targetType" : "\"BIGINT\"" + }, + "queryContext" : [ { + "objectType" : "", + "objectName" : "", + "startIndex" : 8, + "stopIndex" : 42, + "fragment" : "CAST('9223372036854775808' AS long)" + } ] +} + + +-- !query +SELECT HEX(CAST('abc' AS binary)) +-- !query schema +struct +-- !query output +616263 + + +-- !query +SELECT HEX(CAST(CAST(123 AS byte) AS binary)) +-- !query schema +struct<> +-- !query output +org.apache.spark.sql.catalyst.ExtendedAnalysisException +{ + "condition" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", + "sqlState" : "42K09", + "messageParameters" : { + "config" : "\"spark.sql.ansi.enabled\"", + "configVal" : "'false'", + "sqlExpr" : "\"CAST(CAST(123 AS TINYINT) AS BINARY)\"", + "srcType" : "\"TINYINT\"", + "targetType" : "\"BINARY\"" + }, + "queryContext" : [ { + "objectType" : "", + "objectName" : "", + "startIndex" : 12, + "stopIndex" : 44, + "fragment" : "CAST(CAST(123 AS byte) AS binary)" + } ] +} + + +-- !query +SELECT HEX(CAST(CAST(-123 AS byte) AS binary)) +-- !query schema +struct<> +-- !query output +org.apache.spark.sql.catalyst.ExtendedAnalysisException +{ + "condition" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", + "sqlState" : "42K09", + "messageParameters" : { + "config" : "\"spark.sql.ansi.enabled\"", + "configVal" : "'false'", + "sqlExpr" : "\"CAST(CAST(-123 AS TINYINT) AS BINARY)\"", + "srcType" : "\"TINYINT\"", + "targetType" : "\"BINARY\"" + }, + "queryContext" : [ { + "objectType" : "", + "objectName" : "", + "startIndex" : 12, + "stopIndex" : 45, + "fragment" : "CAST(CAST(-123 AS byte) AS binary)" + } ] +} + + +-- !query +SELECT HEX(CAST(123S AS binary)) +-- !query schema +struct<> +-- !query output +org.apache.spark.sql.catalyst.ExtendedAnalysisException +{ + "condition" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", + "sqlState" : "42K09", + "messageParameters" : { + "config" : "\"spark.sql.ansi.enabled\"", + "configVal" : "'false'", + "sqlExpr" : "\"CAST(123 AS BINARY)\"", + "srcType" : "\"SMALLINT\"", + "targetType" : "\"BINARY\"" + }, + "queryContext" : [ { + "objectType" : "", + "objectName" : "", + "startIndex" : 12, + "stopIndex" : 31, + "fragment" : "CAST(123S AS binary)" + } ] +} + + +-- !query +SELECT HEX(CAST(-123S AS binary)) +-- !query schema +struct<> +-- !query output +org.apache.spark.sql.catalyst.ExtendedAnalysisException +{ + "condition" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", + "sqlState" : "42K09", + "messageParameters" : { + "config" : "\"spark.sql.ansi.enabled\"", + "configVal" : "'false'", + "sqlExpr" : "\"CAST(-123 AS BINARY)\"", + "srcType" : "\"SMALLINT\"", + "targetType" : "\"BINARY\"" + }, + "queryContext" : [ { + "objectType" : "", + "objectName" : "", + "startIndex" : 12, + "stopIndex" : 32, + "fragment" : "CAST(-123S AS binary)" + } ] +} + + +-- !query +SELECT HEX(CAST(123 AS binary)) +-- !query schema +struct<> +-- !query output +org.apache.spark.sql.catalyst.ExtendedAnalysisException +{ + "condition" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", + "sqlState" : "42K09", + "messageParameters" : { + "config" : "\"spark.sql.ansi.enabled\"", + "configVal" : "'false'", + "sqlExpr" : "\"CAST(123 AS BINARY)\"", + "srcType" : "\"INT\"", + "targetType" : "\"BINARY\"" + }, + "queryContext" : [ { + "objectType" : "", + "objectName" : "", + "startIndex" : 12, + "stopIndex" : 30, + "fragment" : "CAST(123 AS binary)" + } ] +} + + +-- !query +SELECT HEX(CAST(-123 AS binary)) +-- !query schema +struct<> +-- !query output +org.apache.spark.sql.catalyst.ExtendedAnalysisException +{ + "condition" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", + "sqlState" : "42K09", + "messageParameters" : { + "config" : "\"spark.sql.ansi.enabled\"", + "configVal" : "'false'", + "sqlExpr" : "\"CAST(-123 AS BINARY)\"", + "srcType" : "\"INT\"", + "targetType" : "\"BINARY\"" + }, + "queryContext" : [ { + "objectType" : "", + "objectName" : "", + "startIndex" : 12, + "stopIndex" : 31, + "fragment" : "CAST(-123 AS binary)" + } ] +} + + +-- !query +SELECT HEX(CAST(123L AS binary)) +-- !query schema +struct<> +-- !query output +org.apache.spark.sql.catalyst.ExtendedAnalysisException +{ + "condition" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", + "sqlState" : "42K09", + "messageParameters" : { + "config" : "\"spark.sql.ansi.enabled\"", + "configVal" : "'false'", + "sqlExpr" : "\"CAST(123 AS BINARY)\"", + "srcType" : "\"BIGINT\"", + "targetType" : "\"BINARY\"" + }, + "queryContext" : [ { + "objectType" : "", + "objectName" : "", + "startIndex" : 12, + "stopIndex" : 31, + "fragment" : "CAST(123L AS binary)" + } ] +} + + +-- !query +SELECT HEX(CAST(-123L AS binary)) +-- !query schema +struct<> +-- !query output +org.apache.spark.sql.catalyst.ExtendedAnalysisException +{ + "condition" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", + "sqlState" : "42K09", + "messageParameters" : { + "config" : "\"spark.sql.ansi.enabled\"", + "configVal" : "'false'", + "sqlExpr" : "\"CAST(-123 AS BINARY)\"", + "srcType" : "\"BIGINT\"", + "targetType" : "\"BINARY\"" + }, + "queryContext" : [ { + "objectType" : "", + "objectName" : "", + "startIndex" : 12, + "stopIndex" : 32, + "fragment" : "CAST(-123L AS binary)" + } ] +} + + +-- !query +DESC FUNCTION boolean +-- !query schema +struct +-- !query output +Class: org.apache.spark.sql.catalyst.expressions.Cast +Function: boolean +Usage: boolean(expr) - Casts the value `expr` to the target data type `boolean`. + + +-- !query +DESC FUNCTION EXTENDED boolean +-- !query schema +struct +-- !query output +Class: org.apache.spark.sql.catalyst.expressions.Cast +Extended Usage: + No example/argument for boolean. + + Since: 2.0.1 + +Function: boolean +Usage: boolean(expr) - Casts the value `expr` to the target data type `boolean`. + + +-- !query +SELECT CAST('interval 3 month 1 hour' AS interval) +-- !query schema +struct +-- !query output +3 months 1 hours + + +-- !query +SELECT CAST("interval '3-1' year to month" AS interval year to month) +-- !query schema +struct +-- !query output +3-1 + + +-- !query +SELECT CAST("interval '3 00:00:01' day to second" AS interval day to second) +-- !query schema +struct +-- !query output +3 00:00:01.000000000 + + +-- !query +SELECT CAST(interval 3 month 1 hour AS string) +-- !query schema +struct<> +-- !query output +org.apache.spark.sql.catalyst.parser.ParseException +{ + "condition" : "_LEGACY_ERROR_TEMP_0029", + "messageParameters" : { + "literal" : "interval 3 month 1 hour" + }, + "queryContext" : [ { + "objectType" : "", + "objectName" : "", + "startIndex" : 13, + "stopIndex" : 35, + "fragment" : "interval 3 month 1 hour" + } ] +} + + +-- !query +SELECT CAST(interval 3 year 1 month AS string) +-- !query schema +struct +-- !query output +INTERVAL '3-1' YEAR TO MONTH + + +-- !query +SELECT CAST(interval 3 day 1 second AS string) +-- !query schema +struct +-- !query output +INTERVAL '3 00:00:01' DAY TO SECOND + + +-- !query +select cast(' 1' as tinyint) +-- !query schema +struct +-- !query output +1 + + +-- !query +select cast(' 1\t' as tinyint) +-- !query schema +struct +-- !query output +1 + + +-- !query +select cast(' 1' as smallint) +-- !query schema +struct +-- !query output +1 + + +-- !query +select cast(' 1' as INT) +-- !query schema +struct +-- !query output +1 + + +-- !query +select cast(' 1' as bigint) +-- !query schema +struct +-- !query output +1 + + +-- !query +select cast(' 1' as float) +-- !query schema +struct +-- !query output +1.0 + + +-- !query +select cast(' 1 ' as DOUBLE) +-- !query schema +struct +-- !query output +1.0 + + +-- !query +select cast('1.0 ' as DEC) +-- !query schema +struct +-- !query output +1 + + +-- !query +select cast('1中文' as tinyint) +-- !query schema +struct<> +-- !query output +org.apache.spark.SparkNumberFormatException +{ + "condition" : "CAST_INVALID_INPUT", + "sqlState" : "22018", + "messageParameters" : { + "expression" : "'1中文'", + "sourceType" : "\"STRING\"", + "targetType" : "\"TINYINT\"" + }, + "queryContext" : [ { + "objectType" : "", + "objectName" : "", + "startIndex" : 8, + "stopIndex" : 29, + "fragment" : "cast('1中文' as tinyint)" + } ] +} + + +-- !query +select cast('1中文' as smallint) +-- !query schema +struct<> +-- !query output +org.apache.spark.SparkNumberFormatException +{ + "condition" : "CAST_INVALID_INPUT", + "sqlState" : "22018", + "messageParameters" : { + "expression" : "'1中文'", + "sourceType" : "\"STRING\"", + "targetType" : "\"SMALLINT\"" + }, + "queryContext" : [ { + "objectType" : "", + "objectName" : "", + "startIndex" : 8, + "stopIndex" : 30, + "fragment" : "cast('1中文' as smallint)" + } ] +} + + +-- !query +select cast('1中文' as INT) +-- !query schema +struct<> +-- !query output +org.apache.spark.SparkNumberFormatException +{ + "condition" : "CAST_INVALID_INPUT", + "sqlState" : "22018", + "messageParameters" : { + "expression" : "'1中文'", + "sourceType" : "\"STRING\"", + "targetType" : "\"INT\"" + }, + "queryContext" : [ { + "objectType" : "", + "objectName" : "", + "startIndex" : 8, + "stopIndex" : 25, + "fragment" : "cast('1中文' as INT)" + } ] +} + + +-- !query +select cast('中文1' as bigint) +-- !query schema +struct<> +-- !query output +org.apache.spark.SparkNumberFormatException +{ + "condition" : "CAST_INVALID_INPUT", + "sqlState" : "22018", + "messageParameters" : { + "expression" : "'中文1'", + "sourceType" : "\"STRING\"", + "targetType" : "\"BIGINT\"" + }, + "queryContext" : [ { + "objectType" : "", + "objectName" : "", + "startIndex" : 8, + "stopIndex" : 28, + "fragment" : "cast('中文1' as bigint)" + } ] +} + + +-- !query +select cast('1中文' as bigint) +-- !query schema +struct<> +-- !query output +org.apache.spark.SparkNumberFormatException +{ + "condition" : "CAST_INVALID_INPUT", + "sqlState" : "22018", + "messageParameters" : { + "expression" : "'1中文'", + "sourceType" : "\"STRING\"", + "targetType" : "\"BIGINT\"" + }, + "queryContext" : [ { + "objectType" : "", + "objectName" : "", + "startIndex" : 8, + "stopIndex" : 28, + "fragment" : "cast('1中文' as bigint)" + } ] +} + + +-- !query +select cast('\t\t true \n\r ' as boolean) +-- !query schema +struct +-- !query output +true + + +-- !query +select cast('\t\n false \t\r' as boolean) +-- !query schema +struct +-- !query output +false + + +-- !query +select cast('\t\n xyz \t\r' as boolean) +-- !query schema +struct<> +-- !query output +org.apache.spark.SparkRuntimeException +{ + "condition" : "CAST_INVALID_INPUT", + "sqlState" : "22018", + "messageParameters" : { + "expression" : "'\t\n xyz \t\r'", + "sourceType" : "\"STRING\"", + "targetType" : "\"BOOLEAN\"" + }, + "queryContext" : [ { + "objectType" : "", + "objectName" : "", + "startIndex" : 8, + "stopIndex" : 39, + "fragment" : "cast('\\t\\n xyz \\t\\r' as boolean)" + } ] +} + + +-- !query +select cast('23.45' as decimal(4, 2)) +-- !query schema +struct +-- !query output +23.45 + + +-- !query +select cast('123.45' as decimal(4, 2)) +-- !query schema +struct<> +-- !query output +org.apache.spark.SparkArithmeticException +{ + "condition" : "NUMERIC_VALUE_OUT_OF_RANGE.WITH_SUGGESTION", + "sqlState" : "22003", + "messageParameters" : { + "config" : "\"spark.sql.ansi.enabled\"", + "precision" : "4", + "scale" : "2", + "value" : "123.45" + }, + "queryContext" : [ { + "objectType" : "", + "objectName" : "", + "startIndex" : 8, + "stopIndex" : 38, + "fragment" : "cast('123.45' as decimal(4, 2))" + } ] +} + + +-- !query +select cast('xyz' as decimal(4, 2)) +-- !query schema +struct<> +-- !query output +org.apache.spark.SparkNumberFormatException +{ + "condition" : "CAST_INVALID_INPUT", + "sqlState" : "22018", + "messageParameters" : { + "expression" : "'xyz'", + "sourceType" : "\"STRING\"", + "targetType" : "\"DECIMAL(4,2)\"" + }, + "queryContext" : [ { + "objectType" : "", + "objectName" : "", + "startIndex" : 8, + "stopIndex" : 35, + "fragment" : "cast('xyz' as decimal(4, 2))" + } ] +} + + +-- !query +select cast('2022-01-01' as date) +-- !query schema +struct +-- !query output +2022-01-01 + + +-- !query +select cast('a' as date) +-- !query schema +struct<> +-- !query output +org.apache.spark.SparkDateTimeException +{ + "condition" : "CAST_INVALID_INPUT", + "sqlState" : "22018", + "messageParameters" : { + "expression" : "'a'", + "sourceType" : "\"STRING\"", + "targetType" : "\"DATE\"" + }, + "queryContext" : [ { + "objectType" : "", + "objectName" : "", + "startIndex" : 8, + "stopIndex" : 24, + "fragment" : "cast('a' as date)" + } ] +} + + +-- !query +select cast('2022-01-01 00:00:00' as timestamp) +-- !query schema +struct +-- !query output +2022-01-01 00:00:00 + + +-- !query +select cast('a' as timestamp) +-- !query schema +struct<> +-- !query output +org.apache.spark.SparkDateTimeException +{ + "condition" : "CAST_INVALID_INPUT", + "sqlState" : "22018", + "messageParameters" : { + "expression" : "'a'", + "sourceType" : "\"STRING\"", + "targetType" : "\"TIMESTAMP\"" + }, + "queryContext" : [ { + "objectType" : "", + "objectName" : "", + "startIndex" : 8, + "stopIndex" : 29, + "fragment" : "cast('a' as timestamp)" + } ] +} + + +-- !query +select cast('2022-01-01 00:00:00' as timestamp_ntz) +-- !query schema +struct +-- !query output +2022-01-01 00:00:00 + + +-- !query +select cast('a' as timestamp_ntz) +-- !query schema +struct<> +-- !query output +org.apache.spark.SparkDateTimeException +{ + "condition" : "CAST_INVALID_INPUT", + "sqlState" : "22018", + "messageParameters" : { + "expression" : "'a'", + "sourceType" : "\"STRING\"", + "targetType" : "\"TIMESTAMP_NTZ\"" + }, + "queryContext" : [ { + "objectType" : "", + "objectName" : "", + "startIndex" : 8, + "stopIndex" : 33, + "fragment" : "cast('a' as timestamp_ntz)" + } ] +} + + +-- !query +select cast(cast('inf' as double) as timestamp) +-- !query schema +struct<> +-- !query output +org.apache.spark.SparkDateTimeException +{ + "condition" : "CAST_INVALID_INPUT", + "sqlState" : "22018", + "messageParameters" : { + "expression" : "Infinity", + "sourceType" : "\"DOUBLE\"", + "targetType" : "\"TIMESTAMP\"" + }, + "queryContext" : [ { + "objectType" : "", + "objectName" : "", + "startIndex" : 8, + "stopIndex" : 47, + "fragment" : "cast(cast('inf' as double) as timestamp)" + } ] +} + + +-- !query +select cast(cast('inf' as float) as timestamp) +-- !query schema +struct<> +-- !query output +org.apache.spark.SparkDateTimeException +{ + "condition" : "CAST_INVALID_INPUT", + "sqlState" : "22018", + "messageParameters" : { + "expression" : "Infinity", + "sourceType" : "\"DOUBLE\"", + "targetType" : "\"TIMESTAMP\"" + }, + "queryContext" : [ { + "objectType" : "", + "objectName" : "", + "startIndex" : 8, + "stopIndex" : 46, + "fragment" : "cast(cast('inf' as float) as timestamp)" + } ] +} + + +-- !query +select cast(interval '1' year as tinyint) +-- !query schema +struct +-- !query output +1 + + +-- !query +select cast(interval '-10-2' year to month as smallint) +-- !query schema +struct +-- !query output +-122 + + +-- !query +select cast(interval '1000' month as int) +-- !query schema +struct +-- !query output +1000 + + +-- !query +select cast(interval -'10.123456' second as tinyint) +-- !query schema +struct +-- !query output +-10 + + +-- !query +select cast(interval '23:59:59' hour to second as smallint) +-- !query schema +struct<> +-- !query output +org.apache.spark.SparkArithmeticException +{ + "condition" : "CAST_OVERFLOW", + "sqlState" : "22003", + "messageParameters" : { + "sourceType" : "\"INTERVAL HOUR TO SECOND\"", + "targetType" : "\"SMALLINT\"", + "value" : "INTERVAL '23:59:59' HOUR TO SECOND" + } +} + + +-- !query +select cast(interval -'1 02:03:04.123' day to second as int) +-- !query schema +struct +-- !query output +-93784 + + +-- !query +select cast(interval '10' day as bigint) +-- !query schema +struct +-- !query output +10 + + +-- !query +select cast(interval '-1000' month as tinyint) +-- !query schema +struct<> +-- !query output +org.apache.spark.SparkArithmeticException +{ + "condition" : "CAST_OVERFLOW", + "sqlState" : "22003", + "messageParameters" : { + "sourceType" : "\"INTERVAL MONTH\"", + "targetType" : "\"TINYINT\"", + "value" : "INTERVAL '-1000' MONTH" + } +} + + +-- !query +select cast(interval '1000000' second as smallint) +-- !query schema +struct<> +-- !query output +org.apache.spark.SparkArithmeticException +{ + "condition" : "CAST_OVERFLOW", + "sqlState" : "22003", + "messageParameters" : { + "sourceType" : "\"INTERVAL SECOND\"", + "targetType" : "\"SMALLINT\"", + "value" : "INTERVAL '1000000' SECOND" + } +} + + +-- !query +select cast(1Y as interval year) +-- !query schema +struct +-- !query output +1-0 + + +-- !query +select cast(-122S as interval year to month) +-- !query schema +struct +-- !query output +-10-2 + + +-- !query +select cast(ym as interval year to month) from values(-122S) as t(ym) +-- !query schema +struct +-- !query output +-10-2 + + +-- !query +select cast(1000 as interval month) +-- !query schema +struct +-- !query output +83-4 + + +-- !query +select cast(-10L as interval second) +-- !query schema +struct +-- !query output +-0 00:00:10.000000000 + + +-- !query +select cast(100Y as interval hour to second) +-- !query schema +struct +-- !query output +0 00:01:40.000000000 + + +-- !query +select cast(dt as interval hour to second) from values(100Y) as t(dt) +-- !query schema +struct +-- !query output +0 00:01:40.000000000 + + +-- !query +select cast(-1000S as interval day to second) +-- !query schema +struct +-- !query output +-0 00:16:40.000000000 + + +-- !query +select cast(10 as interval day) +-- !query schema +struct +-- !query output +10 00:00:00.000000000 + + +-- !query +select cast(2147483647 as interval year) +-- !query schema +struct<> +-- !query output +org.apache.spark.SparkArithmeticException +{ + "condition" : "CAST_OVERFLOW", + "sqlState" : "22003", + "messageParameters" : { + "sourceType" : "\"INT\"", + "targetType" : "\"INTERVAL YEAR\"", + "value" : "2147483647" + } +} + + +-- !query +select cast(-9223372036854775808L as interval day) +-- !query schema +struct<> +-- !query output +org.apache.spark.SparkArithmeticException +{ + "condition" : "CAST_OVERFLOW", + "sqlState" : "22003", + "messageParameters" : { + "sourceType" : "\"BIGINT\"", + "targetType" : "\"INTERVAL DAY\"", + "value" : "-9223372036854775808L" + } +} + + +-- !query +select cast(interval '-1' year as decimal(10, 0)) +-- !query schema +struct +-- !query output +-1 + + +-- !query +select cast(interval '1.000001' second as decimal(10, 6)) +-- !query schema +struct +-- !query output +1.000001 + + +-- !query +select cast(interval '08:11:10.001' hour to second as decimal(10, 4)) +-- !query schema +struct +-- !query output +29470.0010 + + +-- !query +select cast(interval '1 01:02:03.1' day to second as decimal(8, 1)) +-- !query schema +struct +-- !query output +90123.1 + + +-- !query +select cast(interval '10.123' second as decimal(4, 2)) +-- !query schema +struct +-- !query output +10.12 + + +-- !query +select cast(interval '10.005' second as decimal(4, 2)) +-- !query schema +struct +-- !query output +10.01 + + +-- !query +select cast(interval '10.123' second as decimal(5, 2)) +-- !query schema +struct +-- !query output +10.12 + + +-- !query +select cast(interval '10.123' second as decimal(1, 0)) +-- !query schema +struct<> +-- !query output +org.apache.spark.SparkArithmeticException +{ + "condition" : "NUMERIC_VALUE_OUT_OF_RANGE.WITH_SUGGESTION", + "sqlState" : "22003", + "messageParameters" : { + "config" : "\"spark.sql.ansi.enabled\"", + "precision" : "1", + "scale" : "0", + "value" : "10.123000" + }, + "queryContext" : [ { + "objectType" : "", + "objectName" : "", + "startIndex" : 8, + "stopIndex" : 54, + "fragment" : "cast(interval '10.123' second as decimal(1, 0))" + } ] +} + + +-- !query +select cast(10.123456BD as interval day to second) +-- !query schema +struct +-- !query output +0 00:00:10.123456000 + + +-- !query +select cast(80.654321BD as interval hour to minute) +-- !query schema +struct +-- !query output +0 01:20:00.000000000 + + +-- !query +select cast(-10.123456BD as interval year to month) +-- !query schema +struct +-- !query output +-0-10 + + +-- !query +select cast(10.654321BD as interval month) +-- !query schema +struct +-- !query output +0-11 + + +-- !query +SELECT '1.23' :: int +-- !query schema +struct<> +-- !query output +org.apache.spark.SparkNumberFormatException +{ + "condition" : "CAST_INVALID_INPUT", + "sqlState" : "22018", + "messageParameters" : { + "expression" : "'1.23'", + "sourceType" : "\"STRING\"", + "targetType" : "\"INT\"" + }, + "queryContext" : [ { + "objectType" : "", + "objectName" : "", + "startIndex" : 8, + "stopIndex" : 20, + "fragment" : "'1.23' :: int" + } ] +} + + +-- !query +SELECT 'abc' :: int +-- !query schema +struct<> +-- !query output +org.apache.spark.SparkNumberFormatException +{ + "condition" : "CAST_INVALID_INPUT", + "sqlState" : "22018", + "messageParameters" : { + "expression" : "'abc'", + "sourceType" : "\"STRING\"", + "targetType" : "\"INT\"" + }, + "queryContext" : [ { + "objectType" : "", + "objectName" : "", + "startIndex" : 8, + "stopIndex" : 19, + "fragment" : "'abc' :: int" + } ] +} + + +-- !query +SELECT '12345678901234567890123' :: long +-- !query schema +struct<> +-- !query output +org.apache.spark.SparkNumberFormatException +{ + "condition" : "CAST_INVALID_INPUT", + "sqlState" : "22018", + "messageParameters" : { + "expression" : "'12345678901234567890123'", + "sourceType" : "\"STRING\"", + "targetType" : "\"BIGINT\"" + }, + "queryContext" : [ { + "objectType" : "", + "objectName" : "", + "startIndex" : 8, + "stopIndex" : 40, + "fragment" : "'12345678901234567890123' :: long" + } ] +} + + +-- !query +SELECT '' :: int +-- !query schema +struct<> +-- !query output +org.apache.spark.SparkNumberFormatException +{ + "condition" : "CAST_INVALID_INPUT", + "sqlState" : "22018", + "messageParameters" : { + "expression" : "''", + "sourceType" : "\"STRING\"", + "targetType" : "\"INT\"" + }, + "queryContext" : [ { + "objectType" : "", + "objectName" : "", + "startIndex" : 8, + "stopIndex" : 16, + "fragment" : "'' :: int" + } ] +} + + +-- !query +SELECT NULL :: int +-- !query schema +struct +-- !query output +NULL + + +-- !query +SELECT '123.a' :: int +-- !query schema +struct<> +-- !query output +org.apache.spark.SparkNumberFormatException +{ + "condition" : "CAST_INVALID_INPUT", + "sqlState" : "22018", + "messageParameters" : { + "expression" : "'123.a'", + "sourceType" : "\"STRING\"", + "targetType" : "\"INT\"" + }, + "queryContext" : [ { + "objectType" : "", + "objectName" : "", + "startIndex" : 8, + "stopIndex" : 21, + "fragment" : "'123.a' :: int" + } ] +} + + +-- !query +SELECT '-2147483648' :: int +-- !query schema +struct +-- !query output +-2147483648 + + +-- !query +SELECT HEX('abc' :: binary) +-- !query schema +struct +-- !query output +616263 + + +-- !query +SELECT HEX((123 :: byte) :: binary) +-- !query schema +struct<> +-- !query output +org.apache.spark.sql.catalyst.ExtendedAnalysisException +{ + "condition" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", + "sqlState" : "42K09", + "messageParameters" : { + "config" : "\"spark.sql.ansi.enabled\"", + "configVal" : "'false'", + "sqlExpr" : "\"CAST(CAST(123 AS TINYINT) AS BINARY)\"", + "srcType" : "\"TINYINT\"", + "targetType" : "\"BINARY\"" + }, + "queryContext" : [ { + "objectType" : "", + "objectName" : "", + "startIndex" : 12, + "stopIndex" : 34, + "fragment" : "(123 :: byte) :: binary" + } ] +} + + +-- !query +SELECT 'interval 3 month 1 hour' :: interval +-- !query schema +struct +-- !query output +3 months 1 hours + + +-- !query +SELECT interval 3 day 1 second :: string +-- !query schema +struct +-- !query output +INTERVAL '3 00:00:01' DAY TO SECOND + + +-- !query +select ' 1 ' :: DOUBLE +-- !query schema +struct +-- !query output +1.0 + + +-- !query +select '1.0 ' :: DEC +-- !query schema +struct +-- !query output +1 + + +-- !query +select '\t\t true \n\r ' :: boolean +-- !query schema +struct +-- !query output +true + + +-- !query +select '2022-01-01 00:00:00' :: timestamp +-- !query schema +struct +-- !query output +2022-01-01 00:00:00 + + +-- !query +select interval '-10-2' year to month :: smallint +-- !query schema +struct +-- !query output +-122 + + +-- !query +select -10L :: interval second +-- !query schema +struct +-- !query output +-0 00:00:10.000000000 + + +-- !query +select interval '08:11:10.001' hour to second :: decimal(10, 4) +-- !query schema +struct +-- !query output +29470.0010 + + +-- !query +select 10.123456BD :: interval day to second +-- !query schema +struct +-- !query output +0 00:00:10.123456000 + + +-- !query +SELECT '1.23' :: int :: long +-- !query schema +struct<> +-- !query output +org.apache.spark.SparkNumberFormatException +{ + "condition" : "CAST_INVALID_INPUT", + "sqlState" : "22018", + "messageParameters" : { + "expression" : "'1.23'", + "sourceType" : "\"STRING\"", + "targetType" : "\"INT\"" + }, + "queryContext" : [ { + "objectType" : "", + "objectName" : "", + "startIndex" : 8, + "stopIndex" : 20, + "fragment" : "'1.23' :: int" + } ] +} + + +-- !query +SELECT '2147483648' :: long :: int +-- !query schema +struct<> +-- !query output +org.apache.spark.SparkArithmeticException +{ + "condition" : "CAST_OVERFLOW", + "sqlState" : "22003", + "messageParameters" : { + "sourceType" : "\"BIGINT\"", + "targetType" : "\"INT\"", + "value" : "2147483648L" + } +} + + +-- !query +SELECT CAST('2147483648' :: long AS int) +-- !query schema +struct<> +-- !query output +org.apache.spark.SparkArithmeticException +{ + "condition" : "CAST_OVERFLOW", + "sqlState" : "22003", + "messageParameters" : { + "sourceType" : "\"BIGINT\"", + "targetType" : "\"INT\"", + "value" : "2147483648L" + } +} + + +-- !query +SELECT map(1, '123', 2, '456')[1] :: int +-- !query schema +struct +-- !query output +123 + + +-- !query +SELECT '2147483648' :: BINT +-- !query schema +struct<> +-- !query output +org.apache.spark.sql.catalyst.parser.ParseException +{ + "condition" : "UNSUPPORTED_DATATYPE", + "sqlState" : "0A000", + "messageParameters" : { + "typeName" : "\"BINT\"" + }, + "queryContext" : [ { + "objectType" : "", + "objectName" : "", + "startIndex" : 24, + "stopIndex" : 27, + "fragment" : "BINT" + } ] +} + + +-- !query +SELECT '2147483648' :: SELECT +-- !query schema +struct<> +-- !query output +org.apache.spark.sql.catalyst.parser.ParseException +{ + "condition" : "UNSUPPORTED_DATATYPE", + "sqlState" : "0A000", + "messageParameters" : { + "typeName" : "\"SELECT\"" + }, + "queryContext" : [ { + "objectType" : "", + "objectName" : "", + "startIndex" : 24, + "stopIndex" : 29, + "fragment" : "SELECT" + } ] +} + + +-- !query +SELECT FALSE IS NOT NULL :: string +-- !query schema +struct<> +-- !query output +org.apache.spark.sql.catalyst.parser.ParseException +{ + "condition" : "PARSE_SYNTAX_ERROR", + "sqlState" : "42601", + "messageParameters" : { + "error" : "'::'", + "hint" : "" + } +} diff --git a/sql/core/src/test/resources/sql-tests/results/ansi/datetime-parsing-invalid.sql.out b/sql/core/src/test/resources/sql-tests/results/ansi/datetime-parsing-invalid.sql.out new file mode 100644 index 000000000000..736eba0adf71 --- /dev/null +++ b/sql/core/src/test/resources/sql-tests/results/ansi/datetime-parsing-invalid.sql.out @@ -0,0 +1,465 @@ +-- Automatically generated by SQLQueryTestSuite +-- !query +select to_timestamp('294248', 'y') +-- !query schema +struct<> +-- !query output +java.lang.ArithmeticException +long overflow + + +-- !query +select to_timestamp('1', 'yy') +-- !query schema +struct<> +-- !query output +org.apache.spark.SparkDateTimeException +{ + "condition" : "CANNOT_PARSE_TIMESTAMP", + "sqlState" : "22007", + "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", + "message" : "Text '1' could not be parsed at index 0" + } +} + + +-- !query +select to_timestamp('-12', 'yy') +-- !query schema +struct<> +-- !query output +org.apache.spark.SparkDateTimeException +{ + "condition" : "CANNOT_PARSE_TIMESTAMP", + "sqlState" : "22007", + "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", + "message" : "Text '-12' could not be parsed at index 0" + } +} + + +-- !query +select to_timestamp('123', 'yy') +-- !query schema +struct<> +-- !query output +org.apache.spark.SparkDateTimeException +{ + "condition" : "CANNOT_PARSE_TIMESTAMP", + "sqlState" : "22007", + "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", + "message" : "Text '123' could not be parsed, unparsed text found at index 2" + } +} + + +-- !query +select to_timestamp('1', 'yyy') +-- !query schema +struct<> +-- !query output +org.apache.spark.SparkDateTimeException +{ + "condition" : "CANNOT_PARSE_TIMESTAMP", + "sqlState" : "22007", + "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", + "message" : "Text '1' could not be parsed at index 0" + } +} + + +-- !query +select to_timestamp('1234567', 'yyyyyyy') +-- !query schema +struct<> +-- !query output +org.apache.spark.SparkUpgradeException +{ + "condition" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION.DATETIME_PATTERN_RECOGNITION", + "sqlState" : "42K0B", + "messageParameters" : { + "config" : "\"spark.sql.legacy.timeParserPolicy\"", + "docroot" : "https://spark.apache.org/docs/latest", + "pattern" : "'yyyyyyy'" + } +} + + +-- !query +select to_timestamp('366', 'D') +-- !query schema +struct<> +-- !query output +org.apache.spark.SparkDateTimeException +{ + "condition" : "CANNOT_PARSE_TIMESTAMP", + "sqlState" : "22007", + "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", + "message" : "Invalid date 'DayOfYear 366' as '1970' is not a leap year" + } +} + + +-- !query +select to_timestamp('9', 'DD') +-- !query schema +struct<> +-- !query output +org.apache.spark.SparkDateTimeException +{ + "condition" : "CANNOT_PARSE_TIMESTAMP", + "sqlState" : "22007", + "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", + "message" : "Text '9' could not be parsed at index 0" + } +} + + +-- !query +select to_timestamp('366', 'DD') +-- !query schema +struct<> +-- !query output +org.apache.spark.SparkDateTimeException +{ + "condition" : "CANNOT_PARSE_TIMESTAMP", + "sqlState" : "22007", + "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", + "message" : "Invalid date 'DayOfYear 366' as '1970' is not a leap year" + } +} + + +-- !query +select to_timestamp('9', 'DDD') +-- !query schema +struct<> +-- !query output +org.apache.spark.SparkDateTimeException +{ + "condition" : "CANNOT_PARSE_TIMESTAMP", + "sqlState" : "22007", + "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", + "message" : "Text '9' could not be parsed at index 0" + } +} + + +-- !query +select to_timestamp('99', 'DDD') +-- !query schema +struct<> +-- !query output +org.apache.spark.SparkDateTimeException +{ + "condition" : "CANNOT_PARSE_TIMESTAMP", + "sqlState" : "22007", + "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", + "message" : "Text '99' could not be parsed at index 0" + } +} + + +-- !query +select to_timestamp('30-365', 'dd-DDD') +-- !query schema +struct<> +-- !query output +org.apache.spark.SparkDateTimeException +{ + "condition" : "CANNOT_PARSE_TIMESTAMP", + "sqlState" : "22007", + "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", + "message" : "Conflict found: Field DayOfMonth 30 differs from DayOfMonth 31 derived from 1970-12-31." + } +} + + +-- !query +select to_timestamp('11-365', 'MM-DDD') +-- !query schema +struct<> +-- !query output +org.apache.spark.SparkDateTimeException +{ + "condition" : "CANNOT_PARSE_TIMESTAMP", + "sqlState" : "22007", + "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", + "message" : "Conflict found: Field MonthOfYear 11 differs from MonthOfYear 12 derived from 1970-12-31." + } +} + + +-- !query +select to_timestamp('2019-366', 'yyyy-DDD') +-- !query schema +struct<> +-- !query output +org.apache.spark.SparkDateTimeException +{ + "condition" : "CANNOT_PARSE_TIMESTAMP", + "sqlState" : "22007", + "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", + "message" : "Text '2019-366' could not be parsed: Invalid date 'DayOfYear 366' as '2019' is not a leap year" + } +} + + +-- !query +select to_timestamp('12-30-365', 'MM-dd-DDD') +-- !query schema +struct<> +-- !query output +org.apache.spark.SparkDateTimeException +{ + "condition" : "CANNOT_PARSE_TIMESTAMP", + "sqlState" : "22007", + "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", + "message" : "Conflict found: Field DayOfMonth 30 differs from DayOfMonth 31 derived from 1970-12-31." + } +} + + +-- !query +select to_timestamp('2020-01-365', 'yyyy-dd-DDD') +-- !query schema +struct<> +-- !query output +org.apache.spark.SparkDateTimeException +{ + "condition" : "CANNOT_PARSE_TIMESTAMP", + "sqlState" : "22007", + "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", + "message" : "Text '2020-01-365' could not be parsed: Conflict found: Field DayOfMonth 30 differs from DayOfMonth 1 derived from 2020-12-30" + } +} + + +-- !query +select to_timestamp('2020-10-350', 'yyyy-MM-DDD') +-- !query schema +struct<> +-- !query output +org.apache.spark.SparkDateTimeException +{ + "condition" : "CANNOT_PARSE_TIMESTAMP", + "sqlState" : "22007", + "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", + "message" : "Text '2020-10-350' could not be parsed: Conflict found: Field MonthOfYear 12 differs from MonthOfYear 10 derived from 2020-12-15" + } +} + + +-- !query +select to_timestamp('2020-11-31-366', 'yyyy-MM-dd-DDD') +-- !query schema +struct<> +-- !query output +org.apache.spark.SparkDateTimeException +{ + "condition" : "CANNOT_PARSE_TIMESTAMP", + "sqlState" : "22007", + "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", + "message" : "Text '2020-11-31-366' could not be parsed: Invalid date 'NOVEMBER 31'" + } +} + + +-- !query +select from_csv('2018-366', 'date Date', map('dateFormat', 'yyyy-DDD')) +-- !query schema +struct> +-- !query output +{"date":null} + + +-- !query +select to_date("2020-01-27T20:06:11.847", "yyyy-MM-dd HH:mm:ss.SSS") +-- !query schema +struct<> +-- !query output +org.apache.spark.SparkDateTimeException +{ + "condition" : "CANNOT_PARSE_TIMESTAMP", + "sqlState" : "22007", + "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", + "message" : "Text '2020-01-27T20:06:11.847' could not be parsed at index 10" + } +} + + +-- !query +select to_date("Unparseable", "yyyy-MM-dd HH:mm:ss.SSS") +-- !query schema +struct<> +-- !query output +org.apache.spark.SparkDateTimeException +{ + "condition" : "CANNOT_PARSE_TIMESTAMP", + "sqlState" : "22007", + "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", + "message" : "Text 'Unparseable' could not be parsed at index 0" + } +} + + +-- !query +select to_timestamp("2020-01-27T20:06:11.847", "yyyy-MM-dd HH:mm:ss.SSS") +-- !query schema +struct<> +-- !query output +org.apache.spark.SparkDateTimeException +{ + "condition" : "CANNOT_PARSE_TIMESTAMP", + "sqlState" : "22007", + "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", + "message" : "Text '2020-01-27T20:06:11.847' could not be parsed at index 10" + } +} + + +-- !query +select to_timestamp("Unparseable", "yyyy-MM-dd HH:mm:ss.SSS") +-- !query schema +struct<> +-- !query output +org.apache.spark.SparkDateTimeException +{ + "condition" : "CANNOT_PARSE_TIMESTAMP", + "sqlState" : "22007", + "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", + "message" : "Text 'Unparseable' could not be parsed at index 0" + } +} + + +-- !query +select unix_timestamp("2020-01-27T20:06:11.847", "yyyy-MM-dd HH:mm:ss.SSS") +-- !query schema +struct<> +-- !query output +org.apache.spark.SparkDateTimeException +{ + "condition" : "CANNOT_PARSE_TIMESTAMP", + "sqlState" : "22007", + "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", + "message" : "Text '2020-01-27T20:06:11.847' could not be parsed at index 10" + } +} + + +-- !query +select unix_timestamp("Unparseable", "yyyy-MM-dd HH:mm:ss.SSS") +-- !query schema +struct<> +-- !query output +org.apache.spark.SparkDateTimeException +{ + "condition" : "CANNOT_PARSE_TIMESTAMP", + "sqlState" : "22007", + "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", + "message" : "Text 'Unparseable' could not be parsed at index 0" + } +} + + +-- !query +select to_unix_timestamp("2020-01-27T20:06:11.847", "yyyy-MM-dd HH:mm:ss.SSS") +-- !query schema +struct<> +-- !query output +org.apache.spark.SparkDateTimeException +{ + "condition" : "CANNOT_PARSE_TIMESTAMP", + "sqlState" : "22007", + "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", + "message" : "Text '2020-01-27T20:06:11.847' could not be parsed at index 10" + } +} + + +-- !query +select to_unix_timestamp("Unparseable", "yyyy-MM-dd HH:mm:ss.SSS") +-- !query schema +struct<> +-- !query output +org.apache.spark.SparkDateTimeException +{ + "condition" : "CANNOT_PARSE_TIMESTAMP", + "sqlState" : "22007", + "messageParameters" : { + "ansiConfig" : "\"spark.sql.ansi.enabled\"", + "message" : "Text 'Unparseable' could not be parsed at index 0" + } +} + + +-- !query +select cast("Unparseable" as timestamp) +-- !query schema +struct<> +-- !query output +org.apache.spark.SparkDateTimeException +{ + "condition" : "CAST_INVALID_INPUT", + "sqlState" : "22018", + "messageParameters" : { + "expression" : "'Unparseable'", + "sourceType" : "\"STRING\"", + "targetType" : "\"TIMESTAMP\"" + }, + "queryContext" : [ { + "objectType" : "", + "objectName" : "", + "startIndex" : 8, + "stopIndex" : 39, + "fragment" : "cast(\"Unparseable\" as timestamp)" + } ] +} + + +-- !query +select cast("Unparseable" as date) +-- !query schema +struct<> +-- !query output +org.apache.spark.SparkDateTimeException +{ + "condition" : "CAST_INVALID_INPUT", + "sqlState" : "22018", + "messageParameters" : { + "expression" : "'Unparseable'", + "sourceType" : "\"STRING\"", + "targetType" : "\"DATE\"" + }, + "queryContext" : [ { + "objectType" : "", + "objectName" : "", + "startIndex" : 8, + "stopIndex" : 34, + "fragment" : "cast(\"Unparseable\" as date)" + } ] +} diff --git a/sql/core/src/test/resources/sql-tests/results/ansi/decimalArithmeticOperations.sql.out b/sql/core/src/test/resources/sql-tests/results/ansi/decimalArithmeticOperations.sql.out new file mode 100644 index 000000000000..e0b96debee81 --- /dev/null +++ b/sql/core/src/test/resources/sql-tests/results/ansi/decimalArithmeticOperations.sql.out @@ -0,0 +1,365 @@ +-- Automatically generated by SQLQueryTestSuite +-- !query +create table decimals_test(id int, a decimal(38,18), b decimal(38,18)) using parquet +-- !query schema +struct<> +-- !query output + + + +-- !query +insert into decimals_test values(1, 100.0, 999.0), (2, 12345.123, 12345.123), + (3, 0.1234567891011, 1234.1), (4, 123456789123456789.0, 1.123456789123456789) +-- !query schema +struct<> +-- !query output + + + +-- !query +select id, a*10, b/10 from decimals_test order by id +-- !query schema +struct +-- !query output +1 1000.000000000000000 99.900000000000000000 +2 123451.230000000000000 1234.512300000000000000 +3 1.234567891011000 123.410000000000000000 +4 1234567891234567890.000000000000000 0.112345678912345679 + + +-- !query +select 10.3 * 3.0 +-- !query schema +struct<(10.3 * 3.0):decimal(6,2)> +-- !query output +30.90 + + +-- !query +select 10.3000 * 3.0 +-- !query schema +struct<(10.3000 * 3.0):decimal(9,5)> +-- !query output +30.90000 + + +-- !query +select 10.30000 * 30.0 +-- !query schema +struct<(10.30000 * 30.0):decimal(11,6)> +-- !query output +309.000000 + + +-- !query +select 10.300000000000000000 * 3.000000000000000000 +-- !query schema +struct<(10.300000000000000000 * 3.000000000000000000):decimal(38,34)> +-- !query output +30.9000000000000000000000000000000000 + + +-- !query +select 10.300000000000000000 * 3.0000000000000000000 +-- !query schema +struct<(10.300000000000000000 * 3.0000000000000000000):decimal(38,34)> +-- !query output +30.9000000000000000000000000000000000 + + +-- !query +select (5e36BD + 0.1) + 5e36BD +-- !query schema +struct<> +-- !query output +org.apache.spark.SparkArithmeticException +{ + "condition" : "NUMERIC_VALUE_OUT_OF_RANGE.WITH_SUGGESTION", + "sqlState" : "22003", + "messageParameters" : { + "config" : "\"spark.sql.ansi.enabled\"", + "precision" : "38", + "scale" : "1", + "value" : "10000000000000000000000000000000000000.1" + }, + "queryContext" : [ { + "objectType" : "", + "objectName" : "", + "startIndex" : 8, + "stopIndex" : 30, + "fragment" : "(5e36BD + 0.1) + 5e36BD" + } ] +} + + +-- !query +select (-4e36BD - 0.1) - 7e36BD +-- !query schema +struct<> +-- !query output +org.apache.spark.SparkArithmeticException +{ + "condition" : "NUMERIC_VALUE_OUT_OF_RANGE.WITH_SUGGESTION", + "sqlState" : "22003", + "messageParameters" : { + "config" : "\"spark.sql.ansi.enabled\"", + "precision" : "38", + "scale" : "1", + "value" : "-11000000000000000000000000000000000000.1" + }, + "queryContext" : [ { + "objectType" : "", + "objectName" : "", + "startIndex" : 8, + "stopIndex" : 31, + "fragment" : "(-4e36BD - 0.1) - 7e36BD" + } ] +} + + +-- !query +select 12345678901234567890.0 * 12345678901234567890.0 +-- !query schema +struct<> +-- !query output +org.apache.spark.SparkArithmeticException +{ + "condition" : "NUMERIC_VALUE_OUT_OF_RANGE.WITH_SUGGESTION", + "sqlState" : "22003", + "messageParameters" : { + "config" : "\"spark.sql.ansi.enabled\"", + "precision" : "38", + "scale" : "2", + "value" : "152415787532388367501905199875019052100" + }, + "queryContext" : [ { + "objectType" : "", + "objectName" : "", + "startIndex" : 8, + "stopIndex" : 54, + "fragment" : "12345678901234567890.0 * 12345678901234567890.0" + } ] +} + + +-- !query +select 1e35BD / 0.1 +-- !query schema +struct<> +-- !query output +org.apache.spark.SparkArithmeticException +{ + "condition" : "NUMERIC_VALUE_OUT_OF_RANGE.WITH_SUGGESTION", + "sqlState" : "22003", + "messageParameters" : { + "config" : "\"spark.sql.ansi.enabled\"", + "precision" : "38", + "scale" : "6", + "value" : "1000000000000000000000000000000000000.000000000000000000000000000000000000000" + }, + "queryContext" : [ { + "objectType" : "", + "objectName" : "", + "startIndex" : 8, + "stopIndex" : 19, + "fragment" : "1e35BD / 0.1" + } ] +} + + +-- !query +select 123456789123456789.1234567890 * 1.123456789123456789 +-- !query schema +struct<(123456789123456789.1234567890 * 1.123456789123456789):decimal(38,18)> +-- !query output +138698367904130467.654320988515622621 + + +-- !query +select 123456789123456789.1234567890 * 1.123456789123456789 +-- !query schema +struct<(123456789123456789.1234567890 * 1.123456789123456789):decimal(38,18)> +-- !query output +138698367904130467.654320988515622621 + + +-- !query +select 12345678912345.123456789123 / 0.000000012345678 +-- !query schema +struct<(12345678912345.123456789123 / 1.2345678E-8):decimal(38,9)> +-- !query output +1000000073899961059796.725866332 + + +-- !query +select 1.0123456789012345678901234567890123456e36BD / 0.1 +-- !query schema +struct<> +-- !query output +org.apache.spark.SparkArithmeticException +{ + "condition" : "NUMERIC_VALUE_OUT_OF_RANGE.WITH_SUGGESTION", + "sqlState" : "22003", + "messageParameters" : { + "config" : "\"spark.sql.ansi.enabled\"", + "precision" : "38", + "scale" : "6", + "value" : "10123456789012345678901234567890123456.000000000000000000000000000000000000000" + }, + "queryContext" : [ { + "objectType" : "", + "objectName" : "", + "startIndex" : 8, + "stopIndex" : 57, + "fragment" : "1.0123456789012345678901234567890123456e36BD / 0.1" + } ] +} + + +-- !query +select 1.0123456789012345678901234567890123456e35BD / 1.0 +-- !query schema +struct<> +-- !query output +org.apache.spark.SparkArithmeticException +{ + "condition" : "NUMERIC_VALUE_OUT_OF_RANGE.WITH_SUGGESTION", + "sqlState" : "22003", + "messageParameters" : { + "config" : "\"spark.sql.ansi.enabled\"", + "precision" : "38", + "scale" : "6", + "value" : "101234567890123456789012345678901234.560000000000000000000000000000000000000" + }, + "queryContext" : [ { + "objectType" : "", + "objectName" : "", + "startIndex" : 8, + "stopIndex" : 57, + "fragment" : "1.0123456789012345678901234567890123456e35BD / 1.0" + } ] +} + + +-- !query +select 1.0123456789012345678901234567890123456e34BD / 1.0 +-- !query schema +struct<> +-- !query output +org.apache.spark.SparkArithmeticException +{ + "condition" : "NUMERIC_VALUE_OUT_OF_RANGE.WITH_SUGGESTION", + "sqlState" : "22003", + "messageParameters" : { + "config" : "\"spark.sql.ansi.enabled\"", + "precision" : "38", + "scale" : "6", + "value" : "10123456789012345678901234567890123.456000000000000000000000000000000000000" + }, + "queryContext" : [ { + "objectType" : "", + "objectName" : "", + "startIndex" : 8, + "stopIndex" : 57, + "fragment" : "1.0123456789012345678901234567890123456e34BD / 1.0" + } ] +} + + +-- !query +select 1.0123456789012345678901234567890123456e33BD / 1.0 +-- !query schema +struct<> +-- !query output +org.apache.spark.SparkArithmeticException +{ + "condition" : "NUMERIC_VALUE_OUT_OF_RANGE.WITH_SUGGESTION", + "sqlState" : "22003", + "messageParameters" : { + "config" : "\"spark.sql.ansi.enabled\"", + "precision" : "38", + "scale" : "6", + "value" : "1012345678901234567890123456789012.345600000000000000000000000000000000000" + }, + "queryContext" : [ { + "objectType" : "", + "objectName" : "", + "startIndex" : 8, + "stopIndex" : 57, + "fragment" : "1.0123456789012345678901234567890123456e33BD / 1.0" + } ] +} + + +-- !query +select 1.0123456789012345678901234567890123456e32BD / 1.0 +-- !query schema +struct<> +-- !query output +org.apache.spark.SparkArithmeticException +{ + "condition" : "NUMERIC_VALUE_OUT_OF_RANGE.WITH_SUGGESTION", + "sqlState" : "22003", + "messageParameters" : { + "config" : "\"spark.sql.ansi.enabled\"", + "precision" : "38", + "scale" : "6", + "value" : "101234567890123456789012345678901.234560000000000000000000000000000000000" + }, + "queryContext" : [ { + "objectType" : "", + "objectName" : "", + "startIndex" : 8, + "stopIndex" : 57, + "fragment" : "1.0123456789012345678901234567890123456e32BD / 1.0" + } ] +} + + +-- !query +select 1.0123456789012345678901234567890123456e31BD / 1.0 +-- !query schema +struct<(10123456789012345678901234567890.123456 / 1.0):decimal(38,6)> +-- !query output +10123456789012345678901234567890.123456 + + +-- !query +select 1.0123456789012345678901234567890123456e31BD / 0.1 +-- !query schema +struct<> +-- !query output +org.apache.spark.SparkArithmeticException +{ + "condition" : "NUMERIC_VALUE_OUT_OF_RANGE.WITH_SUGGESTION", + "sqlState" : "22003", + "messageParameters" : { + "config" : "\"spark.sql.ansi.enabled\"", + "precision" : "38", + "scale" : "6", + "value" : "101234567890123456789012345678901.234560000000000000000000000000000000000" + }, + "queryContext" : [ { + "objectType" : "", + "objectName" : "", + "startIndex" : 8, + "stopIndex" : 57, + "fragment" : "1.0123456789012345678901234567890123456e31BD / 0.1" + } ] +} + + +-- !query +select 1.0123456789012345678901234567890123456e31BD / 10.0 +-- !query schema +struct<(10123456789012345678901234567890.123456 / 10.0):decimal(38,6)> +-- !query output +1012345678901234567890123456789.012346 + + +-- !query +drop table decimals_test +-- !query schema +struct<> +-- !query output + diff --git a/sql/core/src/test/resources/sql-tests/results/ansi/math.sql.out b/sql/core/src/test/resources/sql-tests/results/ansi/math.sql.out new file mode 100644 index 000000000000..ed9f0469ee6a --- /dev/null +++ b/sql/core/src/test/resources/sql-tests/results/ansi/math.sql.out @@ -0,0 +1,903 @@ +-- Automatically generated by SQLQueryTestSuite +-- !query +SELECT round(25y, 1) +-- !query schema +struct +-- !query output +25 + + +-- !query +SELECT round(25y, 0) +-- !query schema +struct +-- !query output +25 + + +-- !query +SELECT round(25y, -1) +-- !query schema +struct +-- !query output +30 + + +-- !query +SELECT round(25y, -2) +-- !query schema +struct +-- !query output +0 + + +-- !query +SELECT round(25y, -3) +-- !query schema +struct +-- !query output +0 + + +-- !query +SELECT round(127y, -1) +-- !query schema +struct<> +-- !query output +org.apache.spark.SparkArithmeticException +{ + "condition" : "ARITHMETIC_OVERFLOW", + "sqlState" : "22003", + "messageParameters" : { + "alternative" : "", + "config" : "\"spark.sql.ansi.enabled\"", + "message" : "Overflow" + }, + "queryContext" : [ { + "objectType" : "", + "objectName" : "", + "startIndex" : 8, + "stopIndex" : 22, + "fragment" : "round(127y, -1)" + } ] +} + + +-- !query +SELECT round(-128y, -1) +-- !query schema +struct<> +-- !query output +org.apache.spark.SparkArithmeticException +{ + "condition" : "ARITHMETIC_OVERFLOW", + "sqlState" : "22003", + "messageParameters" : { + "alternative" : "", + "config" : "\"spark.sql.ansi.enabled\"", + "message" : "Overflow" + }, + "queryContext" : [ { + "objectType" : "", + "objectName" : "", + "startIndex" : 8, + "stopIndex" : 23, + "fragment" : "round(-128y, -1)" + } ] +} + + +-- !query +SELECT round(525s, 1) +-- !query schema +struct +-- !query output +525 + + +-- !query +SELECT round(525s, 0) +-- !query schema +struct +-- !query output +525 + + +-- !query +SELECT round(525s, -1) +-- !query schema +struct +-- !query output +530 + + +-- !query +SELECT round(525s, -2) +-- !query schema +struct +-- !query output +500 + + +-- !query +SELECT round(525s, -3) +-- !query schema +struct +-- !query output +1000 + + +-- !query +SELECT round(32767s, -1) +-- !query schema +struct<> +-- !query output +org.apache.spark.SparkArithmeticException +{ + "condition" : "ARITHMETIC_OVERFLOW", + "sqlState" : "22003", + "messageParameters" : { + "alternative" : "", + "config" : "\"spark.sql.ansi.enabled\"", + "message" : "Overflow" + }, + "queryContext" : [ { + "objectType" : "", + "objectName" : "", + "startIndex" : 8, + "stopIndex" : 24, + "fragment" : "round(32767s, -1)" + } ] +} + + +-- !query +SELECT round(-32768s, -1) +-- !query schema +struct<> +-- !query output +org.apache.spark.SparkArithmeticException +{ + "condition" : "ARITHMETIC_OVERFLOW", + "sqlState" : "22003", + "messageParameters" : { + "alternative" : "", + "config" : "\"spark.sql.ansi.enabled\"", + "message" : "Overflow" + }, + "queryContext" : [ { + "objectType" : "", + "objectName" : "", + "startIndex" : 8, + "stopIndex" : 25, + "fragment" : "round(-32768s, -1)" + } ] +} + + +-- !query +SELECT round(525, 1) +-- !query schema +struct +-- !query output +525 + + +-- !query +SELECT round(525, 0) +-- !query schema +struct +-- !query output +525 + + +-- !query +SELECT round(525, -1) +-- !query schema +struct +-- !query output +530 + + +-- !query +SELECT round(525, -2) +-- !query schema +struct +-- !query output +500 + + +-- !query +SELECT round(525, -3) +-- !query schema +struct +-- !query output +1000 + + +-- !query +SELECT round(2147483647, -1) +-- !query schema +struct<> +-- !query output +org.apache.spark.SparkArithmeticException +{ + "condition" : "ARITHMETIC_OVERFLOW", + "sqlState" : "22003", + "messageParameters" : { + "alternative" : "", + "config" : "\"spark.sql.ansi.enabled\"", + "message" : "Overflow" + }, + "queryContext" : [ { + "objectType" : "", + "objectName" : "", + "startIndex" : 8, + "stopIndex" : 28, + "fragment" : "round(2147483647, -1)" + } ] +} + + +-- !query +SELECT round(-2147483647, -1) +-- !query schema +struct<> +-- !query output +org.apache.spark.SparkArithmeticException +{ + "condition" : "ARITHMETIC_OVERFLOW", + "sqlState" : "22003", + "messageParameters" : { + "alternative" : "", + "config" : "\"spark.sql.ansi.enabled\"", + "message" : "Overflow" + }, + "queryContext" : [ { + "objectType" : "", + "objectName" : "", + "startIndex" : 8, + "stopIndex" : 29, + "fragment" : "round(-2147483647, -1)" + } ] +} + + +-- !query +SELECT round(525L, 1) +-- !query schema +struct +-- !query output +525 + + +-- !query +SELECT round(525L, 0) +-- !query schema +struct +-- !query output +525 + + +-- !query +SELECT round(525L, -1) +-- !query schema +struct +-- !query output +530 + + +-- !query +SELECT round(525L, -2) +-- !query schema +struct +-- !query output +500 + + +-- !query +SELECT round(525L, -3) +-- !query schema +struct +-- !query output +1000 + + +-- !query +SELECT round(9223372036854775807L, -1) +-- !query schema +struct<> +-- !query output +org.apache.spark.SparkArithmeticException +{ + "condition" : "ARITHMETIC_OVERFLOW", + "sqlState" : "22003", + "messageParameters" : { + "alternative" : "", + "config" : "\"spark.sql.ansi.enabled\"", + "message" : "Overflow" + }, + "queryContext" : [ { + "objectType" : "", + "objectName" : "", + "startIndex" : 8, + "stopIndex" : 38, + "fragment" : "round(9223372036854775807L, -1)" + } ] +} + + +-- !query +SELECT round(-9223372036854775808L, -1) +-- !query schema +struct<> +-- !query output +org.apache.spark.SparkArithmeticException +{ + "condition" : "ARITHMETIC_OVERFLOW", + "sqlState" : "22003", + "messageParameters" : { + "alternative" : "", + "config" : "\"spark.sql.ansi.enabled\"", + "message" : "Overflow" + }, + "queryContext" : [ { + "objectType" : "", + "objectName" : "", + "startIndex" : 8, + "stopIndex" : 39, + "fragment" : "round(-9223372036854775808L, -1)" + } ] +} + + +-- !query +SELECT bround(25y, 1) +-- !query schema +struct +-- !query output +25 + + +-- !query +SELECT bround(25y, 0) +-- !query schema +struct +-- !query output +25 + + +-- !query +SELECT bround(25y, -1) +-- !query schema +struct +-- !query output +20 + + +-- !query +SELECT bround(25y, -2) +-- !query schema +struct +-- !query output +0 + + +-- !query +SELECT bround(25y, -3) +-- !query schema +struct +-- !query output +0 + + +-- !query +SELECT bround(127y, -1) +-- !query schema +struct<> +-- !query output +org.apache.spark.SparkArithmeticException +{ + "condition" : "ARITHMETIC_OVERFLOW", + "sqlState" : "22003", + "messageParameters" : { + "alternative" : "", + "config" : "\"spark.sql.ansi.enabled\"", + "message" : "Overflow" + }, + "queryContext" : [ { + "objectType" : "", + "objectName" : "", + "startIndex" : 8, + "stopIndex" : 23, + "fragment" : "bround(127y, -1)" + } ] +} + + +-- !query +SELECT bround(-128y, -1) +-- !query schema +struct<> +-- !query output +org.apache.spark.SparkArithmeticException +{ + "condition" : "ARITHMETIC_OVERFLOW", + "sqlState" : "22003", + "messageParameters" : { + "alternative" : "", + "config" : "\"spark.sql.ansi.enabled\"", + "message" : "Overflow" + }, + "queryContext" : [ { + "objectType" : "", + "objectName" : "", + "startIndex" : 8, + "stopIndex" : 24, + "fragment" : "bround(-128y, -1)" + } ] +} + + +-- !query +SELECT bround(525s, 1) +-- !query schema +struct +-- !query output +525 + + +-- !query +SELECT bround(525s, 0) +-- !query schema +struct +-- !query output +525 + + +-- !query +SELECT bround(525s, -1) +-- !query schema +struct +-- !query output +520 + + +-- !query +SELECT bround(525s, -2) +-- !query schema +struct +-- !query output +500 + + +-- !query +SELECT bround(525s, -3) +-- !query schema +struct +-- !query output +1000 + + +-- !query +SELECT bround(32767s, -1) +-- !query schema +struct<> +-- !query output +org.apache.spark.SparkArithmeticException +{ + "condition" : "ARITHMETIC_OVERFLOW", + "sqlState" : "22003", + "messageParameters" : { + "alternative" : "", + "config" : "\"spark.sql.ansi.enabled\"", + "message" : "Overflow" + }, + "queryContext" : [ { + "objectType" : "", + "objectName" : "", + "startIndex" : 8, + "stopIndex" : 25, + "fragment" : "bround(32767s, -1)" + } ] +} + + +-- !query +SELECT bround(-32768s, -1) +-- !query schema +struct<> +-- !query output +org.apache.spark.SparkArithmeticException +{ + "condition" : "ARITHMETIC_OVERFLOW", + "sqlState" : "22003", + "messageParameters" : { + "alternative" : "", + "config" : "\"spark.sql.ansi.enabled\"", + "message" : "Overflow" + }, + "queryContext" : [ { + "objectType" : "", + "objectName" : "", + "startIndex" : 8, + "stopIndex" : 26, + "fragment" : "bround(-32768s, -1)" + } ] +} + + +-- !query +SELECT bround(525, 1) +-- !query schema +struct +-- !query output +525 + + +-- !query +SELECT bround(525, 0) +-- !query schema +struct +-- !query output +525 + + +-- !query +SELECT bround(525, -1) +-- !query schema +struct +-- !query output +520 + + +-- !query +SELECT bround(525, -2) +-- !query schema +struct +-- !query output +500 + + +-- !query +SELECT bround(525, -3) +-- !query schema +struct +-- !query output +1000 + + +-- !query +SELECT bround(2147483647, -1) +-- !query schema +struct<> +-- !query output +org.apache.spark.SparkArithmeticException +{ + "condition" : "ARITHMETIC_OVERFLOW", + "sqlState" : "22003", + "messageParameters" : { + "alternative" : "", + "config" : "\"spark.sql.ansi.enabled\"", + "message" : "Overflow" + }, + "queryContext" : [ { + "objectType" : "", + "objectName" : "", + "startIndex" : 8, + "stopIndex" : 29, + "fragment" : "bround(2147483647, -1)" + } ] +} + + +-- !query +SELECT bround(-2147483647, -1) +-- !query schema +struct<> +-- !query output +org.apache.spark.SparkArithmeticException +{ + "condition" : "ARITHMETIC_OVERFLOW", + "sqlState" : "22003", + "messageParameters" : { + "alternative" : "", + "config" : "\"spark.sql.ansi.enabled\"", + "message" : "Overflow" + }, + "queryContext" : [ { + "objectType" : "", + "objectName" : "", + "startIndex" : 8, + "stopIndex" : 30, + "fragment" : "bround(-2147483647, -1)" + } ] +} + + +-- !query +SELECT bround(525L, 1) +-- !query schema +struct +-- !query output +525 + + +-- !query +SELECT bround(525L, 0) +-- !query schema +struct +-- !query output +525 + + +-- !query +SELECT bround(525L, -1) +-- !query schema +struct +-- !query output +520 + + +-- !query +SELECT bround(525L, -2) +-- !query schema +struct +-- !query output +500 + + +-- !query +SELECT bround(525L, -3) +-- !query schema +struct +-- !query output +1000 + + +-- !query +SELECT bround(9223372036854775807L, -1) +-- !query schema +struct<> +-- !query output +org.apache.spark.SparkArithmeticException +{ + "condition" : "ARITHMETIC_OVERFLOW", + "sqlState" : "22003", + "messageParameters" : { + "alternative" : "", + "config" : "\"spark.sql.ansi.enabled\"", + "message" : "Overflow" + }, + "queryContext" : [ { + "objectType" : "", + "objectName" : "", + "startIndex" : 8, + "stopIndex" : 39, + "fragment" : "bround(9223372036854775807L, -1)" + } ] +} + + +-- !query +SELECT bround(-9223372036854775808L, -1) +-- !query schema +struct<> +-- !query output +org.apache.spark.SparkArithmeticException +{ + "condition" : "ARITHMETIC_OVERFLOW", + "sqlState" : "22003", + "messageParameters" : { + "alternative" : "", + "config" : "\"spark.sql.ansi.enabled\"", + "message" : "Overflow" + }, + "queryContext" : [ { + "objectType" : "", + "objectName" : "", + "startIndex" : 8, + "stopIndex" : 40, + "fragment" : "bround(-9223372036854775808L, -1)" + } ] +} + + +-- !query +SELECT conv('100', 2, 10) +-- !query schema +struct +-- !query output +4 + + +-- !query +SELECT conv(-10, 16, -10) +-- !query schema +struct +-- !query output +-16 + + +-- !query +SELECT conv('9223372036854775808', 10, 16) +-- !query schema +struct +-- !query output +8000000000000000 + + +-- !query +SELECT conv('92233720368547758070', 10, 16) +-- !query schema +struct<> +-- !query output +org.apache.spark.SparkArithmeticException +{ + "condition" : "ARITHMETIC_OVERFLOW", + "sqlState" : "22003", + "messageParameters" : { + "alternative" : "", + "config" : "\"spark.sql.ansi.enabled\"", + "message" : "Overflow in function conv()" + }, + "queryContext" : [ { + "objectType" : "", + "objectName" : "", + "startIndex" : 8, + "stopIndex" : 43, + "fragment" : "conv('92233720368547758070', 10, 16)" + } ] +} + + +-- !query +SELECT conv('9223372036854775807', 36, 10) +-- !query schema +struct<> +-- !query output +org.apache.spark.SparkArithmeticException +{ + "condition" : "ARITHMETIC_OVERFLOW", + "sqlState" : "22003", + "messageParameters" : { + "alternative" : "", + "config" : "\"spark.sql.ansi.enabled\"", + "message" : "Overflow in function conv()" + }, + "queryContext" : [ { + "objectType" : "", + "objectName" : "", + "startIndex" : 8, + "stopIndex" : 42, + "fragment" : "conv('9223372036854775807', 36, 10)" + } ] +} + + +-- !query +SELECT conv('-9223372036854775807', 36, 10) +-- !query schema +struct<> +-- !query output +org.apache.spark.SparkArithmeticException +{ + "condition" : "ARITHMETIC_OVERFLOW", + "sqlState" : "22003", + "messageParameters" : { + "alternative" : "", + "config" : "\"spark.sql.ansi.enabled\"", + "message" : "Overflow in function conv()" + }, + "queryContext" : [ { + "objectType" : "", + "objectName" : "", + "startIndex" : 8, + "stopIndex" : 43, + "fragment" : "conv('-9223372036854775807', 36, 10)" + } ] +} + + +-- !query +SELECT BIN(0) +-- !query schema +struct +-- !query output +0 + + +-- !query +SELECT BIN(25) +-- !query schema +struct +-- !query output +11001 + + +-- !query +SELECT BIN(25L) +-- !query schema +struct +-- !query output +11001 + + +-- !query +SELECT BIN(25.5) +-- !query schema +struct +-- !query output +11001 + + +-- !query +SELECT POSITIVE(0Y) +-- !query schema +struct<(+ 0):tinyint> +-- !query output +0 + + +-- !query +SELECT POSITIVE(25) +-- !query schema +struct<(+ 25):int> +-- !query output +25 + + +-- !query +SELECT POSITIVE(-25L) +-- !query schema +struct<(+ -25):bigint> +-- !query output +-25 + + +-- !query +SELECT POSITIVE(25.5) +-- !query schema +struct<(+ 25.5):decimal(3,1)> +-- !query output +25.5 + + +-- !query +SELECT POSITIVE("25.5") +-- !query schema +struct<(+ 25.5):double> +-- !query output +25.5 + + +-- !query +SELECT POSITIVE("invalid") +-- !query schema +struct<> +-- !query output +org.apache.spark.SparkNumberFormatException +{ + "condition" : "CAST_INVALID_INPUT", + "sqlState" : "22018", + "messageParameters" : { + "expression" : "'invalid'", + "sourceType" : "\"STRING\"", + "targetType" : "\"DOUBLE\"" + }, + "queryContext" : [ { + "objectType" : "", + "objectName" : "", + "startIndex" : 8, + "stopIndex" : 26, + "fragment" : "POSITIVE(\"invalid\")" + } ] +} + + +-- !query +SELECT POSITIVE(null) +-- !query schema +struct<(+ NULL):double> +-- !query output +NULL diff --git a/sql/core/src/test/resources/sql-tests/results/array.sql.out b/sql/core/src/test/resources/sql-tests/results/array.sql.out index b5dabfb47f43..0c141c08d436 100644 --- a/sql/core/src/test/resources/sql-tests/results/array.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/array.sql.out @@ -129,7 +129,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"1\"", @@ -182,7 +182,7 @@ struct<> -- !query output org.apache.spark.SparkArrayIndexOutOfBoundsException { - "errorClass" : "INVALID_ARRAY_INDEX_IN_ELEMENT_AT", + "condition" : "INVALID_ARRAY_INDEX_IN_ELEMENT_AT", "sqlState" : "22003", "messageParameters" : { "arraySize" : "3", @@ -205,7 +205,7 @@ struct<> -- !query output org.apache.spark.SparkArrayIndexOutOfBoundsException { - "errorClass" : "INVALID_ARRAY_INDEX_IN_ELEMENT_AT", + "condition" : "INVALID_ARRAY_INDEX_IN_ELEMENT_AT", "sqlState" : "22003", "messageParameters" : { "arraySize" : "3", @@ -228,7 +228,7 @@ struct<> -- !query output org.apache.spark.SparkRuntimeException { - "errorClass" : "INVALID_INDEX_OF_ZERO", + "condition" : "INVALID_INDEX_OF_ZERO", "sqlState" : "22003", "queryContext" : [ { "objectType" : "", @@ -247,7 +247,7 @@ struct<> -- !query output org.apache.spark.SparkArrayIndexOutOfBoundsException { - "errorClass" : "INVALID_ARRAY_INDEX", + "condition" : "INVALID_ARRAY_INDEX", "sqlState" : "22003", "messageParameters" : { "arraySize" : "2", @@ -270,7 +270,7 @@ struct<> -- !query output org.apache.spark.SparkArrayIndexOutOfBoundsException { - "errorClass" : "INVALID_ARRAY_INDEX", + "condition" : "INVALID_ARRAY_INDEX", "sqlState" : "22003", "messageParameters" : { "arraySize" : "2", @@ -293,7 +293,7 @@ struct<> -- !query output org.apache.spark.SparkArrayIndexOutOfBoundsException { - "errorClass" : "INVALID_ARRAY_INDEX", + "condition" : "INVALID_ARRAY_INDEX", "sqlState" : "22003", "messageParameters" : { "arraySize" : "2", @@ -348,7 +348,7 @@ struct<> -- !query output org.apache.spark.SparkArrayIndexOutOfBoundsException { - "errorClass" : "INVALID_ARRAY_INDEX", + "condition" : "INVALID_ARRAY_INDEX", "sqlState" : "22003", "messageParameters" : { "arraySize" : "3", @@ -371,7 +371,7 @@ struct<> -- !query output org.apache.spark.SparkArrayIndexOutOfBoundsException { - "errorClass" : "INVALID_ARRAY_INDEX", + "condition" : "INVALID_ARRAY_INDEX", "sqlState" : "22003", "messageParameters" : { "arraySize" : "3", @@ -426,7 +426,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"map(a, 1, b, 2)\"", @@ -540,7 +540,7 @@ struct<> -- !query output org.apache.spark.SparkRuntimeException { - "errorClass" : "INVALID_INDEX_OF_ZERO", + "condition" : "INVALID_INDEX_OF_ZERO", "sqlState" : "22003", "queryContext" : [ { "objectType" : "", @@ -575,7 +575,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.ARRAY_FUNCTION_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.ARRAY_FUNCTION_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "\"ARRAY\"", @@ -681,7 +681,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"id\"", diff --git a/sql/core/src/test/resources/sql-tests/results/bitwise.sql.out b/sql/core/src/test/resources/sql-tests/results/bitwise.sql.out index 7233b0d0ae49..a6c889fac7c1 100644 --- a/sql/core/src/test/resources/sql-tests/results/bitwise.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/bitwise.sql.out @@ -150,7 +150,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"bit count\"", @@ -176,7 +176,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"a\"", @@ -295,7 +295,7 @@ struct<> -- !query output org.apache.spark.SparkIllegalArgumentException { - "errorClass" : "INVALID_PARAMETER_VALUE.BIT_POSITION_RANGE", + "condition" : "INVALID_PARAMETER_VALUE.BIT_POSITION_RANGE", "sqlState" : "22023", "messageParameters" : { "functionName" : "`getbit`", @@ -313,7 +313,7 @@ struct<> -- !query output org.apache.spark.SparkIllegalArgumentException { - "errorClass" : "INVALID_PARAMETER_VALUE.BIT_POSITION_RANGE", + "condition" : "INVALID_PARAMETER_VALUE.BIT_POSITION_RANGE", "sqlState" : "22023", "messageParameters" : { "functionName" : "`getbit`", @@ -355,7 +355,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'>'", @@ -371,7 +371,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'<'", @@ -387,7 +387,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'>>'", @@ -403,7 +403,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'<'", @@ -419,7 +419,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'>'", diff --git a/sql/core/src/test/resources/sql-tests/results/cast.sql.out b/sql/core/src/test/resources/sql-tests/results/cast.sql.out index 0dbdf1d9975c..4399570b3076 100644 --- a/sql/core/src/test/resources/sql-tests/results/cast.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/cast.sql.out @@ -6,7 +6,7 @@ struct<> -- !query output org.apache.spark.SparkNumberFormatException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'1.23'", @@ -30,7 +30,7 @@ struct<> -- !query output org.apache.spark.SparkNumberFormatException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'1.23'", @@ -54,7 +54,7 @@ struct<> -- !query output org.apache.spark.SparkNumberFormatException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'-4.56'", @@ -78,7 +78,7 @@ struct<> -- !query output org.apache.spark.SparkNumberFormatException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'-4.56'", @@ -102,7 +102,7 @@ struct<> -- !query output org.apache.spark.SparkNumberFormatException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'abc'", @@ -126,7 +126,7 @@ struct<> -- !query output org.apache.spark.SparkNumberFormatException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'abc'", @@ -150,7 +150,7 @@ struct<> -- !query output org.apache.spark.SparkNumberFormatException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'abc'", @@ -174,7 +174,7 @@ struct<> -- !query output org.apache.spark.SparkNumberFormatException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'abc'", @@ -198,7 +198,7 @@ struct<> -- !query output org.apache.spark.SparkNumberFormatException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'1234567890123'", @@ -222,7 +222,7 @@ struct<> -- !query output org.apache.spark.SparkNumberFormatException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'12345678901234567890123'", @@ -246,7 +246,7 @@ struct<> -- !query output org.apache.spark.SparkNumberFormatException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "''", @@ -270,7 +270,7 @@ struct<> -- !query output org.apache.spark.SparkNumberFormatException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "''", @@ -294,7 +294,7 @@ struct<> -- !query output org.apache.spark.SparkNumberFormatException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "''", @@ -318,7 +318,7 @@ struct<> -- !query output org.apache.spark.SparkNumberFormatException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "''", @@ -358,7 +358,7 @@ struct<> -- !query output org.apache.spark.SparkNumberFormatException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'123.a'", @@ -382,7 +382,7 @@ struct<> -- !query output org.apache.spark.SparkNumberFormatException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'123.a'", @@ -406,7 +406,7 @@ struct<> -- !query output org.apache.spark.SparkNumberFormatException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'123.a'", @@ -430,7 +430,7 @@ struct<> -- !query output org.apache.spark.SparkNumberFormatException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'123.a'", @@ -462,7 +462,7 @@ struct<> -- !query output org.apache.spark.SparkNumberFormatException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'-2147483649'", @@ -494,7 +494,7 @@ struct<> -- !query output org.apache.spark.SparkNumberFormatException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'2147483648'", @@ -526,7 +526,7 @@ struct<> -- !query output org.apache.spark.SparkNumberFormatException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'-9223372036854775809'", @@ -558,7 +558,7 @@ struct<> -- !query output org.apache.spark.SparkNumberFormatException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'9223372036854775808'", @@ -590,7 +590,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", + "condition" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", "sqlState" : "42K09", "messageParameters" : { "config" : "\"spark.sql.ansi.enabled\"", @@ -616,7 +616,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", + "condition" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", "sqlState" : "42K09", "messageParameters" : { "config" : "\"spark.sql.ansi.enabled\"", @@ -642,7 +642,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", + "condition" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", "sqlState" : "42K09", "messageParameters" : { "config" : "\"spark.sql.ansi.enabled\"", @@ -668,7 +668,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", + "condition" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", "sqlState" : "42K09", "messageParameters" : { "config" : "\"spark.sql.ansi.enabled\"", @@ -694,7 +694,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", + "condition" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", "sqlState" : "42K09", "messageParameters" : { "config" : "\"spark.sql.ansi.enabled\"", @@ -720,7 +720,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", + "condition" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", "sqlState" : "42K09", "messageParameters" : { "config" : "\"spark.sql.ansi.enabled\"", @@ -746,7 +746,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", + "condition" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", "sqlState" : "42K09", "messageParameters" : { "config" : "\"spark.sql.ansi.enabled\"", @@ -772,7 +772,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", + "condition" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", "sqlState" : "42K09", "messageParameters" : { "config" : "\"spark.sql.ansi.enabled\"", @@ -847,7 +847,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0029", + "condition" : "_LEGACY_ERROR_TEMP_0029", "messageParameters" : { "literal" : "interval 3 month 1 hour" }, @@ -948,7 +948,7 @@ struct<> -- !query output org.apache.spark.SparkNumberFormatException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'1中文'", @@ -972,7 +972,7 @@ struct<> -- !query output org.apache.spark.SparkNumberFormatException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'1中文'", @@ -996,7 +996,7 @@ struct<> -- !query output org.apache.spark.SparkNumberFormatException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'1中文'", @@ -1020,7 +1020,7 @@ struct<> -- !query output org.apache.spark.SparkNumberFormatException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'中文1'", @@ -1044,7 +1044,7 @@ struct<> -- !query output org.apache.spark.SparkNumberFormatException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'1中文'", @@ -1086,7 +1086,7 @@ struct<> -- !query output org.apache.spark.SparkRuntimeException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'\t\n xyz \t\r'", @@ -1118,7 +1118,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "NUMERIC_VALUE_OUT_OF_RANGE.WITH_SUGGESTION", + "condition" : "NUMERIC_VALUE_OUT_OF_RANGE.WITH_SUGGESTION", "sqlState" : "22003", "messageParameters" : { "config" : "\"spark.sql.ansi.enabled\"", @@ -1143,7 +1143,7 @@ struct<> -- !query output org.apache.spark.SparkNumberFormatException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'xyz'", @@ -1175,7 +1175,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'a'", @@ -1207,7 +1207,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'a'", @@ -1239,7 +1239,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'a'", @@ -1263,7 +1263,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "Infinity", @@ -1287,7 +1287,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "Infinity", @@ -1343,7 +1343,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "CAST_OVERFLOW", + "condition" : "CAST_OVERFLOW", "sqlState" : "22003", "messageParameters" : { "sourceType" : "\"INTERVAL HOUR TO SECOND\"", @@ -1376,7 +1376,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "CAST_OVERFLOW", + "condition" : "CAST_OVERFLOW", "sqlState" : "22003", "messageParameters" : { "sourceType" : "\"INTERVAL MONTH\"", @@ -1393,7 +1393,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "CAST_OVERFLOW", + "condition" : "CAST_OVERFLOW", "sqlState" : "22003", "messageParameters" : { "sourceType" : "\"INTERVAL SECOND\"", @@ -1482,7 +1482,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "CAST_OVERFLOW", + "condition" : "CAST_OVERFLOW", "sqlState" : "22003", "messageParameters" : { "sourceType" : "\"INT\"", @@ -1499,7 +1499,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "CAST_OVERFLOW", + "condition" : "CAST_OVERFLOW", "sqlState" : "22003", "messageParameters" : { "sourceType" : "\"BIGINT\"", @@ -1572,7 +1572,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "NUMERIC_VALUE_OUT_OF_RANGE.WITH_SUGGESTION", + "condition" : "NUMERIC_VALUE_OUT_OF_RANGE.WITH_SUGGESTION", "sqlState" : "22003", "messageParameters" : { "config" : "\"spark.sql.ansi.enabled\"", @@ -1629,7 +1629,7 @@ struct<> -- !query output org.apache.spark.SparkNumberFormatException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'1.23'", @@ -1653,7 +1653,7 @@ struct<> -- !query output org.apache.spark.SparkNumberFormatException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'abc'", @@ -1677,7 +1677,7 @@ struct<> -- !query output org.apache.spark.SparkNumberFormatException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'12345678901234567890123'", @@ -1701,7 +1701,7 @@ struct<> -- !query output org.apache.spark.SparkNumberFormatException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "''", @@ -1733,7 +1733,7 @@ struct<> -- !query output org.apache.spark.SparkNumberFormatException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'123.a'", @@ -1773,7 +1773,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", + "condition" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", "sqlState" : "42K09", "messageParameters" : { "config" : "\"spark.sql.ansi.enabled\"", @@ -1880,7 +1880,7 @@ struct<> -- !query output org.apache.spark.SparkNumberFormatException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'1.23'", @@ -1904,7 +1904,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "CAST_OVERFLOW", + "condition" : "CAST_OVERFLOW", "sqlState" : "22003", "messageParameters" : { "sourceType" : "\"BIGINT\"", @@ -1921,7 +1921,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "CAST_OVERFLOW", + "condition" : "CAST_OVERFLOW", "sqlState" : "22003", "messageParameters" : { "sourceType" : "\"BIGINT\"", @@ -1946,7 +1946,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "UNSUPPORTED_DATATYPE", + "condition" : "UNSUPPORTED_DATATYPE", "sqlState" : "0A000", "messageParameters" : { "typeName" : "\"BINT\"" @@ -1968,7 +1968,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "UNSUPPORTED_DATATYPE", + "condition" : "UNSUPPORTED_DATATYPE", "sqlState" : "0A000", "messageParameters" : { "typeName" : "\"SELECT\"" @@ -1990,7 +1990,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'::'", diff --git a/sql/core/src/test/resources/sql-tests/results/ceil-floor-with-scale-param.sql.out b/sql/core/src/test/resources/sql-tests/results/ceil-floor-with-scale-param.sql.out index 86f54665ad01..96440690cf4b 100644 --- a/sql/core/src/test/resources/sql-tests/results/ceil-floor-with-scale-param.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/ceil-floor-with-scale-param.sql.out @@ -94,7 +94,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "NON_FOLDABLE_ARGUMENT", + "condition" : "NON_FOLDABLE_ARGUMENT", "sqlState" : "42K08", "messageParameters" : { "funcName" : "`ceil`", @@ -118,7 +118,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "NON_FOLDABLE_ARGUMENT", + "condition" : "NON_FOLDABLE_ARGUMENT", "sqlState" : "42K08", "messageParameters" : { "funcName" : "`ceil`", @@ -142,7 +142,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", + "condition" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", "sqlState" : "42605", "messageParameters" : { "actualNum" : "3", @@ -255,7 +255,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "NON_FOLDABLE_ARGUMENT", + "condition" : "NON_FOLDABLE_ARGUMENT", "sqlState" : "42K08", "messageParameters" : { "funcName" : "`floor`", @@ -279,7 +279,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "NON_FOLDABLE_ARGUMENT", + "condition" : "NON_FOLDABLE_ARGUMENT", "sqlState" : "42K08", "messageParameters" : { "funcName" : "`floor`", @@ -303,7 +303,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", + "condition" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", "sqlState" : "42605", "messageParameters" : { "actualNum" : "3", diff --git a/sql/core/src/test/resources/sql-tests/results/change-column.sql.out b/sql/core/src/test/resources/sql-tests/results/change-column.sql.out index a6110543159a..2670d778d3db 100644 --- a/sql/core/src/test/resources/sql-tests/results/change-column.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/change-column.sql.out @@ -24,7 +24,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0035", + "condition" : "_LEGACY_ERROR_TEMP_0035", "messageParameters" : { "message" : "ALTER TABLE table CHANGE COLUMN requires a TYPE, a SET/DROP, a COMMENT, or a FIRST/AFTER" }, @@ -55,7 +55,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "UNSUPPORTED_FEATURE.TABLE_OPERATION", + "condition" : "UNSUPPORTED_FEATURE.TABLE_OPERATION", "sqlState" : "0A000", "messageParameters" : { "operation" : "RENAME COLUMN", @@ -81,7 +81,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "NOT_SUPPORTED_CHANGE_COLUMN", + "condition" : "NOT_SUPPORTED_CHANGE_COLUMN", "sqlState" : "0A000", "messageParameters" : { "newName" : "`a`", @@ -117,7 +117,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "UNSUPPORTED_FEATURE.TABLE_OPERATION", + "condition" : "UNSUPPORTED_FEATURE.TABLE_OPERATION", "sqlState" : "0A000", "messageParameters" : { "operation" : "ALTER COLUMN ... FIRST | AFTER", @@ -133,7 +133,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "UNSUPPORTED_FEATURE.TABLE_OPERATION", + "condition" : "UNSUPPORTED_FEATURE.TABLE_OPERATION", "sqlState" : "0A000", "messageParameters" : { "operation" : "ALTER COLUMN ... FIRST | AFTER", @@ -219,7 +219,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`invalid_col`", @@ -278,7 +278,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "EXPECT_TABLE_NOT_VIEW.NO_ALTERNATIVE", + "condition" : "EXPECT_TABLE_NOT_VIEW.NO_ALTERNATIVE", "sqlState" : "42809", "messageParameters" : { "operation" : "ALTER TABLE ... CHANGE COLUMN", @@ -309,7 +309,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "EXPECT_TABLE_NOT_VIEW.NO_ALTERNATIVE", + "condition" : "EXPECT_TABLE_NOT_VIEW.NO_ALTERNATIVE", "sqlState" : "42809", "messageParameters" : { "operation" : "ALTER TABLE ... CHANGE COLUMN", diff --git a/sql/core/src/test/resources/sql-tests/results/charvarchar.sql.out b/sql/core/src/test/resources/sql-tests/results/charvarchar.sql.out index 3026e463947f..93ff8dd4b320 100644 --- a/sql/core/src/test/resources/sql-tests/results/charvarchar.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/charvarchar.sql.out @@ -262,7 +262,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "NOT_SUPPORTED_CHANGE_COLUMN", + "condition" : "NOT_SUPPORTED_CHANGE_COLUMN", "sqlState" : "0A000", "messageParameters" : { "newName" : "`c`", @@ -614,7 +614,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "_LEGACY_ERROR_TEMP_1232", + "condition" : "_LEGACY_ERROR_TEMP_1232", "messageParameters" : { "partitionColumnNames" : "v2, c2", "specKeys" : "v2", diff --git a/sql/core/src/test/resources/sql-tests/results/collations.sql.out b/sql/core/src/test/resources/sql-tests/results/collations.sql.out index fbfde3d78c1b..8c150b1de03e 100644 --- a/sql/core/src/test/resources/sql-tests/results/collations.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/collations.sql.out @@ -227,7 +227,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -254,7 +254,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -281,7 +281,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -308,7 +308,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -335,7 +335,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -587,7 +587,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "COLLATION_MISMATCH.IMPLICIT", + "condition" : "COLLATION_MISMATCH.IMPLICIT", "sqlState" : "42P21", "messageParameters" : { "implicitTypes" : "\"STRING\", \"STRING COLLATE UTF8_LCASE\"" @@ -602,7 +602,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "COLLATION_MISMATCH.EXPLICIT", + "condition" : "COLLATION_MISMATCH.EXPLICIT", "sqlState" : "42P21", "messageParameters" : { "explicitTypes" : "\"STRING\", \"STRING COLLATE UTF8_LCASE\"" @@ -625,7 +625,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"collate(text, unicode_ai)\"", @@ -921,7 +921,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "COLLATION_MISMATCH.IMPLICIT", + "condition" : "COLLATION_MISMATCH.IMPLICIT", "sqlState" : "42P21", "messageParameters" : { "implicitTypes" : "\"STRING\", \"STRING COLLATE UTF8_LCASE\"" @@ -936,7 +936,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "COLLATION_MISMATCH.EXPLICIT", + "condition" : "COLLATION_MISMATCH.EXPLICIT", "sqlState" : "42P21", "messageParameters" : { "explicitTypes" : "\"STRING\", \"STRING COLLATE UTF8_LCASE\"" @@ -1039,7 +1039,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "COLLATION_MISMATCH.IMPLICIT", + "condition" : "COLLATION_MISMATCH.IMPLICIT", "sqlState" : "42P21", "messageParameters" : { "implicitTypes" : "\"STRING\", \"STRING COLLATE UTF8_LCASE\"" @@ -1054,7 +1054,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "COLLATION_MISMATCH.EXPLICIT", + "condition" : "COLLATION_MISMATCH.EXPLICIT", "sqlState" : "42P21", "messageParameters" : { "explicitTypes" : "\"STRING\", \"STRING COLLATE UTF8_LCASE\"" @@ -1157,7 +1157,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "COLLATION_MISMATCH.IMPLICIT", + "condition" : "COLLATION_MISMATCH.IMPLICIT", "sqlState" : "42P21", "messageParameters" : { "implicitTypes" : "\"STRING\", \"STRING COLLATE UTF8_LCASE\"" @@ -1194,7 +1194,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "COLLATION_MISMATCH.EXPLICIT", + "condition" : "COLLATION_MISMATCH.EXPLICIT", "sqlState" : "42P21", "messageParameters" : { "explicitTypes" : "\"STRING\", \"STRING COLLATE UTF8_LCASE\"" @@ -1239,7 +1239,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"collate(utf8_binary, unicode_ai)\"", @@ -1331,7 +1331,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "COLLATION_MISMATCH.IMPLICIT", + "condition" : "COLLATION_MISMATCH.IMPLICIT", "sqlState" : "42P21", "messageParameters" : { "implicitTypes" : "\"STRING\", \"STRING COLLATE UTF8_LCASE\"" @@ -1368,7 +1368,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "COLLATION_MISMATCH.EXPLICIT", + "condition" : "COLLATION_MISMATCH.EXPLICIT", "sqlState" : "42P21", "messageParameters" : { "explicitTypes" : "\"STRING\", \"STRING COLLATE UTF8_LCASE\"" @@ -1427,7 +1427,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"collate(utf8_binary, unicode_ai)\"", @@ -1519,7 +1519,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "COLLATION_MISMATCH.IMPLICIT", + "condition" : "COLLATION_MISMATCH.IMPLICIT", "sqlState" : "42P21", "messageParameters" : { "implicitTypes" : "\"STRING\", \"STRING COLLATE UTF8_LCASE\"" @@ -1556,7 +1556,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "COLLATION_MISMATCH.EXPLICIT", + "condition" : "COLLATION_MISMATCH.EXPLICIT", "sqlState" : "42P21", "messageParameters" : { "explicitTypes" : "\"STRING\", \"STRING COLLATE UTF8_LCASE\"" @@ -1615,7 +1615,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"collate(utf8_binary, unicode_ai)\"", @@ -1707,7 +1707,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "COLLATION_MISMATCH.IMPLICIT", + "condition" : "COLLATION_MISMATCH.IMPLICIT", "sqlState" : "42P21", "messageParameters" : { "implicitTypes" : "\"STRING\", \"STRING COLLATE UTF8_LCASE\"" @@ -1744,7 +1744,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "COLLATION_MISMATCH.EXPLICIT", + "condition" : "COLLATION_MISMATCH.EXPLICIT", "sqlState" : "42P21", "messageParameters" : { "explicitTypes" : "\"STRING\", \"STRING COLLATE UTF8_LCASE\"" @@ -1803,7 +1803,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"collate(utf8_binary, unicode_ai)\"", @@ -1873,7 +1873,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "COLLATION_MISMATCH.IMPLICIT", + "condition" : "COLLATION_MISMATCH.IMPLICIT", "sqlState" : "42P21", "messageParameters" : { "implicitTypes" : "\"STRING\", \"STRING COLLATE UTF8_LCASE\"" @@ -1910,7 +1910,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "COLLATION_MISMATCH.EXPLICIT", + "condition" : "COLLATION_MISMATCH.EXPLICIT", "sqlState" : "42P21", "messageParameters" : { "explicitTypes" : "\"STRING\", \"STRING COLLATE UTF8_LCASE\"" @@ -2035,7 +2035,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "COLLATION_MISMATCH.IMPLICIT", + "condition" : "COLLATION_MISMATCH.IMPLICIT", "sqlState" : "42P21", "messageParameters" : { "implicitTypes" : "\"STRING\", \"STRING COLLATE UTF8_LCASE\"" @@ -2072,7 +2072,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "COLLATION_MISMATCH.EXPLICIT", + "condition" : "COLLATION_MISMATCH.EXPLICIT", "sqlState" : "42P21", "messageParameters" : { "explicitTypes" : "\"STRING\", \"STRING COLLATE UTF8_LCASE\"" @@ -2131,7 +2131,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"collate(utf8_binary, unicode_ai)\"", @@ -2245,7 +2245,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "COLLATION_MISMATCH.IMPLICIT", + "condition" : "COLLATION_MISMATCH.IMPLICIT", "sqlState" : "42P21", "messageParameters" : { "implicitTypes" : "\"STRING\", \"STRING COLLATE UTF8_LCASE\"" @@ -2260,7 +2260,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "COLLATION_MISMATCH.EXPLICIT", + "condition" : "COLLATION_MISMATCH.EXPLICIT", "sqlState" : "42P21", "messageParameters" : { "explicitTypes" : "\"STRING COLLATE UTF8_LCASE\", \"STRING\"" @@ -2297,7 +2297,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"utf8_binary\"", @@ -2389,7 +2389,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "COLLATION_MISMATCH.IMPLICIT", + "condition" : "COLLATION_MISMATCH.IMPLICIT", "sqlState" : "42P21", "messageParameters" : { "implicitTypes" : "\"STRING\", \"STRING COLLATE UTF8_LCASE\"" @@ -2426,7 +2426,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "COLLATION_MISMATCH.EXPLICIT", + "condition" : "COLLATION_MISMATCH.EXPLICIT", "sqlState" : "42P21", "messageParameters" : { "explicitTypes" : "\"STRING\", \"STRING COLLATE UTF8_LCASE\"" @@ -2485,7 +2485,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"collate(utf8_binary, unicode_ai)\"", @@ -2577,7 +2577,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "COLLATION_MISMATCH.IMPLICIT", + "condition" : "COLLATION_MISMATCH.IMPLICIT", "sqlState" : "42P21", "messageParameters" : { "implicitTypes" : "\"STRING\", \"STRING COLLATE UTF8_LCASE\"" @@ -2614,7 +2614,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "COLLATION_MISMATCH.EXPLICIT", + "condition" : "COLLATION_MISMATCH.EXPLICIT", "sqlState" : "42P21", "messageParameters" : { "explicitTypes" : "\"STRING\", \"STRING COLLATE UTF8_LCASE\"" @@ -2673,7 +2673,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"collate(utf8_binary, unicode_ai)\"", @@ -3250,7 +3250,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "COLLATION_MISMATCH.IMPLICIT", + "condition" : "COLLATION_MISMATCH.IMPLICIT", "sqlState" : "42P21", "messageParameters" : { "implicitTypes" : "\"STRING\", \"STRING COLLATE UTF8_LCASE\"" @@ -3287,7 +3287,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "COLLATION_MISMATCH.EXPLICIT", + "condition" : "COLLATION_MISMATCH.EXPLICIT", "sqlState" : "42P21", "messageParameters" : { "explicitTypes" : "\"STRING\", \"STRING COLLATE UTF8_LCASE\"" @@ -3621,7 +3621,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "COLLATION_MISMATCH.IMPLICIT", + "condition" : "COLLATION_MISMATCH.IMPLICIT", "sqlState" : "42P21", "messageParameters" : { "implicitTypes" : "\"STRING\", \"STRING COLLATE UTF8_LCASE\"" @@ -3658,7 +3658,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "COLLATION_MISMATCH.EXPLICIT", + "condition" : "COLLATION_MISMATCH.EXPLICIT", "sqlState" : "42P21", "messageParameters" : { "explicitTypes" : "\"STRING\", \"STRING COLLATE UTF8_LCASE\"" @@ -4157,7 +4157,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "COLLATION_MISMATCH.IMPLICIT", + "condition" : "COLLATION_MISMATCH.IMPLICIT", "sqlState" : "42P21", "messageParameters" : { "implicitTypes" : "\"STRING\", \"STRING COLLATE UTF8_LCASE\"" @@ -4194,7 +4194,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "COLLATION_MISMATCH.EXPLICIT", + "condition" : "COLLATION_MISMATCH.EXPLICIT", "sqlState" : "42P21", "messageParameters" : { "explicitTypes" : "\"STRING\", \"STRING COLLATE UTF8_LCASE\"" @@ -4319,7 +4319,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "COLLATION_MISMATCH.IMPLICIT", + "condition" : "COLLATION_MISMATCH.IMPLICIT", "sqlState" : "42P21", "messageParameters" : { "implicitTypes" : "\"STRING\", \"STRING COLLATE UTF8_LCASE\"" @@ -4356,7 +4356,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "COLLATION_MISMATCH.EXPLICIT", + "condition" : "COLLATION_MISMATCH.EXPLICIT", "sqlState" : "42P21", "messageParameters" : { "explicitTypes" : "\"STRING\", \"STRING COLLATE UTF8_LCASE\"" @@ -4481,7 +4481,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "COLLATION_MISMATCH.IMPLICIT", + "condition" : "COLLATION_MISMATCH.IMPLICIT", "sqlState" : "42P21", "messageParameters" : { "implicitTypes" : "\"STRING\", \"STRING COLLATE UTF8_LCASE\"" @@ -4518,7 +4518,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "COLLATION_MISMATCH.EXPLICIT", + "condition" : "COLLATION_MISMATCH.EXPLICIT", "sqlState" : "42P21", "messageParameters" : { "explicitTypes" : "\"STRING\", \"STRING COLLATE UTF8_LCASE\"" @@ -4577,7 +4577,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"collate(utf8_binary, unicode_ai)\"", @@ -4669,7 +4669,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "COLLATION_MISMATCH.IMPLICIT", + "condition" : "COLLATION_MISMATCH.IMPLICIT", "sqlState" : "42P21", "messageParameters" : { "implicitTypes" : "\"STRING COLLATE UTF8_LCASE\", \"STRING\"" @@ -4692,7 +4692,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "COLLATION_MISMATCH.EXPLICIT", + "condition" : "COLLATION_MISMATCH.EXPLICIT", "sqlState" : "42P21", "messageParameters" : { "explicitTypes" : "\"STRING COLLATE UTF8_LCASE\", \"STRING\"" @@ -4751,7 +4751,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"collate(utf8_lcase, unicode_ai)\"", @@ -4843,7 +4843,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "COLLATION_MISMATCH.IMPLICIT", + "condition" : "COLLATION_MISMATCH.IMPLICIT", "sqlState" : "42P21", "messageParameters" : { "implicitTypes" : "\"STRING\", \"STRING COLLATE UTF8_LCASE\"" @@ -4880,7 +4880,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "COLLATION_MISMATCH.EXPLICIT", + "condition" : "COLLATION_MISMATCH.EXPLICIT", "sqlState" : "42P21", "messageParameters" : { "explicitTypes" : "\"STRING\", \"STRING COLLATE UTF8_LCASE\"" @@ -4939,7 +4939,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"collate(utf8_binary, unicode_ai)\"", @@ -5031,7 +5031,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "COLLATION_MISMATCH.IMPLICIT", + "condition" : "COLLATION_MISMATCH.IMPLICIT", "sqlState" : "42P21", "messageParameters" : { "implicitTypes" : "\"STRING COLLATE UTF8_LCASE\", \"STRING\"" @@ -5054,7 +5054,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "COLLATION_MISMATCH.EXPLICIT", + "condition" : "COLLATION_MISMATCH.EXPLICIT", "sqlState" : "42P21", "messageParameters" : { "explicitTypes" : "\"STRING COLLATE UTF8_LCASE\", \"STRING\"" @@ -5113,7 +5113,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"collate(utf8_lcase, unicode_ai)\"", @@ -5205,7 +5205,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "COLLATION_MISMATCH.IMPLICIT", + "condition" : "COLLATION_MISMATCH.IMPLICIT", "sqlState" : "42P21", "messageParameters" : { "implicitTypes" : "\"STRING COLLATE UTF8_LCASE\", \"STRING\"" @@ -5242,7 +5242,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "COLLATION_MISMATCH.EXPLICIT", + "condition" : "COLLATION_MISMATCH.EXPLICIT", "sqlState" : "42P21", "messageParameters" : { "explicitTypes" : "\"STRING COLLATE UTF8_LCASE\", \"STRING\"" @@ -5301,7 +5301,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"collate(utf8_lcase, unicode_ai)\"", diff --git a/sql/core/src/test/resources/sql-tests/results/column-resolution-aggregate.sql.out b/sql/core/src/test/resources/sql-tests/results/column-resolution-aggregate.sql.out index bc4942be7860..ae080ee65f65 100644 --- a/sql/core/src/test/resources/sql-tests/results/column-resolution-aggregate.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/column-resolution-aggregate.sql.out @@ -22,7 +22,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "MISSING_AGGREGATION", + "condition" : "MISSING_AGGREGATION", "sqlState" : "42803", "messageParameters" : { "expression" : "\"b\"", @@ -47,7 +47,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "MISSING_AGGREGATION", + "condition" : "MISSING_AGGREGATION", "sqlState" : "42803", "messageParameters" : { "expression" : "\"a\"", @@ -63,7 +63,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "MISSING_AGGREGATION", + "condition" : "MISSING_AGGREGATION", "sqlState" : "42803", "messageParameters" : { "expression" : "\"x\"", @@ -79,7 +79,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "MISSING_AGGREGATION", + "condition" : "MISSING_AGGREGATION", "sqlState" : "42803", "messageParameters" : { "expression" : "\"b\"", @@ -95,7 +95,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "UNSUPPORTED_FEATURE.LATERAL_COLUMN_ALIAS_IN_GROUP_BY", + "condition" : "UNSUPPORTED_FEATURE.LATERAL_COLUMN_ALIAS_IN_GROUP_BY", "sqlState" : "0A000" } @@ -107,7 +107,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "UNSUPPORTED_FEATURE.LATERAL_COLUMN_ALIAS_IN_GROUP_BY", + "condition" : "UNSUPPORTED_FEATURE.LATERAL_COLUMN_ALIAS_IN_GROUP_BY", "sqlState" : "0A000" } diff --git a/sql/core/src/test/resources/sql-tests/results/columnresolution-negative.sql.out b/sql/core/src/test/resources/sql-tests/results/columnresolution-negative.sql.out index 385ad35002be..a56df4130938 100644 --- a/sql/core/src/test/resources/sql-tests/results/columnresolution-negative.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/columnresolution-negative.sql.out @@ -70,7 +70,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "AMBIGUOUS_REFERENCE", + "condition" : "AMBIGUOUS_REFERENCE", "sqlState" : "42704", "messageParameters" : { "name" : "`i1`", @@ -93,7 +93,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "AMBIGUOUS_REFERENCE", + "condition" : "AMBIGUOUS_REFERENCE", "sqlState" : "42704", "messageParameters" : { "name" : "`t1`.`i1`", @@ -116,7 +116,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "AMBIGUOUS_REFERENCE", + "condition" : "AMBIGUOUS_REFERENCE", "sqlState" : "42704", "messageParameters" : { "name" : "`mydb1`.`t1`.`i1`", @@ -139,7 +139,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "AMBIGUOUS_REFERENCE", + "condition" : "AMBIGUOUS_REFERENCE", "sqlState" : "42704", "messageParameters" : { "name" : "`i1`", @@ -162,7 +162,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "AMBIGUOUS_REFERENCE", + "condition" : "AMBIGUOUS_REFERENCE", "sqlState" : "42704", "messageParameters" : { "name" : "`t1`.`i1`", @@ -193,7 +193,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "AMBIGUOUS_REFERENCE", + "condition" : "AMBIGUOUS_REFERENCE", "sqlState" : "42704", "messageParameters" : { "name" : "`i1`", @@ -216,7 +216,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "AMBIGUOUS_REFERENCE", + "condition" : "AMBIGUOUS_REFERENCE", "sqlState" : "42704", "messageParameters" : { "name" : "`t1`.`i1`", @@ -239,7 +239,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "AMBIGUOUS_REFERENCE", + "condition" : "AMBIGUOUS_REFERENCE", "sqlState" : "42704", "messageParameters" : { "name" : "`i1`", @@ -262,7 +262,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "AMBIGUOUS_REFERENCE", + "condition" : "AMBIGUOUS_REFERENCE", "sqlState" : "42704", "messageParameters" : { "name" : "`t1`.`i1`", @@ -285,7 +285,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`db1`.`t1`.`i1`", @@ -324,7 +324,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`mydb1`.`t1`", @@ -347,7 +347,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "CANNOT_RESOLVE_STAR_EXPAND", + "condition" : "CANNOT_RESOLVE_STAR_EXPAND", "sqlState" : "42704", "messageParameters" : { "columns" : "`i1`", @@ -370,7 +370,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`t1`", @@ -401,7 +401,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`mydb1`.`t1`.`i1`", @@ -456,7 +456,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "INCOMPATIBLE_VIEW_SCHEMA_CHANGE", + "condition" : "INCOMPATIBLE_VIEW_SCHEMA_CHANGE", "sqlState" : "51024", "messageParameters" : { "actualCols" : "[]", @@ -507,7 +507,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "INCOMPATIBLE_VIEW_SCHEMA_CHANGE", + "condition" : "INCOMPATIBLE_VIEW_SCHEMA_CHANGE", "sqlState" : "51024", "messageParameters" : { "actualCols" : "[]", diff --git a/sql/core/src/test/resources/sql-tests/results/comments.sql.out b/sql/core/src/test/resources/sql-tests/results/comments.sql.out index 1e2fc99a798c..93ff8c25892b 100644 --- a/sql/core/src/test/resources/sql-tests/results/comments.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/comments.sql.out @@ -133,7 +133,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "UNCLOSED_BRACKETED_COMMENT", + "condition" : "UNCLOSED_BRACKETED_COMMENT", "sqlState" : "42601" } @@ -154,6 +154,6 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "UNCLOSED_BRACKETED_COMMENT", + "condition" : "UNCLOSED_BRACKETED_COMMENT", "sqlState" : "42601" } diff --git a/sql/core/src/test/resources/sql-tests/results/conditional-functions.sql.out b/sql/core/src/test/resources/sql-tests/results/conditional-functions.sql.out index aa8a600f8756..8d1c315115c7 100644 --- a/sql/core/src/test/resources/sql-tests/results/conditional-functions.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/conditional-functions.sql.out @@ -142,7 +142,7 @@ struct<> -- !query output org.apache.spark.SparkNumberFormatException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'abc'", @@ -177,7 +177,7 @@ struct<> -- !query output org.apache.spark.SparkNumberFormatException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'abc'", diff --git a/sql/core/src/test/resources/sql-tests/results/count.sql.out b/sql/core/src/test/resources/sql-tests/results/count.sql.out index 042092279929..6dc87f7d5cdc 100644 --- a/sql/core/src/test/resources/sql-tests/results/count.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/count.sql.out @@ -147,7 +147,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "WRONG_NUM_ARGS.WITH_SUGGESTION", + "condition" : "WRONG_NUM_ARGS.WITH_SUGGESTION", "sqlState" : "42605", "messageParameters" : { "actualNum" : "0", @@ -191,7 +191,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "_LEGACY_ERROR_TEMP_1021", + "condition" : "_LEGACY_ERROR_TEMP_1021", "messageParameters" : { "targetString" : "testData" } diff --git a/sql/core/src/test/resources/sql-tests/results/csv-functions.sql.out b/sql/core/src/test/resources/sql-tests/results/csv-functions.sql.out index 19b24cc12017..cd9542d7e7aa 100644 --- a/sql/core/src/test/resources/sql-tests/results/csv-functions.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/csv-functions.sql.out @@ -22,7 +22,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_SCHEMA.NON_STRING_LITERAL", + "condition" : "INVALID_SCHEMA.NON_STRING_LITERAL", "sqlState" : "42K07", "messageParameters" : { "inputSchema" : "\"1\"" @@ -44,7 +44,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'InvalidType'", @@ -67,7 +67,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_SCHEMA.NON_STRUCT_TYPE", + "condition" : "INVALID_SCHEMA.NON_STRUCT_TYPE", "sqlState" : "42K07", "messageParameters" : { "dataType" : "\"ARRAY\"", @@ -90,7 +90,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_OPTIONS.NON_MAP_FUNCTION", + "condition" : "INVALID_OPTIONS.NON_MAP_FUNCTION", "sqlState" : "42K06", "queryContext" : [ { "objectType" : "", @@ -109,7 +109,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_OPTIONS.NON_STRING_TYPE", + "condition" : "INVALID_OPTIONS.NON_STRING_TYPE", "sqlState" : "42K06", "messageParameters" : { "mapType" : "\"MAP\"" @@ -131,7 +131,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", + "condition" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", "sqlState" : "42605", "messageParameters" : { "actualNum" : "0", @@ -172,7 +172,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_NULL", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_NULL", "sqlState" : "42K09", "messageParameters" : { "exprName" : "csv", @@ -203,7 +203,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.NON_FOLDABLE_INPUT", + "condition" : "DATATYPE_MISMATCH.NON_FOLDABLE_INPUT", "sqlState" : "42K09", "messageParameters" : { "inputExpr" : "\"csvField\"", @@ -252,7 +252,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_OPTIONS.NON_MAP_FUNCTION", + "condition" : "INVALID_OPTIONS.NON_MAP_FUNCTION", "sqlState" : "42K06", "queryContext" : [ { "objectType" : "", @@ -271,7 +271,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_OPTIONS.NON_STRING_TYPE", + "condition" : "INVALID_OPTIONS.NON_STRING_TYPE", "sqlState" : "42K06", "messageParameters" : { "mapType" : "\"MAP\"" diff --git a/sql/core/src/test/resources/sql-tests/results/cte-legacy.sql.out b/sql/core/src/test/resources/sql-tests/results/cte-legacy.sql.out index 1255e8b51f30..f4838f5eb39c 100644 --- a/sql/core/src/test/resources/sql-tests/results/cte-legacy.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/cte-legacy.sql.out @@ -68,7 +68,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "TABLE_OR_VIEW_NOT_FOUND", + "condition" : "TABLE_OR_VIEW_NOT_FOUND", "sqlState" : "42P01", "messageParameters" : { "relationName" : "`cte`" @@ -283,7 +283,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "TABLE_OR_VIEW_NOT_FOUND", + "condition" : "TABLE_OR_VIEW_NOT_FOUND", "sqlState" : "42P01", "messageParameters" : { "relationName" : "`t1`" @@ -334,7 +334,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "TABLE_OR_VIEW_NOT_FOUND", + "condition" : "TABLE_OR_VIEW_NOT_FOUND", "sqlState" : "42P01", "messageParameters" : { "relationName" : "`cte_outer`" @@ -367,7 +367,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "TABLE_OR_VIEW_NOT_FOUND", + "condition" : "TABLE_OR_VIEW_NOT_FOUND", "sqlState" : "42P01", "messageParameters" : { "relationName" : "`cte_invisible_inner`" @@ -402,7 +402,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "TABLE_OR_VIEW_NOT_FOUND", + "condition" : "TABLE_OR_VIEW_NOT_FOUND", "sqlState" : "42P01", "messageParameters" : { "relationName" : "`cte_invisible_inner`" diff --git a/sql/core/src/test/resources/sql-tests/results/cte-nested.sql.out b/sql/core/src/test/resources/sql-tests/results/cte-nested.sql.out index 7cf488ce8cad..8190a4d222e2 100644 --- a/sql/core/src/test/resources/sql-tests/results/cte-nested.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/cte-nested.sql.out @@ -68,7 +68,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "TABLE_OR_VIEW_NOT_FOUND", + "condition" : "TABLE_OR_VIEW_NOT_FOUND", "sqlState" : "42P01", "messageParameters" : { "relationName" : "`cte`" @@ -339,7 +339,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "TABLE_OR_VIEW_NOT_FOUND", + "condition" : "TABLE_OR_VIEW_NOT_FOUND", "sqlState" : "42P01", "messageParameters" : { "relationName" : "`cte_invisible_inner`" @@ -374,7 +374,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "TABLE_OR_VIEW_NOT_FOUND", + "condition" : "TABLE_OR_VIEW_NOT_FOUND", "sqlState" : "42P01", "messageParameters" : { "relationName" : "`cte_invisible_inner`" diff --git a/sql/core/src/test/resources/sql-tests/results/cte-nonlegacy.sql.out b/sql/core/src/test/resources/sql-tests/results/cte-nonlegacy.sql.out index 94ef47397eff..71dee1187e30 100644 --- a/sql/core/src/test/resources/sql-tests/results/cte-nonlegacy.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/cte-nonlegacy.sql.out @@ -68,7 +68,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "TABLE_OR_VIEW_NOT_FOUND", + "condition" : "TABLE_OR_VIEW_NOT_FOUND", "sqlState" : "42P01", "messageParameters" : { "relationName" : "`cte`" @@ -96,7 +96,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "AMBIGUOUS_ALIAS_IN_NESTED_CTE", + "condition" : "AMBIGUOUS_ALIAS_IN_NESTED_CTE", "sqlState" : "42KD0", "messageParameters" : { "config" : "\"spark.sql.legacy.ctePrecedencePolicy\"", @@ -141,7 +141,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "AMBIGUOUS_ALIAS_IN_NESTED_CTE", + "condition" : "AMBIGUOUS_ALIAS_IN_NESTED_CTE", "sqlState" : "42KD0", "messageParameters" : { "config" : "\"spark.sql.legacy.ctePrecedencePolicy\"", @@ -203,7 +203,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "AMBIGUOUS_ALIAS_IN_NESTED_CTE", + "condition" : "AMBIGUOUS_ALIAS_IN_NESTED_CTE", "sqlState" : "42KD0", "messageParameters" : { "config" : "\"spark.sql.legacy.ctePrecedencePolicy\"", @@ -226,7 +226,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "AMBIGUOUS_ALIAS_IN_NESTED_CTE", + "condition" : "AMBIGUOUS_ALIAS_IN_NESTED_CTE", "sqlState" : "42KD0", "messageParameters" : { "config" : "\"spark.sql.legacy.ctePrecedencePolicy\"", @@ -250,7 +250,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "AMBIGUOUS_ALIAS_IN_NESTED_CTE", + "condition" : "AMBIGUOUS_ALIAS_IN_NESTED_CTE", "sqlState" : "42KD0", "messageParameters" : { "config" : "\"spark.sql.legacy.ctePrecedencePolicy\"", @@ -272,7 +272,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "AMBIGUOUS_ALIAS_IN_NESTED_CTE", + "condition" : "AMBIGUOUS_ALIAS_IN_NESTED_CTE", "sqlState" : "42KD0", "messageParameters" : { "config" : "\"spark.sql.legacy.ctePrecedencePolicy\"", @@ -309,7 +309,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "AMBIGUOUS_ALIAS_IN_NESTED_CTE", + "condition" : "AMBIGUOUS_ALIAS_IN_NESTED_CTE", "sqlState" : "42KD0", "messageParameters" : { "config" : "\"spark.sql.legacy.ctePrecedencePolicy\"", @@ -330,7 +330,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "AMBIGUOUS_ALIAS_IN_NESTED_CTE", + "condition" : "AMBIGUOUS_ALIAS_IN_NESTED_CTE", "sqlState" : "42KD0", "messageParameters" : { "config" : "\"spark.sql.legacy.ctePrecedencePolicy\"", @@ -411,7 +411,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "TABLE_OR_VIEW_NOT_FOUND", + "condition" : "TABLE_OR_VIEW_NOT_FOUND", "sqlState" : "42P01", "messageParameters" : { "relationName" : "`cte_invisible_inner`" @@ -446,7 +446,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "TABLE_OR_VIEW_NOT_FOUND", + "condition" : "TABLE_OR_VIEW_NOT_FOUND", "sqlState" : "42P01", "messageParameters" : { "relationName" : "`cte_invisible_inner`" diff --git a/sql/core/src/test/resources/sql-tests/results/cte.sql.out b/sql/core/src/test/resources/sql-tests/results/cte.sql.out index 4367ae1d5f2d..8b316207250e 100644 --- a/sql/core/src/test/resources/sql-tests/results/cte.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/cte.sql.out @@ -30,7 +30,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "TABLE_OR_VIEW_NOT_FOUND", + "condition" : "TABLE_OR_VIEW_NOT_FOUND", "sqlState" : "42P01", "messageParameters" : { "relationName" : "`s`" @@ -53,7 +53,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "TABLE_OR_VIEW_NOT_FOUND", + "condition" : "TABLE_OR_VIEW_NOT_FOUND", "sqlState" : "42P01", "messageParameters" : { "relationName" : "`r`" @@ -95,7 +95,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "TABLE_OR_VIEW_NOT_FOUND", + "condition" : "TABLE_OR_VIEW_NOT_FOUND", "sqlState" : "42P01", "messageParameters" : { "relationName" : "`s2`" @@ -187,7 +187,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "')'", @@ -206,7 +206,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "DUPLICATED_CTE_NAMES", + "condition" : "DUPLICATED_CTE_NAMES", "sqlState" : "42602", "messageParameters" : { "duplicateNames" : "`t`" @@ -229,7 +229,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "TABLE_OR_VIEW_NOT_FOUND", + "condition" : "TABLE_OR_VIEW_NOT_FOUND", "sqlState" : "42P01", "messageParameters" : { "relationName" : "`non_existing_table`" @@ -290,7 +290,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_INLINE_TABLE.CANNOT_EVALUATE_EXPRESSION_IN_INLINE_TABLE", + "condition" : "INVALID_INLINE_TABLE.CANNOT_EVALUATE_EXPRESSION_IN_INLINE_TABLE", "sqlState" : "42000", "messageParameters" : { "expr" : "\"outer(t.id)\"" @@ -322,7 +322,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "AMBIGUOUS_REFERENCE", + "condition" : "AMBIGUOUS_REFERENCE", "sqlState" : "42704", "messageParameters" : { "name" : "`same_name`.`x`", @@ -355,7 +355,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`id`", @@ -380,7 +380,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "TABLE_OR_VIEW_NOT_FOUND", + "condition" : "TABLE_OR_VIEW_NOT_FOUND", "sqlState" : "42P01", "messageParameters" : { "relationName" : "`r2`" @@ -429,7 +429,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "TABLE_OR_VIEW_NOT_FOUND", + "condition" : "TABLE_OR_VIEW_NOT_FOUND", "sqlState" : "42P01", "messageParameters" : { "relationName" : "`q`" @@ -516,7 +516,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "DUPLICATED_CTE_NAMES", + "condition" : "DUPLICATED_CTE_NAMES", "sqlState" : "42602", "messageParameters" : { "duplicateNames" : "`cte1`" @@ -547,7 +547,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "TABLE_OR_VIEW_NOT_FOUND", + "condition" : "TABLE_OR_VIEW_NOT_FOUND", "sqlState" : "42P01", "messageParameters" : { "relationName" : "`va`" @@ -569,7 +569,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "TABLE_OR_VIEW_NOT_FOUND", + "condition" : "TABLE_OR_VIEW_NOT_FOUND", "sqlState" : "42P01", "messageParameters" : { "relationName" : "`cte`" diff --git a/sql/core/src/test/resources/sql-tests/results/date.sql.out b/sql/core/src/test/resources/sql-tests/results/date.sql.out index 37cf05c6386d..66d9e5419dd3 100644 --- a/sql/core/src/test/resources/sql-tests/results/date.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/date.sql.out @@ -22,7 +22,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_TYPED_LITERAL", + "condition" : "INVALID_TYPED_LITERAL", "sqlState" : "42604", "messageParameters" : { "value" : "'2020-01-01中文'", @@ -53,7 +53,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "DATETIME_FIELD_OUT_OF_BOUNDS", + "condition" : "DATETIME_FIELD_OUT_OF_BOUNDS", "sqlState" : "22023", "messageParameters" : { "ansiConfig" : "\"spark.sql.ansi.enabled\"", @@ -69,7 +69,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "DATETIME_FIELD_OUT_OF_BOUNDS", + "condition" : "DATETIME_FIELD_OUT_OF_BOUNDS", "sqlState" : "22023", "messageParameters" : { "ansiConfig" : "\"spark.sql.ansi.enabled\"", @@ -85,7 +85,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_TYPED_LITERAL", + "condition" : "INVALID_TYPED_LITERAL", "sqlState" : "42604", "messageParameters" : { "value" : "'015'", @@ -108,7 +108,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_TYPED_LITERAL", + "condition" : "INVALID_TYPED_LITERAL", "sqlState" : "42604", "messageParameters" : { "value" : "'2021-4294967297-11'", @@ -147,7 +147,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", + "condition" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", "sqlState" : "42605", "messageParameters" : { "actualNum" : "1", @@ -204,7 +204,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "CANNOT_PARSE_TIMESTAMP", + "condition" : "CANNOT_PARSE_TIMESTAMP", "sqlState" : "22007", "messageParameters" : { "func" : "`try_to_date`", @@ -270,7 +270,7 @@ struct<> -- !query output org.apache.spark.SparkIllegalArgumentException { - "errorClass" : "ILLEGAL_DAY_OF_WEEK", + "condition" : "ILLEGAL_DAY_OF_WEEK", "sqlState" : "22009", "messageParameters" : { "string" : "xx" @@ -309,7 +309,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'xx'", @@ -381,7 +381,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"1\"", @@ -407,7 +407,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"1.0\"", @@ -433,7 +433,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"10.0\"", @@ -467,7 +467,7 @@ struct<> -- !query output org.apache.spark.SparkNumberFormatException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'1.2'", @@ -555,7 +555,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"1\"", @@ -581,7 +581,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"1.0\"", @@ -607,7 +607,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"10.0\"", @@ -641,7 +641,7 @@ struct<> -- !query output org.apache.spark.SparkNumberFormatException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'1.2'", @@ -729,7 +729,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"10.0\"", @@ -835,7 +835,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"1\"", @@ -861,7 +861,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"DATE '2011-11-11'\"", @@ -916,7 +916,7 @@ struct<> -- !query output org.apache.spark.SparkUpgradeException { - "errorClass" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION.DATETIME_PATTERN_RECOGNITION", + "condition" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION.DATETIME_PATTERN_RECOGNITION", "sqlState" : "42K0B", "messageParameters" : { "config" : "\"spark.sql.legacy.timeParserPolicy\"", @@ -933,7 +933,7 @@ struct<> -- !query output org.apache.spark.SparkUpgradeException { - "errorClass" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION.DATETIME_PATTERN_RECOGNITION", + "condition" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION.DATETIME_PATTERN_RECOGNITION", "sqlState" : "42K0B", "messageParameters" : { "config" : "\"spark.sql.legacy.timeParserPolicy\"", @@ -950,7 +950,7 @@ struct<> -- !query output org.apache.spark.SparkUpgradeException { - "errorClass" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION.DATETIME_PATTERN_RECOGNITION", + "condition" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION.DATETIME_PATTERN_RECOGNITION", "sqlState" : "42K0B", "messageParameters" : { "config" : "\"spark.sql.legacy.timeParserPolicy\"", @@ -1047,7 +1047,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", + "condition" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", "sqlState" : "22023", "messageParameters" : { "functionName" : "`dateadd`", @@ -1071,7 +1071,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", + "condition" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", "sqlState" : "22023", "messageParameters" : { "functionName" : "`date_add`", @@ -1175,7 +1175,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", + "condition" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", "sqlState" : "22023", "messageParameters" : { "functionName" : "`date_diff`", @@ -1199,7 +1199,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", + "condition" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", "sqlState" : "22023", "messageParameters" : { "functionName" : "`datediff`", diff --git a/sql/core/src/test/resources/sql-tests/results/datetime-formatting-invalid.sql.out b/sql/core/src/test/resources/sql-tests/results/datetime-formatting-invalid.sql.out index 66b36befaa33..0826fd26d5be 100644 --- a/sql/core/src/test/resources/sql-tests/results/datetime-formatting-invalid.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/datetime-formatting-invalid.sql.out @@ -6,7 +6,7 @@ struct<> -- !query output org.apache.spark.SparkUpgradeException { - "errorClass" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION.DATETIME_PATTERN_RECOGNITION", + "condition" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION.DATETIME_PATTERN_RECOGNITION", "sqlState" : "42K0B", "messageParameters" : { "config" : "\"spark.sql.legacy.timeParserPolicy\"", @@ -23,7 +23,7 @@ struct<> -- !query output org.apache.spark.SparkUpgradeException { - "errorClass" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION.DATETIME_PATTERN_RECOGNITION", + "condition" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION.DATETIME_PATTERN_RECOGNITION", "sqlState" : "42K0B", "messageParameters" : { "config" : "\"spark.sql.legacy.timeParserPolicy\"", @@ -40,7 +40,7 @@ struct<> -- !query output org.apache.spark.SparkIllegalArgumentException { - "errorClass" : "INVALID_DATETIME_PATTERN.LENGTH", + "condition" : "INVALID_DATETIME_PATTERN.LENGTH", "sqlState" : "22007", "messageParameters" : { "pattern" : "qqqqq" @@ -55,7 +55,7 @@ struct<> -- !query output org.apache.spark.SparkIllegalArgumentException { - "errorClass" : "INVALID_DATETIME_PATTERN.LENGTH", + "condition" : "INVALID_DATETIME_PATTERN.LENGTH", "sqlState" : "22007", "messageParameters" : { "pattern" : "QQQQQ" @@ -70,7 +70,7 @@ struct<> -- !query output org.apache.spark.SparkUpgradeException { - "errorClass" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION.DATETIME_PATTERN_RECOGNITION", + "condition" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION.DATETIME_PATTERN_RECOGNITION", "sqlState" : "42K0B", "messageParameters" : { "config" : "\"spark.sql.legacy.timeParserPolicy\"", @@ -87,7 +87,7 @@ struct<> -- !query output org.apache.spark.SparkUpgradeException { - "errorClass" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION.DATETIME_PATTERN_RECOGNITION", + "condition" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION.DATETIME_PATTERN_RECOGNITION", "sqlState" : "42K0B", "messageParameters" : { "config" : "\"spark.sql.legacy.timeParserPolicy\"", @@ -104,7 +104,7 @@ struct<> -- !query output org.apache.spark.SparkUpgradeException { - "errorClass" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION.DATETIME_PATTERN_RECOGNITION", + "condition" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION.DATETIME_PATTERN_RECOGNITION", "sqlState" : "42K0B", "messageParameters" : { "config" : "\"spark.sql.legacy.timeParserPolicy\"", @@ -121,7 +121,7 @@ struct<> -- !query output org.apache.spark.SparkUpgradeException { - "errorClass" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION.DATETIME_PATTERN_RECOGNITION", + "condition" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION.DATETIME_PATTERN_RECOGNITION", "sqlState" : "42K0B", "messageParameters" : { "config" : "\"spark.sql.legacy.timeParserPolicy\"", @@ -138,7 +138,7 @@ struct<> -- !query output org.apache.spark.SparkUpgradeException { - "errorClass" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION.DATETIME_PATTERN_RECOGNITION", + "condition" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION.DATETIME_PATTERN_RECOGNITION", "sqlState" : "42K0B", "messageParameters" : { "config" : "\"spark.sql.legacy.timeParserPolicy\"", @@ -155,7 +155,7 @@ struct<> -- !query output org.apache.spark.SparkUpgradeException { - "errorClass" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION.DATETIME_PATTERN_RECOGNITION", + "condition" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION.DATETIME_PATTERN_RECOGNITION", "sqlState" : "42K0B", "messageParameters" : { "config" : "\"spark.sql.legacy.timeParserPolicy\"", @@ -172,7 +172,7 @@ struct<> -- !query output org.apache.spark.SparkUpgradeException { - "errorClass" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION.DATETIME_PATTERN_RECOGNITION", + "condition" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION.DATETIME_PATTERN_RECOGNITION", "sqlState" : "42K0B", "messageParameters" : { "config" : "\"spark.sql.legacy.timeParserPolicy\"", @@ -189,7 +189,7 @@ struct<> -- !query output org.apache.spark.SparkUpgradeException { - "errorClass" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION.DATETIME_PATTERN_RECOGNITION", + "condition" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION.DATETIME_PATTERN_RECOGNITION", "sqlState" : "42K0B", "messageParameters" : { "config" : "\"spark.sql.legacy.timeParserPolicy\"", @@ -206,7 +206,7 @@ struct<> -- !query output org.apache.spark.SparkUpgradeException { - "errorClass" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION.DATETIME_PATTERN_RECOGNITION", + "condition" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION.DATETIME_PATTERN_RECOGNITION", "sqlState" : "42K0B", "messageParameters" : { "config" : "\"spark.sql.legacy.timeParserPolicy\"", @@ -223,7 +223,7 @@ struct<> -- !query output org.apache.spark.SparkUpgradeException { - "errorClass" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION.DATETIME_PATTERN_RECOGNITION", + "condition" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION.DATETIME_PATTERN_RECOGNITION", "sqlState" : "42K0B", "messageParameters" : { "config" : "\"spark.sql.legacy.timeParserPolicy\"", @@ -240,7 +240,7 @@ struct<> -- !query output org.apache.spark.SparkUpgradeException { - "errorClass" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION.DATETIME_PATTERN_RECOGNITION", + "condition" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION.DATETIME_PATTERN_RECOGNITION", "sqlState" : "42K0B", "messageParameters" : { "config" : "\"spark.sql.legacy.timeParserPolicy\"", @@ -257,7 +257,7 @@ struct<> -- !query output org.apache.spark.SparkUpgradeException { - "errorClass" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION.DATETIME_PATTERN_RECOGNITION", + "condition" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION.DATETIME_PATTERN_RECOGNITION", "sqlState" : "42K0B", "messageParameters" : { "config" : "\"spark.sql.legacy.timeParserPolicy\"", @@ -274,7 +274,7 @@ struct<> -- !query output org.apache.spark.SparkUpgradeException { - "errorClass" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION.DATETIME_PATTERN_RECOGNITION", + "condition" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION.DATETIME_PATTERN_RECOGNITION", "sqlState" : "42K0B", "messageParameters" : { "config" : "\"spark.sql.legacy.timeParserPolicy\"", @@ -291,7 +291,7 @@ struct<> -- !query output org.apache.spark.SparkUpgradeException { - "errorClass" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION.DATETIME_PATTERN_RECOGNITION", + "condition" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION.DATETIME_PATTERN_RECOGNITION", "sqlState" : "42K0B", "messageParameters" : { "config" : "\"spark.sql.legacy.timeParserPolicy\"", @@ -317,7 +317,7 @@ struct<> -- !query output org.apache.spark.SparkUpgradeException { - "errorClass" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION.DATETIME_PATTERN_RECOGNITION", + "condition" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION.DATETIME_PATTERN_RECOGNITION", "sqlState" : "42K0B", "messageParameters" : { "config" : "\"spark.sql.legacy.timeParserPolicy\"", @@ -343,7 +343,7 @@ struct<> -- !query output org.apache.spark.SparkUpgradeException { - "errorClass" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION.DATETIME_PATTERN_RECOGNITION", + "condition" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION.DATETIME_PATTERN_RECOGNITION", "sqlState" : "42K0B", "messageParameters" : { "config" : "\"spark.sql.legacy.timeParserPolicy\"", @@ -378,7 +378,7 @@ struct<> -- !query output org.apache.spark.SparkIllegalArgumentException { - "errorClass" : "INVALID_DATETIME_PATTERN.ILLEGAL_CHARACTER", + "condition" : "INVALID_DATETIME_PATTERN.ILLEGAL_CHARACTER", "sqlState" : "22007", "messageParameters" : { "c" : "A", @@ -394,7 +394,7 @@ struct<> -- !query output org.apache.spark.SparkIllegalArgumentException { - "errorClass" : "INVALID_DATETIME_PATTERN.ILLEGAL_CHARACTER", + "condition" : "INVALID_DATETIME_PATTERN.ILLEGAL_CHARACTER", "sqlState" : "22007", "messageParameters" : { "c" : "n", @@ -410,7 +410,7 @@ struct<> -- !query output org.apache.spark.SparkIllegalArgumentException { - "errorClass" : "INVALID_DATETIME_PATTERN.ILLEGAL_CHARACTER", + "condition" : "INVALID_DATETIME_PATTERN.ILLEGAL_CHARACTER", "sqlState" : "22007", "messageParameters" : { "c" : "N", @@ -426,7 +426,7 @@ struct<> -- !query output org.apache.spark.SparkIllegalArgumentException { - "errorClass" : "INVALID_DATETIME_PATTERN.ILLEGAL_CHARACTER", + "condition" : "INVALID_DATETIME_PATTERN.ILLEGAL_CHARACTER", "sqlState" : "22007", "messageParameters" : { "c" : "p", @@ -442,7 +442,7 @@ struct<> -- !query output org.apache.spark.SparkUpgradeException { - "errorClass" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION.DATETIME_PATTERN_RECOGNITION", + "condition" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION.DATETIME_PATTERN_RECOGNITION", "sqlState" : "42K0B", "messageParameters" : { "config" : "\"spark.sql.legacy.timeParserPolicy\"", @@ -459,7 +459,7 @@ struct<> -- !query output org.apache.spark.SparkUpgradeException { - "errorClass" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION.DATETIME_PATTERN_RECOGNITION", + "condition" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION.DATETIME_PATTERN_RECOGNITION", "sqlState" : "42K0B", "messageParameters" : { "config" : "\"spark.sql.legacy.timeParserPolicy\"", @@ -476,7 +476,7 @@ struct<> -- !query output org.apache.spark.SparkUpgradeException { - "errorClass" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION.DATETIME_PATTERN_RECOGNITION", + "condition" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION.DATETIME_PATTERN_RECOGNITION", "sqlState" : "42K0B", "messageParameters" : { "config" : "\"spark.sql.legacy.timeParserPolicy\"", @@ -493,7 +493,7 @@ struct<> -- !query output org.apache.spark.SparkUpgradeException { - "errorClass" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION.DATETIME_PATTERN_RECOGNITION", + "condition" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION.DATETIME_PATTERN_RECOGNITION", "sqlState" : "42K0B", "messageParameters" : { "config" : "\"spark.sql.legacy.timeParserPolicy\"", @@ -510,7 +510,7 @@ struct<> -- !query output org.apache.spark.SparkIllegalArgumentException { - "errorClass" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION.DATETIME_WEEK_BASED_PATTERN", + "condition" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION.DATETIME_WEEK_BASED_PATTERN", "sqlState" : "42K0B", "messageParameters" : { "c" : "e" @@ -525,7 +525,7 @@ struct<> -- !query output org.apache.spark.SparkIllegalArgumentException { - "errorClass" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION.DATETIME_WEEK_BASED_PATTERN", + "condition" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION.DATETIME_WEEK_BASED_PATTERN", "sqlState" : "42K0B", "messageParameters" : { "c" : "c" @@ -540,7 +540,7 @@ struct<> -- !query output org.apache.spark.SparkIllegalArgumentException { - "errorClass" : "INVALID_DATETIME_PATTERN.ILLEGAL_CHARACTER", + "condition" : "INVALID_DATETIME_PATTERN.ILLEGAL_CHARACTER", "sqlState" : "22007", "messageParameters" : { "c" : "B", diff --git a/sql/core/src/test/resources/sql-tests/results/datetime-legacy.sql.out b/sql/core/src/test/resources/sql-tests/results/datetime-legacy.sql.out index feae373b52ca..9f68bb87776a 100644 --- a/sql/core/src/test/resources/sql-tests/results/datetime-legacy.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/datetime-legacy.sql.out @@ -22,7 +22,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_TYPED_LITERAL", + "condition" : "INVALID_TYPED_LITERAL", "sqlState" : "42604", "messageParameters" : { "value" : "'2020-01-01中文'", @@ -53,7 +53,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "DATETIME_FIELD_OUT_OF_BOUNDS", + "condition" : "DATETIME_FIELD_OUT_OF_BOUNDS", "sqlState" : "22023", "messageParameters" : { "ansiConfig" : "\"spark.sql.ansi.enabled\"", @@ -69,7 +69,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "DATETIME_FIELD_OUT_OF_BOUNDS", + "condition" : "DATETIME_FIELD_OUT_OF_BOUNDS", "sqlState" : "22023", "messageParameters" : { "ansiConfig" : "\"spark.sql.ansi.enabled\"", @@ -85,7 +85,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_TYPED_LITERAL", + "condition" : "INVALID_TYPED_LITERAL", "sqlState" : "42604", "messageParameters" : { "value" : "'015'", @@ -108,7 +108,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_TYPED_LITERAL", + "condition" : "INVALID_TYPED_LITERAL", "sqlState" : "42604", "messageParameters" : { "value" : "'2021-4294967297-11'", @@ -147,7 +147,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", + "condition" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", "sqlState" : "42605", "messageParameters" : { "actualNum" : "1", @@ -204,7 +204,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "CANNOT_PARSE_TIMESTAMP", + "condition" : "CANNOT_PARSE_TIMESTAMP", "sqlState" : "22007", "messageParameters" : { "func" : "`try_to_date`", @@ -270,7 +270,7 @@ struct<> -- !query output org.apache.spark.SparkIllegalArgumentException { - "errorClass" : "ILLEGAL_DAY_OF_WEEK", + "condition" : "ILLEGAL_DAY_OF_WEEK", "sqlState" : "22009", "messageParameters" : { "string" : "xx" @@ -309,7 +309,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'xx'", @@ -381,7 +381,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"1\"", @@ -407,7 +407,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"1.0\"", @@ -433,7 +433,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"10.0\"", @@ -467,7 +467,7 @@ struct<> -- !query output org.apache.spark.SparkNumberFormatException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'1.2'", @@ -555,7 +555,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"1\"", @@ -581,7 +581,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"1.0\"", @@ -607,7 +607,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"10.0\"", @@ -641,7 +641,7 @@ struct<> -- !query output org.apache.spark.SparkNumberFormatException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'1.2'", @@ -729,7 +729,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"10.0\"", @@ -835,7 +835,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"1\"", @@ -861,7 +861,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"DATE '2011-11-11'\"", @@ -1020,7 +1020,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", + "condition" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", "sqlState" : "22023", "messageParameters" : { "functionName" : "`dateadd`", @@ -1044,7 +1044,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", + "condition" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", "sqlState" : "22023", "messageParameters" : { "functionName" : "`date_add`", @@ -1148,7 +1148,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", + "condition" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", "sqlState" : "22023", "messageParameters" : { "functionName" : "`date_diff`", @@ -1172,7 +1172,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", + "condition" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", "sqlState" : "22023", "messageParameters" : { "functionName" : "`datediff`", @@ -1204,7 +1204,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_TYPED_LITERAL", + "condition" : "INVALID_TYPED_LITERAL", "sqlState" : "42604", "messageParameters" : { "value" : "'2019-01-01中文'", @@ -1227,7 +1227,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_TYPED_LITERAL", + "condition" : "INVALID_TYPED_LITERAL", "sqlState" : "42604", "messageParameters" : { "value" : "'4294967297'", @@ -1250,7 +1250,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_TYPED_LITERAL", + "condition" : "INVALID_TYPED_LITERAL", "sqlState" : "42604", "messageParameters" : { "value" : "'2021-01-01T12:30:4294967297.123456'", @@ -1313,7 +1313,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "INVALID_FRACTION_OF_SECOND", + "condition" : "INVALID_FRACTION_OF_SECOND", "sqlState" : "22023", "messageParameters" : { "secAndMicros" : "60.007" @@ -1344,7 +1344,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "DATETIME_FIELD_OUT_OF_BOUNDS", + "condition" : "DATETIME_FIELD_OUT_OF_BOUNDS", "sqlState" : "22023", "messageParameters" : { "ansiConfig" : "\"spark.sql.ansi.enabled\"", @@ -1376,7 +1376,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "DATETIME_FIELD_OUT_OF_BOUNDS", + "condition" : "DATETIME_FIELD_OUT_OF_BOUNDS", "sqlState" : "22023", "messageParameters" : { "ansiConfig" : "\"spark.sql.ansi.enabled\"", @@ -1392,7 +1392,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "DATETIME_FIELD_OUT_OF_BOUNDS", + "condition" : "DATETIME_FIELD_OUT_OF_BOUNDS", "sqlState" : "22023", "messageParameters" : { "ansiConfig" : "\"spark.sql.ansi.enabled\"", @@ -1582,7 +1582,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "CANNOT_PARSE_TIMESTAMP", + "condition" : "CANNOT_PARSE_TIMESTAMP", "sqlState" : "22007", "messageParameters" : { "func" : "`try_to_timestamp`", @@ -1598,7 +1598,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "CANNOT_PARSE_TIMESTAMP", + "condition" : "CANNOT_PARSE_TIMESTAMP", "sqlState" : "22007", "messageParameters" : { "func" : "`try_to_timestamp`", @@ -1614,7 +1614,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "CANNOT_PARSE_TIMESTAMP", + "condition" : "CANNOT_PARSE_TIMESTAMP", "sqlState" : "22007", "messageParameters" : { "func" : "`try_to_timestamp`", @@ -1630,7 +1630,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "CANNOT_PARSE_TIMESTAMP", + "condition" : "CANNOT_PARSE_TIMESTAMP", "sqlState" : "22007", "messageParameters" : { "func" : "`try_to_timestamp`", @@ -1646,7 +1646,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "CANNOT_PARSE_TIMESTAMP", + "condition" : "CANNOT_PARSE_TIMESTAMP", "sqlState" : "22007", "messageParameters" : { "func" : "`try_to_timestamp`", @@ -1662,7 +1662,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "CANNOT_PARSE_TIMESTAMP", + "condition" : "CANNOT_PARSE_TIMESTAMP", "sqlState" : "22007", "messageParameters" : { "func" : "`try_to_timestamp`", @@ -1678,7 +1678,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "CANNOT_PARSE_TIMESTAMP", + "condition" : "CANNOT_PARSE_TIMESTAMP", "sqlState" : "22007", "messageParameters" : { "func" : "`try_to_timestamp`", @@ -1694,7 +1694,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "CANNOT_PARSE_TIMESTAMP", + "condition" : "CANNOT_PARSE_TIMESTAMP", "sqlState" : "22007", "messageParameters" : { "func" : "`try_to_timestamp`", @@ -1710,7 +1710,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "CANNOT_PARSE_TIMESTAMP", + "condition" : "CANNOT_PARSE_TIMESTAMP", "sqlState" : "22007", "messageParameters" : { "func" : "`try_to_timestamp`", @@ -1726,7 +1726,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "CANNOT_PARSE_TIMESTAMP", + "condition" : "CANNOT_PARSE_TIMESTAMP", "sqlState" : "22007", "messageParameters" : { "func" : "`try_to_timestamp`", @@ -1742,7 +1742,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "CANNOT_PARSE_TIMESTAMP", + "condition" : "CANNOT_PARSE_TIMESTAMP", "sqlState" : "22007", "messageParameters" : { "func" : "`try_to_timestamp`", @@ -1758,7 +1758,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "CANNOT_PARSE_TIMESTAMP", + "condition" : "CANNOT_PARSE_TIMESTAMP", "sqlState" : "22007", "messageParameters" : { "func" : "`try_to_timestamp`", @@ -1774,7 +1774,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "CANNOT_PARSE_TIMESTAMP", + "condition" : "CANNOT_PARSE_TIMESTAMP", "sqlState" : "22007", "messageParameters" : { "func" : "`try_to_timestamp`", @@ -1790,7 +1790,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "CANNOT_PARSE_TIMESTAMP", + "condition" : "CANNOT_PARSE_TIMESTAMP", "sqlState" : "22007", "messageParameters" : { "func" : "`try_to_timestamp`", @@ -1806,7 +1806,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "CANNOT_PARSE_TIMESTAMP", + "condition" : "CANNOT_PARSE_TIMESTAMP", "sqlState" : "22007", "messageParameters" : { "func" : "`try_to_timestamp`", @@ -1822,7 +1822,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "CANNOT_PARSE_TIMESTAMP", + "condition" : "CANNOT_PARSE_TIMESTAMP", "sqlState" : "22007", "messageParameters" : { "func" : "`try_to_timestamp`", @@ -1838,7 +1838,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "CANNOT_PARSE_TIMESTAMP", + "condition" : "CANNOT_PARSE_TIMESTAMP", "sqlState" : "22007", "messageParameters" : { "func" : "`try_to_timestamp`", @@ -1854,7 +1854,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "CANNOT_PARSE_TIMESTAMP", + "condition" : "CANNOT_PARSE_TIMESTAMP", "sqlState" : "22007", "messageParameters" : { "func" : "`try_to_timestamp`", @@ -1870,7 +1870,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "CANNOT_PARSE_TIMESTAMP", + "condition" : "CANNOT_PARSE_TIMESTAMP", "sqlState" : "22007", "messageParameters" : { "func" : "`try_to_timestamp`", @@ -1886,7 +1886,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "CANNOT_PARSE_TIMESTAMP", + "condition" : "CANNOT_PARSE_TIMESTAMP", "sqlState" : "22007", "messageParameters" : { "func" : "`try_to_timestamp`", @@ -1902,7 +1902,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "CANNOT_PARSE_TIMESTAMP", + "condition" : "CANNOT_PARSE_TIMESTAMP", "sqlState" : "22007", "messageParameters" : { "func" : "`try_to_timestamp`", @@ -1974,7 +1974,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "CANNOT_PARSE_TIMESTAMP", + "condition" : "CANNOT_PARSE_TIMESTAMP", "sqlState" : "22007", "messageParameters" : { "func" : "`try_to_timestamp`", @@ -2086,7 +2086,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"TIMESTAMP\"", @@ -2110,7 +2110,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"TIMESTAMP\"", @@ -2134,7 +2134,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP\"", @@ -2158,7 +2158,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"VOID\"", @@ -2205,7 +2205,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "CANNOT_PARSE_TIMESTAMP", + "condition" : "CANNOT_PARSE_TIMESTAMP", "sqlState" : "22007", "messageParameters" : { "func" : "`try_to_timestamp`", @@ -2293,7 +2293,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", + "condition" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", "sqlState" : "22023", "messageParameters" : { "functionName" : "`timestampadd`", @@ -2317,7 +2317,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", + "condition" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", "sqlState" : "22023", "messageParameters" : { "functionName" : "`timestampadd`", @@ -2373,7 +2373,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", + "condition" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", "sqlState" : "22023", "messageParameters" : { "functionName" : "`timestampdiff`", @@ -2397,7 +2397,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", + "condition" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", "sqlState" : "22023", "messageParameters" : { "functionName" : "`timestampdiff`", @@ -2453,7 +2453,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", + "condition" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", "sqlState" : "22023", "messageParameters" : { "functionName" : "`timediff`", @@ -2477,7 +2477,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", + "condition" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", "sqlState" : "22023", "messageParameters" : { "functionName" : "`timediff`", diff --git a/sql/core/src/test/resources/sql-tests/results/datetime-parsing-invalid.sql.out b/sql/core/src/test/resources/sql-tests/results/datetime-parsing-invalid.sql.out index cdc027846c34..3a7537221d98 100644 --- a/sql/core/src/test/resources/sql-tests/results/datetime-parsing-invalid.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/datetime-parsing-invalid.sql.out @@ -15,7 +15,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "CANNOT_PARSE_TIMESTAMP", + "condition" : "CANNOT_PARSE_TIMESTAMP", "sqlState" : "22007", "messageParameters" : { "func" : "`try_to_timestamp`", @@ -31,7 +31,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "CANNOT_PARSE_TIMESTAMP", + "condition" : "CANNOT_PARSE_TIMESTAMP", "sqlState" : "22007", "messageParameters" : { "func" : "`try_to_timestamp`", @@ -47,7 +47,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "CANNOT_PARSE_TIMESTAMP", + "condition" : "CANNOT_PARSE_TIMESTAMP", "sqlState" : "22007", "messageParameters" : { "func" : "`try_to_timestamp`", @@ -63,7 +63,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "CANNOT_PARSE_TIMESTAMP", + "condition" : "CANNOT_PARSE_TIMESTAMP", "sqlState" : "22007", "messageParameters" : { "func" : "`try_to_timestamp`", @@ -79,7 +79,7 @@ struct<> -- !query output org.apache.spark.SparkUpgradeException { - "errorClass" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION.DATETIME_PATTERN_RECOGNITION", + "condition" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION.DATETIME_PATTERN_RECOGNITION", "sqlState" : "42K0B", "messageParameters" : { "config" : "\"spark.sql.legacy.timeParserPolicy\"", @@ -96,7 +96,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "CANNOT_PARSE_TIMESTAMP", + "condition" : "CANNOT_PARSE_TIMESTAMP", "sqlState" : "22007", "messageParameters" : { "func" : "`try_to_timestamp`", @@ -112,7 +112,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "CANNOT_PARSE_TIMESTAMP", + "condition" : "CANNOT_PARSE_TIMESTAMP", "sqlState" : "22007", "messageParameters" : { "func" : "`try_to_timestamp`", @@ -128,7 +128,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "CANNOT_PARSE_TIMESTAMP", + "condition" : "CANNOT_PARSE_TIMESTAMP", "sqlState" : "22007", "messageParameters" : { "func" : "`try_to_timestamp`", @@ -144,7 +144,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "CANNOT_PARSE_TIMESTAMP", + "condition" : "CANNOT_PARSE_TIMESTAMP", "sqlState" : "22007", "messageParameters" : { "func" : "`try_to_timestamp`", @@ -160,7 +160,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "CANNOT_PARSE_TIMESTAMP", + "condition" : "CANNOT_PARSE_TIMESTAMP", "sqlState" : "22007", "messageParameters" : { "func" : "`try_to_timestamp`", @@ -176,7 +176,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "CANNOT_PARSE_TIMESTAMP", + "condition" : "CANNOT_PARSE_TIMESTAMP", "sqlState" : "22007", "messageParameters" : { "func" : "`try_to_timestamp`", @@ -192,7 +192,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "CANNOT_PARSE_TIMESTAMP", + "condition" : "CANNOT_PARSE_TIMESTAMP", "sqlState" : "22007", "messageParameters" : { "func" : "`try_to_timestamp`", @@ -208,7 +208,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "CANNOT_PARSE_TIMESTAMP", + "condition" : "CANNOT_PARSE_TIMESTAMP", "sqlState" : "22007", "messageParameters" : { "func" : "`try_to_timestamp`", @@ -224,7 +224,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "CANNOT_PARSE_TIMESTAMP", + "condition" : "CANNOT_PARSE_TIMESTAMP", "sqlState" : "22007", "messageParameters" : { "func" : "`try_to_timestamp`", @@ -240,7 +240,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "CANNOT_PARSE_TIMESTAMP", + "condition" : "CANNOT_PARSE_TIMESTAMP", "sqlState" : "22007", "messageParameters" : { "func" : "`try_to_timestamp`", @@ -256,7 +256,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "CANNOT_PARSE_TIMESTAMP", + "condition" : "CANNOT_PARSE_TIMESTAMP", "sqlState" : "22007", "messageParameters" : { "func" : "`try_to_timestamp`", @@ -272,7 +272,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "CANNOT_PARSE_TIMESTAMP", + "condition" : "CANNOT_PARSE_TIMESTAMP", "sqlState" : "22007", "messageParameters" : { "func" : "`try_to_timestamp`", @@ -296,7 +296,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "CANNOT_PARSE_TIMESTAMP", + "condition" : "CANNOT_PARSE_TIMESTAMP", "sqlState" : "22007", "messageParameters" : { "func" : "`try_to_date`", @@ -312,7 +312,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "CANNOT_PARSE_TIMESTAMP", + "condition" : "CANNOT_PARSE_TIMESTAMP", "sqlState" : "22007", "messageParameters" : { "func" : "`try_to_date`", @@ -328,7 +328,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "CANNOT_PARSE_TIMESTAMP", + "condition" : "CANNOT_PARSE_TIMESTAMP", "sqlState" : "22007", "messageParameters" : { "func" : "`try_to_timestamp`", @@ -344,7 +344,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "CANNOT_PARSE_TIMESTAMP", + "condition" : "CANNOT_PARSE_TIMESTAMP", "sqlState" : "22007", "messageParameters" : { "func" : "`try_to_timestamp`", @@ -360,7 +360,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "CANNOT_PARSE_TIMESTAMP", + "condition" : "CANNOT_PARSE_TIMESTAMP", "sqlState" : "22007", "messageParameters" : { "func" : "`try_to_timestamp`", @@ -376,7 +376,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "CANNOT_PARSE_TIMESTAMP", + "condition" : "CANNOT_PARSE_TIMESTAMP", "sqlState" : "22007", "messageParameters" : { "func" : "`try_to_timestamp`", @@ -392,7 +392,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "CANNOT_PARSE_TIMESTAMP", + "condition" : "CANNOT_PARSE_TIMESTAMP", "sqlState" : "22007", "messageParameters" : { "func" : "`try_to_timestamp`", @@ -408,7 +408,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "CANNOT_PARSE_TIMESTAMP", + "condition" : "CANNOT_PARSE_TIMESTAMP", "sqlState" : "22007", "messageParameters" : { "func" : "`try_to_timestamp`", @@ -424,7 +424,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'Unparseable'", @@ -448,7 +448,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'Unparseable'", diff --git a/sql/core/src/test/resources/sql-tests/results/decimalArithmeticOperations.sql.out b/sql/core/src/test/resources/sql-tests/results/decimalArithmeticOperations.sql.out index cb52778c420a..2e7ca64bb986 100644 --- a/sql/core/src/test/resources/sql-tests/results/decimalArithmeticOperations.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/decimalArithmeticOperations.sql.out @@ -14,7 +14,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "DIVIDE_BY_ZERO", + "condition" : "DIVIDE_BY_ZERO", "sqlState" : "22012", "messageParameters" : { "config" : "\"spark.sql.ansi.enabled\"" @@ -36,7 +36,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "DIVIDE_BY_ZERO", + "condition" : "DIVIDE_BY_ZERO", "sqlState" : "22012", "messageParameters" : { "config" : "\"spark.sql.ansi.enabled\"" @@ -58,7 +58,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "DIVIDE_BY_ZERO", + "condition" : "DIVIDE_BY_ZERO", "sqlState" : "22012", "messageParameters" : { "config" : "\"spark.sql.ansi.enabled\"" @@ -167,7 +167,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "NUMERIC_VALUE_OUT_OF_RANGE.WITH_SUGGESTION", + "condition" : "NUMERIC_VALUE_OUT_OF_RANGE.WITH_SUGGESTION", "sqlState" : "22003", "messageParameters" : { "config" : "\"spark.sql.ansi.enabled\"", @@ -192,7 +192,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "NUMERIC_VALUE_OUT_OF_RANGE.WITH_SUGGESTION", + "condition" : "NUMERIC_VALUE_OUT_OF_RANGE.WITH_SUGGESTION", "sqlState" : "22003", "messageParameters" : { "config" : "\"spark.sql.ansi.enabled\"", @@ -217,7 +217,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "NUMERIC_VALUE_OUT_OF_RANGE.WITH_SUGGESTION", + "condition" : "NUMERIC_VALUE_OUT_OF_RANGE.WITH_SUGGESTION", "sqlState" : "22003", "messageParameters" : { "config" : "\"spark.sql.ansi.enabled\"", @@ -242,7 +242,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "NUMERIC_VALUE_OUT_OF_RANGE.WITH_SUGGESTION", + "condition" : "NUMERIC_VALUE_OUT_OF_RANGE.WITH_SUGGESTION", "sqlState" : "22003", "messageParameters" : { "config" : "\"spark.sql.ansi.enabled\"", @@ -267,7 +267,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "NUMERIC_VALUE_OUT_OF_RANGE.WITH_SUGGESTION", + "condition" : "NUMERIC_VALUE_OUT_OF_RANGE.WITH_SUGGESTION", "sqlState" : "22003", "messageParameters" : { "config" : "\"spark.sql.ansi.enabled\"", @@ -401,7 +401,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "NUMERIC_VALUE_OUT_OF_RANGE.WITH_SUGGESTION", + "condition" : "NUMERIC_VALUE_OUT_OF_RANGE.WITH_SUGGESTION", "sqlState" : "22003", "messageParameters" : { "config" : "\"spark.sql.ansi.enabled\"", @@ -469,7 +469,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "NUMERIC_VALUE_OUT_OF_RANGE.WITH_SUGGESTION", + "condition" : "NUMERIC_VALUE_OUT_OF_RANGE.WITH_SUGGESTION", "sqlState" : "22003", "messageParameters" : { "config" : "\"spark.sql.ansi.enabled\"", @@ -502,7 +502,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "NUMERIC_VALUE_OUT_OF_RANGE.WITH_SUGGESTION", + "condition" : "NUMERIC_VALUE_OUT_OF_RANGE.WITH_SUGGESTION", "sqlState" : "22003", "messageParameters" : { "config" : "\"spark.sql.ansi.enabled\"", @@ -527,7 +527,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "NUMERIC_VALUE_OUT_OF_RANGE.WITH_SUGGESTION", + "condition" : "NUMERIC_VALUE_OUT_OF_RANGE.WITH_SUGGESTION", "sqlState" : "22003", "messageParameters" : { "config" : "\"spark.sql.ansi.enabled\"", @@ -552,7 +552,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "NUMERIC_VALUE_OUT_OF_RANGE.WITH_SUGGESTION", + "condition" : "NUMERIC_VALUE_OUT_OF_RANGE.WITH_SUGGESTION", "sqlState" : "22003", "messageParameters" : { "config" : "\"spark.sql.ansi.enabled\"", @@ -577,7 +577,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "NUMERIC_VALUE_OUT_OF_RANGE.WITH_SUGGESTION", + "condition" : "NUMERIC_VALUE_OUT_OF_RANGE.WITH_SUGGESTION", "sqlState" : "22003", "messageParameters" : { "config" : "\"spark.sql.ansi.enabled\"", @@ -602,7 +602,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "NUMERIC_VALUE_OUT_OF_RANGE.WITH_SUGGESTION", + "condition" : "NUMERIC_VALUE_OUT_OF_RANGE.WITH_SUGGESTION", "sqlState" : "22003", "messageParameters" : { "config" : "\"spark.sql.ansi.enabled\"", @@ -627,7 +627,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "NUMERIC_VALUE_OUT_OF_RANGE.WITH_SUGGESTION", + "condition" : "NUMERIC_VALUE_OUT_OF_RANGE.WITH_SUGGESTION", "sqlState" : "22003", "messageParameters" : { "config" : "\"spark.sql.ansi.enabled\"", @@ -652,7 +652,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "NUMERIC_VALUE_OUT_OF_RANGE.WITH_SUGGESTION", + "condition" : "NUMERIC_VALUE_OUT_OF_RANGE.WITH_SUGGESTION", "sqlState" : "22003", "messageParameters" : { "config" : "\"spark.sql.ansi.enabled\"", @@ -677,7 +677,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "NUMERIC_VALUE_OUT_OF_RANGE.WITH_SUGGESTION", + "condition" : "NUMERIC_VALUE_OUT_OF_RANGE.WITH_SUGGESTION", "sqlState" : "22003", "messageParameters" : { "config" : "\"spark.sql.ansi.enabled\"", @@ -702,7 +702,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "NUMERIC_VALUE_OUT_OF_RANGE.WITH_SUGGESTION", + "condition" : "NUMERIC_VALUE_OUT_OF_RANGE.WITH_SUGGESTION", "sqlState" : "22003", "messageParameters" : { "config" : "\"spark.sql.ansi.enabled\"", @@ -727,7 +727,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "NUMERIC_VALUE_OUT_OF_RANGE.WITH_SUGGESTION", + "condition" : "NUMERIC_VALUE_OUT_OF_RANGE.WITH_SUGGESTION", "sqlState" : "22003", "messageParameters" : { "config" : "\"spark.sql.ansi.enabled\"", @@ -784,7 +784,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "NUMERIC_VALUE_OUT_OF_RANGE.WITH_SUGGESTION", + "condition" : "NUMERIC_VALUE_OUT_OF_RANGE.WITH_SUGGESTION", "sqlState" : "22003", "messageParameters" : { "config" : "\"spark.sql.ansi.enabled\"", diff --git a/sql/core/src/test/resources/sql-tests/results/describe-query.sql.out b/sql/core/src/test/resources/sql-tests/results/describe-query.sql.out index c02d4b4adaf8..0dd5a1dc9ba1 100644 --- a/sql/core/src/test/resources/sql-tests/results/describe-query.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/describe-query.sql.out @@ -109,7 +109,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'desc_temp1'", @@ -125,7 +125,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'desc_temp1'", @@ -144,7 +144,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'insert'", diff --git a/sql/core/src/test/resources/sql-tests/results/describe.sql.out b/sql/core/src/test/resources/sql-tests/results/describe.sql.out index 70870131163e..d94582319102 100644 --- a/sql/core/src/test/resources/sql-tests/results/describe.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/describe.sql.out @@ -86,7 +86,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "DESCRIBE_JSON_NOT_EXTENDED", + "condition" : "DESCRIBE_JSON_NOT_EXTENDED", "sqlState" : "0A000", "messageParameters" : { "tableName" : "t" @@ -101,7 +101,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "UNSUPPORTED_FEATURE.DESC_TABLE_COLUMN_JSON", + "condition" : "UNSUPPORTED_FEATURE.DESC_TABLE_COLUMN_JSON", "sqlState" : "0A000" } @@ -406,7 +406,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.analysis.NoSuchPartitionException { - "errorClass" : "PARTITIONS_NOT_FOUND", + "condition" : "PARTITIONS_NOT_FOUND", "sqlState" : "428FT", "messageParameters" : { "partitionList" : "PARTITION (`c` = Us, `d` = 2)", @@ -422,7 +422,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "_LEGACY_ERROR_TEMP_1232", + "condition" : "_LEGACY_ERROR_TEMP_1232", "messageParameters" : { "partitionColumnNames" : "c, d", "specKeys" : "c", @@ -438,7 +438,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_SQL_SYNTAX.EMPTY_PARTITION_VALUE", + "condition" : "INVALID_SQL_SYNTAX.EMPTY_PARTITION_VALUE", "sqlState" : "42000", "messageParameters" : { "partKey" : "`d`" @@ -527,7 +527,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "FORBIDDEN_OPERATION", + "condition" : "FORBIDDEN_OPERATION", "sqlState" : "42809", "messageParameters" : { "objectName" : "`temp_v`", @@ -616,7 +616,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "FORBIDDEN_OPERATION", + "condition" : "FORBIDDEN_OPERATION", "sqlState" : "42809", "messageParameters" : { "objectName" : "`v`", diff --git a/sql/core/src/test/resources/sql-tests/results/double-quoted-identifiers-enabled.sql.out b/sql/core/src/test/resources/sql-tests/results/double-quoted-identifiers-enabled.sql.out index 2444c399a87e..b5fddba63538 100644 --- a/sql/core/src/test/resources/sql-tests/results/double-quoted-identifiers-enabled.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/double-quoted-identifiers-enabled.sql.out @@ -6,7 +6,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "TABLE_OR_VIEW_NOT_FOUND", + "condition" : "TABLE_OR_VIEW_NOT_FOUND", "sqlState" : "42P01", "messageParameters" : { "relationName" : "`not_exist`" @@ -28,7 +28,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.analysis.NoSuchNamespaceException { - "errorClass" : "SCHEMA_NOT_FOUND", + "condition" : "SCHEMA_NOT_FOUND", "sqlState" : "42704", "messageParameters" : { "schemaName" : "`spark_catalog`.`not_exist`" @@ -43,7 +43,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "TABLE_OR_VIEW_NOT_FOUND", + "condition" : "TABLE_OR_VIEW_NOT_FOUND", "sqlState" : "42P01", "messageParameters" : { "relationName" : "`not_exist`" @@ -65,7 +65,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "TABLE_OR_VIEW_NOT_FOUND", + "condition" : "TABLE_OR_VIEW_NOT_FOUND", "sqlState" : "42P01", "messageParameters" : { "relationName" : "`not_exist`" @@ -87,7 +87,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "TABLE_OR_VIEW_NOT_FOUND", + "condition" : "TABLE_OR_VIEW_NOT_FOUND", "sqlState" : "42P01", "messageParameters" : { "relationName" : "`not_exist`" @@ -109,7 +109,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "TABLE_OR_VIEW_NOT_FOUND", + "condition" : "TABLE_OR_VIEW_NOT_FOUND", "sqlState" : "42P01", "messageParameters" : { "relationName" : "`not_exist`" @@ -131,7 +131,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "UNRESOLVED_ROUTINE", + "condition" : "UNRESOLVED_ROUTINE", "sqlState" : "42883", "messageParameters" : { "routineName" : "`not_exist`", @@ -154,7 +154,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "UNRESOLVED_ROUTINE", + "condition" : "UNRESOLVED_ROUTINE", "sqlState" : "42883", "messageParameters" : { "routineName" : "`not_exist`.`not_exist`", @@ -177,7 +177,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "TABLE_OR_VIEW_NOT_FOUND", + "condition" : "TABLE_OR_VIEW_NOT_FOUND", "sqlState" : "42P01", "messageParameters" : { "relationName" : "`hello`" @@ -199,7 +199,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.analysis.NoSuchNamespaceException { - "errorClass" : "SCHEMA_NOT_FOUND", + "condition" : "SCHEMA_NOT_FOUND", "sqlState" : "42704", "messageParameters" : { "schemaName" : "`spark_catalog`.`not_exist`" @@ -214,7 +214,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "TABLE_OR_VIEW_NOT_FOUND", + "condition" : "TABLE_OR_VIEW_NOT_FOUND", "sqlState" : "42P01", "messageParameters" : { "relationName" : "`not_exist`" @@ -236,7 +236,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "TABLE_OR_VIEW_NOT_FOUND", + "condition" : "TABLE_OR_VIEW_NOT_FOUND", "sqlState" : "42P01", "messageParameters" : { "relationName" : "`not_exist`" @@ -258,7 +258,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "TABLE_OR_VIEW_NOT_FOUND", + "condition" : "TABLE_OR_VIEW_NOT_FOUND", "sqlState" : "42P01", "messageParameters" : { "relationName" : "`not_exist`" @@ -280,7 +280,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "TABLE_OR_VIEW_NOT_FOUND", + "condition" : "TABLE_OR_VIEW_NOT_FOUND", "sqlState" : "42P01", "messageParameters" : { "relationName" : "`not_exist`" @@ -302,7 +302,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "UNRESOLVED_ROUTINE", + "condition" : "UNRESOLVED_ROUTINE", "sqlState" : "42883", "messageParameters" : { "routineName" : "`not_exist`", @@ -325,7 +325,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "UNRESOLVED_ROUTINE", + "condition" : "UNRESOLVED_ROUTINE", "sqlState" : "42883", "messageParameters" : { "routineName" : "`not_exist`.`not_exist`", @@ -348,7 +348,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITHOUT_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITHOUT_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`hello`" @@ -370,7 +370,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'\"hello\"'", @@ -386,7 +386,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.analysis.NoSuchTableException { - "errorClass" : "TABLE_OR_VIEW_NOT_FOUND", + "condition" : "TABLE_OR_VIEW_NOT_FOUND", "sqlState" : "42P01", "messageParameters" : { "relationName" : "`spark_catalog`.`default`.`v`" @@ -401,7 +401,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'\"1\"'", diff --git a/sql/core/src/test/resources/sql-tests/results/double-quoted-identifiers.sql.out b/sql/core/src/test/resources/sql-tests/results/double-quoted-identifiers.sql.out index 81a98a60590f..db6096044dce 100644 --- a/sql/core/src/test/resources/sql-tests/results/double-quoted-identifiers.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/double-quoted-identifiers.sql.out @@ -6,7 +6,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'\"not_exist\"'", @@ -22,7 +22,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'\"not_exist\"'", @@ -38,7 +38,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'\"not_exist\"'", @@ -54,7 +54,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'\"not_exist\"'", @@ -70,7 +70,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'\"not_exist\"'", @@ -86,7 +86,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'\"hello\"'", @@ -102,7 +102,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'\"not_exist\"'", @@ -118,7 +118,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'\"not_exist\"'", @@ -134,7 +134,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "TABLE_OR_VIEW_NOT_FOUND", + "condition" : "TABLE_OR_VIEW_NOT_FOUND", "sqlState" : "42P01", "messageParameters" : { "relationName" : "`hello`" @@ -156,7 +156,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.analysis.NoSuchNamespaceException { - "errorClass" : "SCHEMA_NOT_FOUND", + "condition" : "SCHEMA_NOT_FOUND", "sqlState" : "42704", "messageParameters" : { "schemaName" : "`spark_catalog`.`not_exist`" @@ -171,7 +171,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "TABLE_OR_VIEW_NOT_FOUND", + "condition" : "TABLE_OR_VIEW_NOT_FOUND", "sqlState" : "42P01", "messageParameters" : { "relationName" : "`not_exist`" @@ -193,7 +193,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "TABLE_OR_VIEW_NOT_FOUND", + "condition" : "TABLE_OR_VIEW_NOT_FOUND", "sqlState" : "42P01", "messageParameters" : { "relationName" : "`not_exist`" @@ -215,7 +215,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "TABLE_OR_VIEW_NOT_FOUND", + "condition" : "TABLE_OR_VIEW_NOT_FOUND", "sqlState" : "42P01", "messageParameters" : { "relationName" : "`not_exist`" @@ -237,7 +237,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "TABLE_OR_VIEW_NOT_FOUND", + "condition" : "TABLE_OR_VIEW_NOT_FOUND", "sqlState" : "42P01", "messageParameters" : { "relationName" : "`not_exist`" @@ -259,7 +259,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "UNRESOLVED_ROUTINE", + "condition" : "UNRESOLVED_ROUTINE", "sqlState" : "42883", "messageParameters" : { "routineName" : "`not_exist`", @@ -282,7 +282,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "UNRESOLVED_ROUTINE", + "condition" : "UNRESOLVED_ROUTINE", "sqlState" : "42883", "messageParameters" : { "routineName" : "`not_exist`.`not_exist`", @@ -369,7 +369,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'\"myschema\"'", @@ -386,7 +386,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'\"myview\"'", @@ -402,7 +402,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'\"a2\"'", @@ -418,7 +418,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'\"myview\"'", @@ -434,7 +434,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'\"myschema\"'", diff --git a/sql/core/src/test/resources/sql-tests/results/except-all.sql.out b/sql/core/src/test/resources/sql-tests/results/except-all.sql.out index 437a5553d385..701269db94a2 100644 --- a/sql/core/src/test/resources/sql-tests/results/except-all.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/except-all.sql.out @@ -139,7 +139,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -229,7 +229,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "NUM_COLUMNS_MISMATCH", + "condition" : "NUM_COLUMNS_MISMATCH", "sqlState" : "42826", "messageParameters" : { "firstNumColumns" : "1", diff --git a/sql/core/src/test/resources/sql-tests/results/execute-immediate.sql.out b/sql/core/src/test/resources/sql-tests/results/execute-immediate.sql.out index 21ea4436f4fa..d9ecc1bbf186 100644 --- a/sql/core/src/test/resources/sql-tests/results/execute-immediate.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/execute-immediate.sql.out @@ -389,7 +389,7 @@ struct<> -- !query output org.apache.spark.SparkNumberFormatException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'invalid_cast_error_expected'", @@ -413,7 +413,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_STATEMENT_FOR_EXECUTE_INTO", + "condition" : "INVALID_STATEMENT_FOR_EXECUTE_INTO", "sqlState" : "07501", "messageParameters" : { "sqlString" : "INSERT INTO X VALUES (?)" @@ -428,7 +428,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "UNRESOLVED_VARIABLE", + "condition" : "UNRESOLVED_VARIABLE", "sqlState" : "42883", "messageParameters" : { "searchPath" : "`system`.`session`", @@ -451,7 +451,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_QUERY_MIXED_QUERY_PARAMETERS", + "condition" : "INVALID_QUERY_MIXED_QUERY_PARAMETERS", "sqlState" : "42613" } @@ -463,7 +463,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "ALL_PARAMETERS_MUST_BE_NAMED", + "condition" : "ALL_PARAMETERS_MUST_BE_NAMED", "sqlState" : "07001", "messageParameters" : { "exprs" : "\"1\"" @@ -478,7 +478,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'SELCT'", @@ -501,7 +501,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'SELCT'", @@ -524,7 +524,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_VARIABLE_TYPE_FOR_QUERY_EXECUTE_IMMEDIATE", + "condition" : "INVALID_VARIABLE_TYPE_FOR_QUERY_EXECUTE_IMMEDIATE", "sqlState" : "42K09", "messageParameters" : { "varType" : "\"INT\"" @@ -555,7 +555,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "UNSUPPORTED_EXPR_FOR_PARAMETER", + "condition" : "UNSUPPORTED_EXPR_FOR_PARAMETER", "sqlState" : "42K0E", "messageParameters" : { "invalidExprSql" : "\"CONCAT(a, me1)\"" @@ -577,7 +577,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "UNSUPPORTED_EXPR_FOR_PARAMETER", + "condition" : "UNSUPPORTED_EXPR_FOR_PARAMETER", "sqlState" : "42K0E", "messageParameters" : { "invalidExprSql" : "\"scalarsubquery()\"" @@ -599,7 +599,7 @@ struct<> -- !query output org.apache.spark.SparkNumberFormatException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'name1'", @@ -623,7 +623,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'('", @@ -639,7 +639,7 @@ struct<> -- !query output org.apache.spark.SparkException { - "errorClass" : "ROW_SUBQUERY_TOO_MANY_ROWS", + "condition" : "ROW_SUBQUERY_TOO_MANY_ROWS", "sqlState" : "21000" } @@ -651,7 +651,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "ASSIGNMENT_ARITY_MISMATCH", + "condition" : "ASSIGNMENT_ARITY_MISMATCH", "sqlState" : "42802", "messageParameters" : { "numExpr" : "2", @@ -667,7 +667,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "ASSIGNMENT_ARITY_MISMATCH", + "condition" : "ASSIGNMENT_ARITY_MISMATCH", "sqlState" : "42802", "messageParameters" : { "numExpr" : "1", @@ -683,7 +683,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "EXEC_IMMEDIATE_DUPLICATE_ARGUMENT_ALIASES", + "condition" : "EXEC_IMMEDIATE_DUPLICATE_ARGUMENT_ALIASES", "sqlState" : "42701", "messageParameters" : { "aliases" : "`first`" @@ -721,7 +721,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "ALL_PARAMETERS_MUST_BE_NAMED", + "condition" : "ALL_PARAMETERS_MUST_BE_NAMED", "sqlState" : "07001", "messageParameters" : { "exprs" : "\"p\"" @@ -736,7 +736,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "DUPLICATE_ASSIGNMENTS", + "condition" : "DUPLICATE_ASSIGNMENTS", "sqlState" : "42701", "messageParameters" : { "nameList" : "`res_id`" @@ -751,7 +751,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "NESTED_EXECUTE_IMMEDIATE", + "condition" : "NESTED_EXECUTE_IMMEDIATE", "sqlState" : "07501", "messageParameters" : { "sqlString" : "EXECUTE IMMEDIATE 'SELECT ID FROM TBL_VIEW WHERE ID = ? USING 10'" @@ -774,7 +774,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "NULL_QUERY_STRING_EXECUTE_IMMEDIATE", + "condition" : "NULL_QUERY_STRING_EXECUTE_IMMEDIATE", "sqlState" : "22004", "messageParameters" : { "varName" : "`sql_string`" diff --git a/sql/core/src/test/resources/sql-tests/results/extract.sql.out b/sql/core/src/test/resources/sql-tests/results/extract.sql.out index 02a7315be7a6..cf061d29f559 100644 --- a/sql/core/src/test/resources/sql-tests/results/extract.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/extract.sql.out @@ -318,7 +318,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_EXTRACT_FIELD", + "condition" : "INVALID_EXTRACT_FIELD", "sqlState" : "42601", "messageParameters" : { "expr" : "\"c\"", @@ -341,7 +341,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_EXTRACT_FIELD", + "condition" : "INVALID_EXTRACT_FIELD", "sqlState" : "42601", "messageParameters" : { "expr" : "\"i\"", @@ -364,7 +364,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_EXTRACT_FIELD", + "condition" : "INVALID_EXTRACT_FIELD", "sqlState" : "42601", "messageParameters" : { "expr" : "\"j\"", @@ -691,7 +691,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_EXTRACT_FIELD", + "condition" : "INVALID_EXTRACT_FIELD", "sqlState" : "42601", "messageParameters" : { "expr" : "\"c\"", @@ -714,7 +714,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "NON_FOLDABLE_ARGUMENT", + "condition" : "NON_FOLDABLE_ARGUMENT", "sqlState" : "42K08", "messageParameters" : { "funcName" : "`date_part`", @@ -746,7 +746,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "NON_FOLDABLE_ARGUMENT", + "condition" : "NON_FOLDABLE_ARGUMENT", "sqlState" : "42K08", "messageParameters" : { "funcName" : "`date_part`", @@ -970,7 +970,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_EXTRACT_FIELD", + "condition" : "INVALID_EXTRACT_FIELD", "sqlState" : "42601", "messageParameters" : { "expr" : "\"INTERVAL '2-1' YEAR TO MONTH\"", @@ -993,7 +993,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_EXTRACT_FIELD", + "condition" : "INVALID_EXTRACT_FIELD", "sqlState" : "42601", "messageParameters" : { "expr" : "\"INTERVAL '2-1' YEAR TO MONTH\"", @@ -1016,7 +1016,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_EXTRACT_FIELD", + "condition" : "INVALID_EXTRACT_FIELD", "sqlState" : "42601", "messageParameters" : { "expr" : "\"INTERVAL '2-1' YEAR TO MONTH\"", @@ -1143,7 +1143,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_EXTRACT_FIELD", + "condition" : "INVALID_EXTRACT_FIELD", "sqlState" : "42601", "messageParameters" : { "expr" : "\"INTERVAL '123 12:34:56.789123' DAY TO SECOND\"", @@ -1166,7 +1166,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_EXTRACT_FIELD", + "condition" : "INVALID_EXTRACT_FIELD", "sqlState" : "42601", "messageParameters" : { "expr" : "\"INTERVAL '123 12:34:56.789123' DAY TO SECOND\"", diff --git a/sql/core/src/test/resources/sql-tests/results/group-analytics.sql.out b/sql/core/src/test/resources/sql-tests/results/group-analytics.sql.out index f7f76242a4e6..39f26f63b722 100644 --- a/sql/core/src/test/resources/sql-tests/results/group-analytics.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/group-analytics.sql.out @@ -132,7 +132,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0051", + "condition" : "_LEGACY_ERROR_TEMP_0051", "messageParameters" : { "element" : "ROLLUP" }, @@ -201,7 +201,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0051", + "condition" : "_LEGACY_ERROR_TEMP_0051", "messageParameters" : { "element" : "CUBE" }, @@ -466,7 +466,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNSUPPORTED_GROUPING_EXPRESSION", + "condition" : "UNSUPPORTED_GROUPING_EXPRESSION", "sqlState" : "42K0E", "queryContext" : [ { "objectType" : "", @@ -485,7 +485,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNSUPPORTED_GROUPING_EXPRESSION", + "condition" : "UNSUPPORTED_GROUPING_EXPRESSION", "sqlState" : "42K0E", "queryContext" : [ { "objectType" : "", @@ -531,7 +531,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "UNSUPPORTED_GROUPING_EXPRESSION", + "condition" : "UNSUPPORTED_GROUPING_EXPRESSION", "sqlState" : "42K0E" } @@ -543,7 +543,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "UNSUPPORTED_GROUPING_EXPRESSION", + "condition" : "UNSUPPORTED_GROUPING_EXPRESSION", "sqlState" : "42K0E" } @@ -601,7 +601,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "UNSUPPORTED_GROUPING_EXPRESSION", + "condition" : "UNSUPPORTED_GROUPING_EXPRESSION", "sqlState" : "42K0E" } @@ -613,7 +613,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "UNSUPPORTED_GROUPING_EXPRESSION", + "condition" : "UNSUPPORTED_GROUPING_EXPRESSION", "sqlState" : "42K0E" } diff --git a/sql/core/src/test/resources/sql-tests/results/group-by-all-duckdb.sql.out b/sql/core/src/test/resources/sql-tests/results/group-by-all-duckdb.sql.out index 30b5846ceff8..baceb7adab5f 100644 --- a/sql/core/src/test/resources/sql-tests/results/group-by-all-duckdb.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/group-by-all-duckdb.sql.out @@ -57,7 +57,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_ALL_IN_GROUP_BY", + "condition" : "UNRESOLVED_ALL_IN_GROUP_BY", "sqlState" : "42803", "queryContext" : [ { "objectType" : "", @@ -103,7 +103,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`c1`", diff --git a/sql/core/src/test/resources/sql-tests/results/group-by-all-mosha.sql.out b/sql/core/src/test/resources/sql-tests/results/group-by-all-mosha.sql.out index f0708c56224f..ed2513be7d60 100644 --- a/sql/core/src/test/resources/sql-tests/results/group-by-all-mosha.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/group-by-all-mosha.sql.out @@ -58,7 +58,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_ALL_IN_GROUP_BY", + "condition" : "UNRESOLVED_ALL_IN_GROUP_BY", "sqlState" : "42803", "queryContext" : [ { "objectType" : "", @@ -103,7 +103,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`i`", @@ -126,7 +126,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`i`", diff --git a/sql/core/src/test/resources/sql-tests/results/group-by-all.sql.out b/sql/core/src/test/resources/sql-tests/results/group-by-all.sql.out index bf3c1a3c21b8..299e2adf4021 100644 --- a/sql/core/src/test/resources/sql-tests/results/group-by-all.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/group-by-all.sql.out @@ -195,7 +195,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_ALL_IN_GROUP_BY", + "condition" : "UNRESOLVED_ALL_IN_GROUP_BY", "sqlState" : "42803", "queryContext" : [ { "objectType" : "", @@ -214,7 +214,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_ALL_IN_GROUP_BY", + "condition" : "UNRESOLVED_ALL_IN_GROUP_BY", "sqlState" : "42803", "queryContext" : [ { "objectType" : "", @@ -255,7 +255,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_ALL_IN_GROUP_BY", + "condition" : "UNRESOLVED_ALL_IN_GROUP_BY", "sqlState" : "42803", "queryContext" : [ { "objectType" : "", @@ -274,7 +274,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`non_exist`", diff --git a/sql/core/src/test/resources/sql-tests/results/group-by-filter.sql.out b/sql/core/src/test/resources/sql-tests/results/group-by-filter.sql.out index 57d4681e2dd9..a3dc66442da3 100644 --- a/sql/core/src/test/resources/sql-tests/results/group-by-filter.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/group-by-filter.sql.out @@ -61,7 +61,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "MISSING_GROUP_BY", + "condition" : "MISSING_GROUP_BY", "sqlState" : "42803", "queryContext" : [ { "objectType" : "", @@ -251,7 +251,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "MISSING_AGGREGATION", + "condition" : "MISSING_AGGREGATION", "sqlState" : "42803", "messageParameters" : { "expression" : "\"a\"", @@ -738,7 +738,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "MISSING_AGGREGATION", + "condition" : "MISSING_AGGREGATION", "sqlState" : "42803", "messageParameters" : { "expression" : "\"a\"", @@ -901,7 +901,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_AGGREGATE_FILTER.NON_DETERMINISTIC", + "condition" : "INVALID_AGGREGATE_FILTER.NON_DETERMINISTIC", "sqlState" : "42903", "messageParameters" : { "filterExpr" : "\"(rand(num2) > 1)\"" @@ -923,7 +923,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_AGGREGATE_FILTER.NOT_BOOLEAN", + "condition" : "INVALID_AGGREGATE_FILTER.NOT_BOOLEAN", "sqlState" : "42903", "messageParameters" : { "filterExpr" : "\"str\"" @@ -945,7 +945,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_AGGREGATE_FILTER.CONTAINS_AGGREGATE", + "condition" : "INVALID_AGGREGATE_FILTER.CONTAINS_AGGREGATE", "sqlState" : "42903", "messageParameters" : { "aggExpr" : "\"max(num2)\"", @@ -968,7 +968,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_AGGREGATE_FILTER.CONTAINS_WINDOW_FUNCTION", + "condition" : "INVALID_AGGREGATE_FILTER.CONTAINS_WINDOW_FUNCTION", "sqlState" : "42903", "messageParameters" : { "filterExpr" : "\"(nth_value(num2, 2) OVER (ORDER BY num2 ASC NULLS FIRST RANGE BETWEEN UNBOUNDED PRECEDING AND CURRENT ROW) > 1)\"", diff --git a/sql/core/src/test/resources/sql-tests/results/group-by-ordinal.sql.out b/sql/core/src/test/resources/sql-tests/results/group-by-ordinal.sql.out index 0f29c27268c8..55f3eeddf42f 100644 --- a/sql/core/src/test/resources/sql-tests/results/group-by-ordinal.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/group-by-ordinal.sql.out @@ -93,7 +93,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "GROUP_BY_POS_OUT_OF_RANGE", + "condition" : "GROUP_BY_POS_OUT_OF_RANGE", "sqlState" : "42805", "messageParameters" : { "index" : "-1", @@ -116,7 +116,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "GROUP_BY_POS_OUT_OF_RANGE", + "condition" : "GROUP_BY_POS_OUT_OF_RANGE", "sqlState" : "42805", "messageParameters" : { "index" : "0", @@ -139,7 +139,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "GROUP_BY_POS_OUT_OF_RANGE", + "condition" : "GROUP_BY_POS_OUT_OF_RANGE", "sqlState" : "42805", "messageParameters" : { "index" : "3", @@ -162,7 +162,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "GROUP_BY_POS_AGGREGATE", + "condition" : "GROUP_BY_POS_AGGREGATE", "sqlState" : "42903", "messageParameters" : { "aggExpr" : "sum(data.b) AS `sum(b)`", @@ -185,7 +185,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "GROUP_BY_POS_AGGREGATE", + "condition" : "GROUP_BY_POS_AGGREGATE", "sqlState" : "42903", "messageParameters" : { "aggExpr" : "(sum(data.b) + CAST(2 AS BIGINT)) AS `(sum(b) + 2)`", @@ -223,7 +223,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "STAR_GROUP_BY_POS", + "condition" : "STAR_GROUP_BY_POS", "sqlState" : "0A000" } @@ -423,7 +423,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "GROUP_BY_POS_OUT_OF_RANGE", + "condition" : "GROUP_BY_POS_OUT_OF_RANGE", "sqlState" : "42805", "messageParameters" : { "index" : "-1", @@ -446,7 +446,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "GROUP_BY_POS_AGGREGATE", + "condition" : "GROUP_BY_POS_AGGREGATE", "sqlState" : "42903", "messageParameters" : { "aggExpr" : "count(1) AS `count(1)`", @@ -469,7 +469,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "GROUP_BY_POS_OUT_OF_RANGE", + "condition" : "GROUP_BY_POS_OUT_OF_RANGE", "sqlState" : "42805", "messageParameters" : { "index" : "-1", @@ -492,7 +492,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "GROUP_BY_POS_AGGREGATE", + "condition" : "GROUP_BY_POS_AGGREGATE", "sqlState" : "42903", "messageParameters" : { "aggExpr" : "count(1) AS `count(1)`", diff --git a/sql/core/src/test/resources/sql-tests/results/group-by.sql.out b/sql/core/src/test/resources/sql-tests/results/group-by.sql.out index 98ad1a0a5bba..0f61924aa425 100644 --- a/sql/core/src/test/resources/sql-tests/results/group-by.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/group-by.sql.out @@ -16,7 +16,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "MISSING_GROUP_BY", + "condition" : "MISSING_GROUP_BY", "sqlState" : "42803", "queryContext" : [ { "objectType" : "", @@ -54,7 +54,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "MISSING_AGGREGATION", + "condition" : "MISSING_AGGREGATION", "sqlState" : "42803", "messageParameters" : { "expression" : "\"a\"", @@ -125,7 +125,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "MISSING_AGGREGATION", + "condition" : "MISSING_AGGREGATION", "sqlState" : "42803", "messageParameters" : { "expression" : "\"a\"", @@ -200,7 +200,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`non_existing`", @@ -223,7 +223,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "GROUP_BY_AGGREGATE", + "condition" : "GROUP_BY_AGGREGATE", "sqlState" : "42903", "messageParameters" : { "sqlExpr" : "count(testdata.b)" @@ -254,7 +254,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "MISSING_AGGREGATION", + "condition" : "MISSING_AGGREGATION", "sqlState" : "42803", "messageParameters" : { "expression" : "\"k\"", @@ -278,7 +278,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`k`", @@ -364,7 +364,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "MISSING_GROUP_BY", + "condition" : "MISSING_GROUP_BY", "sqlState" : "42803", "queryContext" : [ { "objectType" : "", @@ -408,7 +408,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INVALID_WHERE_CONDITION", + "condition" : "INVALID_WHERE_CONDITION", "sqlState" : "42903", "messageParameters" : { "condition" : "\"(max(id) > 0)\"", @@ -552,7 +552,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"1\"", @@ -578,7 +578,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"1\"", @@ -604,7 +604,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"1\"", @@ -638,7 +638,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"1.0\"", @@ -664,7 +664,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"1.0\"", @@ -801,7 +801,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INVALID_WHERE_CONDITION", + "condition" : "INVALID_WHERE_CONDITION", "sqlState" : "42903", "messageParameters" : { "condition" : "\"(count(1) > 1)\"", @@ -824,7 +824,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INVALID_WHERE_CONDITION", + "condition" : "INVALID_WHERE_CONDITION", "sqlState" : "42903", "messageParameters" : { "condition" : "\"((count(1) + 1) > 1)\"", @@ -847,7 +847,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INVALID_WHERE_CONDITION", + "condition" : "INVALID_WHERE_CONDITION", "sqlState" : "42903", "messageParameters" : { "condition" : "\"(((k = 1) OR (k = 2)) OR (((count(1) + 1) > 1) OR (max(k) > 1)))\"", diff --git a/sql/core/src/test/resources/sql-tests/results/grouping_set.sql.out b/sql/core/src/test/resources/sql-tests/results/grouping_set.sql.out index 5849acd094fb..1579aec1e1b2 100644 --- a/sql/core/src/test/resources/sql-tests/results/grouping_set.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/grouping_set.sql.out @@ -135,7 +135,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'ROLLUP'", @@ -151,7 +151,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'CUBE'", @@ -167,7 +167,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "MISSING_AGGREGATION", + "condition" : "MISSING_AGGREGATION", "sqlState" : "42803", "messageParameters" : { "expression" : "\"c1\"", diff --git a/sql/core/src/test/resources/sql-tests/results/having.sql.out b/sql/core/src/test/resources/sql-tests/results/having.sql.out index c9d588642636..53bbecb58e4b 100644 --- a/sql/core/src/test/resources/sql-tests/results/having.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/having.sql.out @@ -36,7 +36,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"INT\"", diff --git a/sql/core/src/test/resources/sql-tests/results/higher-order-functions.sql.out b/sql/core/src/test/resources/sql-tests/results/higher-order-functions.sql.out index 7bfc35a61e09..1ac44e7447de 100644 --- a/sql/core/src/test/resources/sql-tests/results/higher-order-functions.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/higher-order-functions.sql.out @@ -18,7 +18,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_LAMBDA_FUNCTION_CALL.NON_HIGHER_ORDER_FUNCTION", + "condition" : "INVALID_LAMBDA_FUNCTION_CALL.NON_HIGHER_ORDER_FUNCTION", "sqlState" : "42K0D", "messageParameters" : { "class" : "org.apache.spark.sql.catalyst.expressions.Upper" @@ -40,7 +40,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_LAMBDA_FUNCTION_CALL.NON_HIGHER_ORDER_FUNCTION", + "condition" : "INVALID_LAMBDA_FUNCTION_CALL.NON_HIGHER_ORDER_FUNCTION", "sqlState" : "42K0D", "messageParameters" : { "class" : "org.apache.spark.sql.catalyst.expressions.CeilExpressionBuilder$" diff --git a/sql/core/src/test/resources/sql-tests/results/hll.sql.out b/sql/core/src/test/resources/sql-tests/results/hll.sql.out index ea9dc8e8af17..ffcd8eb79be9 100644 --- a/sql/core/src/test/resources/sql-tests/results/hll.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/hll.sql.out @@ -79,7 +79,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"col\"", @@ -106,7 +106,7 @@ struct<> -- !query output org.apache.spark.SparkRuntimeException { - "errorClass" : "HLL_INVALID_LG_K", + "condition" : "HLL_INVALID_LG_K", "sqlState" : "22546", "messageParameters" : { "function" : "`hll_sketch_agg`", @@ -125,7 +125,7 @@ struct<> -- !query output org.apache.spark.SparkRuntimeException { - "errorClass" : "HLL_INVALID_LG_K", + "condition" : "HLL_INVALID_LG_K", "sqlState" : "22546", "messageParameters" : { "function" : "`hll_sketch_agg`", @@ -151,7 +151,7 @@ struct<> -- !query output org.apache.spark.SparkRuntimeException { - "errorClass" : "HLL_UNION_DIFFERENT_LG_K", + "condition" : "HLL_UNION_DIFFERENT_LG_K", "sqlState" : "22000", "messageParameters" : { "function" : "`hll_union`", @@ -173,7 +173,7 @@ struct<> -- !query output org.apache.spark.SparkRuntimeException { - "errorClass" : "HLL_UNION_DIFFERENT_LG_K", + "condition" : "HLL_UNION_DIFFERENT_LG_K", "sqlState" : "22000", "messageParameters" : { "function" : "`hll_union_agg`", @@ -196,7 +196,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"1\"", @@ -222,7 +222,7 @@ struct<> -- !query output org.apache.spark.SparkRuntimeException { - "errorClass" : "HLL_INVALID_INPUT_SKETCH_BUFFER", + "condition" : "HLL_INVALID_INPUT_SKETCH_BUFFER", "sqlState" : "22546", "messageParameters" : { "function" : "`hll_sketch_estimate`" @@ -237,7 +237,7 @@ struct<> -- !query output org.apache.spark.SparkRuntimeException { - "errorClass" : "HLL_INVALID_INPUT_SKETCH_BUFFER", + "condition" : "HLL_INVALID_INPUT_SKETCH_BUFFER", "sqlState" : "22546", "messageParameters" : { "function" : "`hll_union`" @@ -253,7 +253,7 @@ struct<> -- !query output org.apache.spark.SparkRuntimeException { - "errorClass" : "HLL_INVALID_INPUT_SKETCH_BUFFER", + "condition" : "HLL_INVALID_INPUT_SKETCH_BUFFER", "sqlState" : "22546", "messageParameters" : { "function" : "`hll_union_agg`" diff --git a/sql/core/src/test/resources/sql-tests/results/identifier-clause.sql.out b/sql/core/src/test/resources/sql-tests/results/identifier-clause.sql.out index 596745b4ba5d..c4f9e9656f51 100644 --- a/sql/core/src/test/resources/sql-tests/results/identifier-clause.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/identifier-clause.sql.out @@ -102,7 +102,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "UNSUPPORTED_FEATURE.TABLE_OPERATION", + "condition" : "UNSUPPORTED_FEATURE.TABLE_OPERATION", "sqlState" : "0A000", "messageParameters" : { "operation" : "DELETE", @@ -118,7 +118,7 @@ struct<> -- !query output org.apache.spark.SparkUnsupportedOperationException { - "errorClass" : "_LEGACY_ERROR_TEMP_2096", + "condition" : "_LEGACY_ERROR_TEMP_2096", "messageParameters" : { "ddl" : "UPDATE TABLE" } @@ -133,7 +133,7 @@ struct<> -- !query output org.apache.spark.SparkUnsupportedOperationException { - "errorClass" : "_LEGACY_ERROR_TEMP_2096", + "condition" : "_LEGACY_ERROR_TEMP_2096", "messageParameters" : { "ddl" : "MERGE INTO TABLE" } @@ -283,7 +283,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "UNSUPPORTED_FEATURE.TABLE_OPERATION", + "condition" : "UNSUPPORTED_FEATURE.TABLE_OPERATION", "sqlState" : "0A000", "messageParameters" : { "operation" : "REPLACE TABLE", @@ -420,7 +420,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "NOT_A_PARTITIONED_TABLE", + "condition" : "NOT_A_PARTITIONED_TABLE", "sqlState" : "42809", "messageParameters" : { "operation" : "MSCK REPAIR TABLE", @@ -714,7 +714,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'1'", @@ -737,7 +737,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_EMPTY_STATEMENT", + "condition" : "PARSE_EMPTY_STATEMENT", "sqlState" : "42617", "queryContext" : [ { "objectType" : "", @@ -756,7 +756,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "NOT_A_CONSTANT_STRING.NULL", + "condition" : "NOT_A_CONSTANT_STRING.NULL", "sqlState" : "42601", "messageParameters" : { "expr" : "CAST(NULL AS STRING)", @@ -779,7 +779,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "NOT_A_CONSTANT_STRING.WRONG_TYPE", + "condition" : "NOT_A_CONSTANT_STRING.WRONG_TYPE", "sqlState" : "42601", "messageParameters" : { "dataType" : "int", @@ -803,7 +803,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "NOT_A_CONSTANT_STRING.NOT_CONSTANT", + "condition" : "NOT_A_CONSTANT_STRING.NOT_CONSTANT", "sqlState" : "42601", "messageParameters" : { "expr" : "substr('HELLO', 1, CAST((rand() + CAST(1 AS DOUBLE)) AS INT))", @@ -826,7 +826,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "UNRESOLVED_ROUTINE", + "condition" : "UNRESOLVED_ROUTINE", "sqlState" : "42883", "messageParameters" : { "routineName" : "`IDENTIFIER`", @@ -849,7 +849,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "NOT_A_CONSTANT_STRING.WRONG_TYPE", + "condition" : "NOT_A_CONSTANT_STRING.WRONG_TYPE", "sqlState" : "42601", "messageParameters" : { "dataType" : "int", @@ -873,7 +873,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "REQUIRES_SINGLE_PART_NAMESPACE", + "condition" : "REQUIRES_SINGLE_PART_NAMESPACE", "sqlState" : "42K05", "messageParameters" : { "namespace" : "`a`.`b`", @@ -889,7 +889,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "REQUIRES_SINGLE_PART_NAMESPACE", + "condition" : "REQUIRES_SINGLE_PART_NAMESPACE", "sqlState" : "42K05", "messageParameters" : { "namespace" : "`a`.`b`", @@ -905,7 +905,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "REQUIRES_SINGLE_PART_NAMESPACE", + "condition" : "REQUIRES_SINGLE_PART_NAMESPACE", "sqlState" : "42K05", "messageParameters" : { "namespace" : "`a`.`b`", @@ -921,7 +921,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "REQUIRES_SINGLE_PART_NAMESPACE", + "condition" : "REQUIRES_SINGLE_PART_NAMESPACE", "sqlState" : "42K05", "messageParameters" : { "namespace" : "`a`.`b`", @@ -937,7 +937,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "REQUIRES_SINGLE_PART_NAMESPACE", + "condition" : "REQUIRES_SINGLE_PART_NAMESPACE", "sqlState" : "42K05", "messageParameters" : { "namespace" : "`a`.`b`.`c`", @@ -953,7 +953,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "NOT_A_CONSTANT_STRING.WRONG_TYPE", + "condition" : "NOT_A_CONSTANT_STRING.WRONG_TYPE", "sqlState" : "42601", "messageParameters" : { "dataType" : "int", @@ -977,7 +977,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "IDENTIFIER_TOO_MANY_NAME_PARTS", + "condition" : "IDENTIFIER_TOO_MANY_NAME_PARTS", "sqlState" : "42601", "messageParameters" : { "identifier" : "`a`.`b`.`c`.`d`" @@ -999,7 +999,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_SQL_SYNTAX.CREATE_TEMP_FUNC_WITH_DATABASE", + "condition" : "INVALID_SQL_SYNTAX.CREATE_TEMP_FUNC_WITH_DATABASE", "sqlState" : "42000", "messageParameters" : { "database" : "`default`" @@ -1021,7 +1021,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_SQL_SYNTAX.MULTI_PART_NAME", + "condition" : "INVALID_SQL_SYNTAX.MULTI_PART_NAME", "sqlState" : "42000", "messageParameters" : { "name" : "`default`.`myDoubleAvg`", @@ -1044,7 +1044,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "TEMP_VIEW_NAME_TOO_MANY_NAME_PARTS", + "condition" : "TEMP_VIEW_NAME_TOO_MANY_NAME_PARTS", "sqlState" : "428EK", "messageParameters" : { "actualName" : "`default`.`v`" @@ -1175,7 +1175,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "''x.win''", @@ -1191,7 +1191,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'('", @@ -1207,7 +1207,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`t`", @@ -1230,7 +1230,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "''a''", @@ -1246,7 +1246,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "''a''", @@ -1262,7 +1262,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_SQL_SYNTAX.INVALID_TABLE_VALUED_FUNC_NAME", + "condition" : "INVALID_SQL_SYNTAX.INVALID_TABLE_VALUED_FUNC_NAME", "sqlState" : "42000", "messageParameters" : { "funcName" : "`s`.`IDENTIFIER`" @@ -1284,7 +1284,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'.'", @@ -1300,7 +1300,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'.'", @@ -1316,7 +1316,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "''win''", @@ -1332,7 +1332,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'WINDOW'", @@ -1348,7 +1348,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "''v''", @@ -1364,7 +1364,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'('", @@ -1380,7 +1380,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'('", @@ -1396,7 +1396,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'('", diff --git a/sql/core/src/test/resources/sql-tests/results/ilike-all.sql.out b/sql/core/src/test/resources/sql-tests/results/ilike-all.sql.out index 96dfe0f97c3c..3953b32d8532 100644 --- a/sql/core/src/test/resources/sql-tests/results/ilike-all.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/ilike-all.sql.out @@ -130,7 +130,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0064", + "condition" : "_LEGACY_ERROR_TEMP_0064", "messageParameters" : { "msg" : "Expected something between '(' and ')'." }, diff --git a/sql/core/src/test/resources/sql-tests/results/ilike-any.sql.out b/sql/core/src/test/resources/sql-tests/results/ilike-any.sql.out index c8bd38d32462..ff27a23c72d2 100644 --- a/sql/core/src/test/resources/sql-tests/results/ilike-any.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/ilike-any.sql.out @@ -136,7 +136,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0064", + "condition" : "_LEGACY_ERROR_TEMP_0064", "messageParameters" : { "msg" : "Expected something between '(' and ')'." }, diff --git a/sql/core/src/test/resources/sql-tests/results/inline-table.sql.out b/sql/core/src/test/resources/sql-tests/results/inline-table.sql.out index 0a2c7b0f55ed..80a0e3d5c9d3 100644 --- a/sql/core/src/test/resources/sql-tests/results/inline-table.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/inline-table.sql.out @@ -111,7 +111,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_INLINE_TABLE.CANNOT_EVALUATE_EXPRESSION_IN_INLINE_TABLE", + "condition" : "INVALID_INLINE_TABLE.CANNOT_EVALUATE_EXPRESSION_IN_INLINE_TABLE", "sqlState" : "42000", "messageParameters" : { "expr" : "\"rand(5)\"" @@ -133,7 +133,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_INLINE_TABLE.NUM_COLUMNS_MISMATCH", + "condition" : "INVALID_INLINE_TABLE.NUM_COLUMNS_MISMATCH", "sqlState" : "42000", "messageParameters" : { "actualNumCols" : "1", @@ -157,7 +157,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_INLINE_TABLE.INCOMPATIBLE_TYPES_IN_INLINE_TABLE", + "condition" : "INVALID_INLINE_TABLE.INCOMPATIBLE_TYPES_IN_INLINE_TABLE", "sqlState" : "42000", "messageParameters" : { "colName" : "`b`" @@ -179,7 +179,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_INLINE_TABLE.NUM_COLUMNS_MISMATCH", + "condition" : "INVALID_INLINE_TABLE.NUM_COLUMNS_MISMATCH", "sqlState" : "42000", "messageParameters" : { "actualNumCols" : "1", @@ -203,7 +203,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "UNRESOLVED_ROUTINE", + "condition" : "UNRESOLVED_ROUTINE", "sqlState" : "42883", "messageParameters" : { "routineName" : "`random_not_exist_func`", @@ -226,7 +226,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_INLINE_TABLE.CANNOT_EVALUATE_EXPRESSION_IN_INLINE_TABLE", + "condition" : "INVALID_INLINE_TABLE.CANNOT_EVALUATE_EXPRESSION_IN_INLINE_TABLE", "sqlState" : "42000", "messageParameters" : { "expr" : "\"count(1)\"" diff --git a/sql/core/src/test/resources/sql-tests/results/intersect-all.sql.out b/sql/core/src/test/resources/sql-tests/results/intersect-all.sql.out index 44f95f225abe..4a75429bddf5 100644 --- a/sql/core/src/test/resources/sql-tests/results/intersect-all.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/intersect-all.sql.out @@ -96,7 +96,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -125,7 +125,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "NUM_COLUMNS_MISMATCH", + "condition" : "NUM_COLUMNS_MISMATCH", "sqlState" : "42826", "messageParameters" : { "firstNumColumns" : "1", diff --git a/sql/core/src/test/resources/sql-tests/results/interval.sql.out b/sql/core/src/test/resources/sql-tests/results/interval.sql.out index 4e012df792de..cef6e1114e6a 100644 --- a/sql/core/src/test/resources/sql-tests/results/interval.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/interval.sql.out @@ -14,7 +14,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0029", + "condition" : "_LEGACY_ERROR_TEMP_0029", "messageParameters" : { "literal" : "interval 4 month 2 weeks 3 microseconds" }, @@ -127,7 +127,7 @@ struct<> -- !query output org.apache.spark.SparkNumberFormatException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'a'", @@ -151,7 +151,7 @@ struct<> -- !query output org.apache.spark.SparkNumberFormatException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'a'", @@ -175,7 +175,7 @@ struct<> -- !query output org.apache.spark.SparkNumberFormatException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'a'", @@ -199,7 +199,7 @@ struct<> -- !query output org.apache.spark.SparkNumberFormatException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'a'", @@ -239,7 +239,7 @@ struct<> -- !query output org.apache.spark.SparkNumberFormatException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'a'", @@ -263,7 +263,7 @@ struct<> -- !query output org.apache.spark.SparkNumberFormatException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'a'", @@ -287,7 +287,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"STRING\"", @@ -311,7 +311,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"STRING\"", @@ -335,7 +335,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "INTERVAL_DIVIDED_BY_ZERO", + "condition" : "INTERVAL_DIVIDED_BY_ZERO", "sqlState" : "22012", "queryContext" : [ { "objectType" : "", @@ -378,7 +378,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "INTERVAL_DIVIDED_BY_ZERO", + "condition" : "INTERVAL_DIVIDED_BY_ZERO", "sqlState" : "22012", "queryContext" : [ { "objectType" : "", @@ -421,7 +421,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"INT\"", @@ -445,7 +445,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"INT\"", @@ -469,7 +469,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"VOID\"", @@ -493,7 +493,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"VOID\"", @@ -517,7 +517,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0029", + "condition" : "_LEGACY_ERROR_TEMP_0029", "messageParameters" : { "literal" : "interval '-1 month 1 day -1 second'" }, @@ -554,7 +554,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0029", + "condition" : "_LEGACY_ERROR_TEMP_0029", "messageParameters" : { "literal" : "interval -1 month 1 day -1 second" }, @@ -591,7 +591,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0029", + "condition" : "_LEGACY_ERROR_TEMP_0029", "messageParameters" : { "literal" : "interval '-1 month 1 day -1 second'" }, @@ -628,7 +628,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0029", + "condition" : "_LEGACY_ERROR_TEMP_0029", "messageParameters" : { "literal" : "interval -1 month 1 day -1 second" }, @@ -873,7 +873,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "NUMERIC_VALUE_OUT_OF_RANGE.WITH_SUGGESTION", + "condition" : "NUMERIC_VALUE_OUT_OF_RANGE.WITH_SUGGESTION", "sqlState" : "22003", "messageParameters" : { "config" : "\"spark.sql.ansi.enabled\"", @@ -938,7 +938,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "INTERVAL_ARITHMETIC_OVERFLOW.WITHOUT_SUGGESTION", + "condition" : "INTERVAL_ARITHMETIC_OVERFLOW.WITHOUT_SUGGESTION", "sqlState" : "22015", "queryContext" : [ { "objectType" : "", @@ -989,7 +989,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "INTERVAL_ARITHMETIC_OVERFLOW.WITHOUT_SUGGESTION", + "condition" : "INTERVAL_ARITHMETIC_OVERFLOW.WITHOUT_SUGGESTION", "sqlState" : "22015", "queryContext" : [ { "objectType" : "", @@ -1016,7 +1016,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "INTERVAL_ARITHMETIC_OVERFLOW.WITHOUT_SUGGESTION", + "condition" : "INTERVAL_ARITHMETIC_OVERFLOW.WITHOUT_SUGGESTION", "sqlState" : "22015", "queryContext" : [ { "objectType" : "", @@ -1099,7 +1099,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0029", + "condition" : "_LEGACY_ERROR_TEMP_0029", "messageParameters" : { "literal" : "interval 1 year 2 month 3 week 4 day 5 hour 6 minute 7 seconds 8 millisecond 9 microsecond" }, @@ -1144,7 +1144,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0029", + "condition" : "_LEGACY_ERROR_TEMP_0029", "messageParameters" : { "literal" : "interval '30' year '25' month '-100' day '40' hour '80' minute '299.889987299' second" }, @@ -1285,7 +1285,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_INTERVAL_FORMAT.UNMATCHED_FORMAT_STRING_WITH_NOTICE", + "condition" : "INVALID_INTERVAL_FORMAT.UNMATCHED_FORMAT_STRING_WITH_NOTICE", "sqlState" : "22006", "messageParameters" : { "input" : "20 15:40:32.99899999", @@ -1310,7 +1310,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_INTERVAL_FORMAT.UNMATCHED_FORMAT_STRING_WITH_NOTICE", + "condition" : "INVALID_INTERVAL_FORMAT.UNMATCHED_FORMAT_STRING_WITH_NOTICE", "sqlState" : "22006", "messageParameters" : { "input" : "20 15:40:32.99899999", @@ -1335,7 +1335,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_INTERVAL_FORMAT.UNMATCHED_FORMAT_STRING_WITH_NOTICE", + "condition" : "INVALID_INTERVAL_FORMAT.UNMATCHED_FORMAT_STRING_WITH_NOTICE", "sqlState" : "22006", "messageParameters" : { "input" : "15:40:32.99899999", @@ -1360,7 +1360,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_INTERVAL_FORMAT.UNMATCHED_FORMAT_STRING_WITH_NOTICE", + "condition" : "INVALID_INTERVAL_FORMAT.UNMATCHED_FORMAT_STRING_WITH_NOTICE", "sqlState" : "22006", "messageParameters" : { "input" : "15:40.99899999", @@ -1385,7 +1385,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_INTERVAL_FORMAT.UNMATCHED_FORMAT_STRING_WITH_NOTICE", + "condition" : "INVALID_INTERVAL_FORMAT.UNMATCHED_FORMAT_STRING_WITH_NOTICE", "sqlState" : "22006", "messageParameters" : { "input" : "15:40", @@ -1410,7 +1410,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_INTERVAL_FORMAT.UNMATCHED_FORMAT_STRING_WITH_NOTICE", + "condition" : "INVALID_INTERVAL_FORMAT.UNMATCHED_FORMAT_STRING_WITH_NOTICE", "sqlState" : "22006", "messageParameters" : { "input" : "20 40:32.99899999", @@ -1435,7 +1435,7 @@ struct<> -- !query output org.apache.spark.SparkIllegalArgumentException { - "errorClass" : "INVALID_INTERVAL_FORMAT.INVALID_UNIT", + "condition" : "INVALID_INTERVAL_FORMAT.INVALID_UNIT", "sqlState" : "22006", "messageParameters" : { "input" : " 10 nanoseconds", @@ -1531,7 +1531,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'fake_unit'", @@ -1547,7 +1547,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0027", + "condition" : "_LEGACY_ERROR_TEMP_0027", "queryContext" : [ { "objectType" : "", "objectName" : "", @@ -1565,7 +1565,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_INTERVAL_FORMAT.UNSUPPORTED_FROM_TO_EXPRESSION", + "condition" : "INVALID_INTERVAL_FORMAT.UNSUPPORTED_FROM_TO_EXPRESSION", "sqlState" : "22006", "messageParameters" : { "from" : "year", @@ -1589,7 +1589,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0024", + "condition" : "_LEGACY_ERROR_TEMP_0024", "queryContext" : [ { "objectType" : "", "objectName" : "", @@ -1607,7 +1607,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0024", + "condition" : "_LEGACY_ERROR_TEMP_0024", "queryContext" : [ { "objectType" : "", "objectName" : "", @@ -1625,7 +1625,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0024", + "condition" : "_LEGACY_ERROR_TEMP_0024", "queryContext" : [ { "objectType" : "", "objectName" : "", @@ -1643,7 +1643,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0024", + "condition" : "_LEGACY_ERROR_TEMP_0024", "queryContext" : [ { "objectType" : "", "objectName" : "", @@ -1661,7 +1661,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0024", + "condition" : "_LEGACY_ERROR_TEMP_0024", "queryContext" : [ { "objectType" : "", "objectName" : "", @@ -1679,7 +1679,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0024", + "condition" : "_LEGACY_ERROR_TEMP_0024", "queryContext" : [ { "objectType" : "", "objectName" : "", @@ -1697,7 +1697,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0024", + "condition" : "_LEGACY_ERROR_TEMP_0024", "queryContext" : [ { "objectType" : "", "objectName" : "", @@ -1715,7 +1715,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "UNRESOLVED_ROUTINE", + "condition" : "UNRESOLVED_ROUTINE", "sqlState" : "42883", "messageParameters" : { "routineName" : "`interval`", @@ -1738,7 +1738,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "UNRESOLVED_ROUTINE", + "condition" : "UNRESOLVED_ROUTINE", "sqlState" : "42883", "messageParameters" : { "routineName" : "`interval`", @@ -1761,7 +1761,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'day'", @@ -1777,7 +1777,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "UNRESOLVED_ROUTINE", + "condition" : "UNRESOLVED_ROUTINE", "sqlState" : "42883", "messageParameters" : { "routineName" : "`interval`", @@ -1800,7 +1800,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "UNRESOLVED_ROUTINE", + "condition" : "UNRESOLVED_ROUTINE", "sqlState" : "42883", "messageParameters" : { "routineName" : "`interval`", @@ -1823,7 +1823,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'days'", @@ -1847,7 +1847,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_INTERVAL_FORMAT.INTERVAL_PARSING", + "condition" : "INVALID_INTERVAL_FORMAT.INTERVAL_PARSING", "sqlState" : "22006", "messageParameters" : { "input" : "178956970-8", @@ -1906,7 +1906,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"INTERVAL YEAR\"", @@ -1946,7 +1946,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"INTERVAL YEAR\"", @@ -1970,7 +1970,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"INTERVAL YEAR\"", @@ -1994,7 +1994,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'4 11:11'", @@ -2018,7 +2018,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'4 12:12:12'", @@ -2050,7 +2050,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"INTERVAL YEAR\"", @@ -2074,7 +2074,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"INTERVAL YEAR\"", @@ -2098,7 +2098,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'1'", @@ -2122,7 +2122,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'1'", @@ -2146,7 +2146,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"INTERVAL '2-2' YEAR TO MONTH\"", @@ -2172,7 +2172,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"INTERVAL '2-2' YEAR TO MONTH\"", @@ -2198,7 +2198,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"INTERVAL '2-2' YEAR TO MONTH\"", @@ -2224,7 +2224,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"INTERVAL DAY\"", @@ -2248,7 +2248,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"1\"", @@ -2274,7 +2274,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"INT\"", @@ -2298,7 +2298,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"1\"", @@ -2324,7 +2324,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"INTERVAL MONTH\"", @@ -2396,7 +2396,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_INTERVAL_FORMAT.UNMATCHED_FORMAT_STRING", + "condition" : "INVALID_INTERVAL_FORMAT.UNMATCHED_FORMAT_STRING", "sqlState" : "22006", "messageParameters" : { "input" : "-\t2-2\t", @@ -2429,7 +2429,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_INTERVAL_FORMAT.UNMATCHED_FORMAT_STRING_WITH_NOTICE", + "condition" : "INVALID_INTERVAL_FORMAT.UNMATCHED_FORMAT_STRING_WITH_NOTICE", "sqlState" : "22006", "messageParameters" : { "input" : "\n-\t10\t 12:34:46.789\t", @@ -2454,7 +2454,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_TYPED_LITERAL", + "condition" : "INVALID_TYPED_LITERAL", "sqlState" : "42604", "messageParameters" : { "value" : "'中文 interval 1 day'", @@ -2477,7 +2477,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_TYPED_LITERAL", + "condition" : "INVALID_TYPED_LITERAL", "sqlState" : "42604", "messageParameters" : { "value" : "'interval中文 1 day'", @@ -2500,7 +2500,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_TYPED_LITERAL", + "condition" : "INVALID_TYPED_LITERAL", "sqlState" : "42604", "messageParameters" : { "value" : "'interval 1中文day'", @@ -2523,7 +2523,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "INTERVAL_ARITHMETIC_OVERFLOW.WITHOUT_SUGGESTION", + "condition" : "INTERVAL_ARITHMETIC_OVERFLOW.WITHOUT_SUGGESTION", "sqlState" : "22015" } @@ -2535,7 +2535,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "INTERVAL_ARITHMETIC_OVERFLOW.WITH_SUGGESTION", + "condition" : "INTERVAL_ARITHMETIC_OVERFLOW.WITH_SUGGESTION", "sqlState" : "22015", "messageParameters" : { "functionName" : "`try_subtract`" @@ -2550,7 +2550,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "INTERVAL_ARITHMETIC_OVERFLOW.WITH_SUGGESTION", + "condition" : "INTERVAL_ARITHMETIC_OVERFLOW.WITH_SUGGESTION", "sqlState" : "22015", "messageParameters" : { "functionName" : "`try_add`" @@ -2635,7 +2635,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_TYPED_LITERAL", + "condition" : "INVALID_TYPED_LITERAL", "sqlState" : "42604", "messageParameters" : { "value" : "'+'", @@ -2658,7 +2658,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_TYPED_LITERAL", + "condition" : "INVALID_TYPED_LITERAL", "sqlState" : "42604", "messageParameters" : { "value" : "'+.'", @@ -2681,7 +2681,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_TYPED_LITERAL", + "condition" : "INVALID_TYPED_LITERAL", "sqlState" : "42604", "messageParameters" : { "value" : "'1'", @@ -2704,7 +2704,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_TYPED_LITERAL", + "condition" : "INVALID_TYPED_LITERAL", "sqlState" : "42604", "messageParameters" : { "value" : "'1.2'", @@ -2727,7 +2727,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_TYPED_LITERAL", + "condition" : "INVALID_TYPED_LITERAL", "sqlState" : "42604", "messageParameters" : { "value" : "'- 2'", @@ -2750,7 +2750,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_TYPED_LITERAL", + "condition" : "INVALID_TYPED_LITERAL", "sqlState" : "42604", "messageParameters" : { "value" : "'1 day -'", @@ -2773,7 +2773,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_TYPED_LITERAL", + "condition" : "INVALID_TYPED_LITERAL", "sqlState" : "42604", "messageParameters" : { "value" : "'1 day 1'", @@ -2796,7 +2796,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0026", + "condition" : "_LEGACY_ERROR_TEMP_0026", "messageParameters" : { "value" : "1 day 2" }, @@ -2817,7 +2817,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0026", + "condition" : "_LEGACY_ERROR_TEMP_0026", "messageParameters" : { "value" : "interval 1" }, @@ -2862,7 +2862,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "INTERVAL_ARITHMETIC_OVERFLOW.WITH_SUGGESTION", + "condition" : "INTERVAL_ARITHMETIC_OVERFLOW.WITH_SUGGESTION", "sqlState" : "22015", "messageParameters" : { "functionName" : "`try_divide`" @@ -2884,7 +2884,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "INTERVAL_ARITHMETIC_OVERFLOW.WITH_SUGGESTION", + "condition" : "INTERVAL_ARITHMETIC_OVERFLOW.WITH_SUGGESTION", "sqlState" : "22015", "messageParameters" : { "functionName" : "`try_divide`" @@ -2940,7 +2940,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "INTERVAL_ARITHMETIC_OVERFLOW.WITH_SUGGESTION", + "condition" : "INTERVAL_ARITHMETIC_OVERFLOW.WITH_SUGGESTION", "sqlState" : "22015", "messageParameters" : { "functionName" : "`try_divide`" @@ -2962,7 +2962,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "INTERVAL_ARITHMETIC_OVERFLOW.WITH_SUGGESTION", + "condition" : "INTERVAL_ARITHMETIC_OVERFLOW.WITH_SUGGESTION", "sqlState" : "22015", "messageParameters" : { "functionName" : "`try_divide`" @@ -3098,7 +3098,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0063", + "condition" : "_LEGACY_ERROR_TEMP_0063", "messageParameters" : { "msg" : "requirement failed: day 106751992 outside range [0, 106751991]" }, @@ -3119,7 +3119,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0063", + "condition" : "_LEGACY_ERROR_TEMP_0063", "messageParameters" : { "msg" : "requirement failed: day 106751992 outside range [0, 106751991]" }, @@ -3140,7 +3140,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0063", + "condition" : "_LEGACY_ERROR_TEMP_0063", "messageParameters" : { "msg" : "requirement failed: hour 2562047789 outside range [0, 2562047788]" }, @@ -3161,7 +3161,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0063", + "condition" : "_LEGACY_ERROR_TEMP_0063", "messageParameters" : { "msg" : "requirement failed: hour 2562047789 outside range [0, 2562047788]" }, @@ -3182,7 +3182,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0063", + "condition" : "_LEGACY_ERROR_TEMP_0063", "messageParameters" : { "msg" : "requirement failed: minute 153722867281 outside range [0, 153722867280]" }, @@ -3203,7 +3203,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0063", + "condition" : "_LEGACY_ERROR_TEMP_0063", "messageParameters" : { "msg" : "requirement failed: minute 153722867281 outside range [0, 153722867280]" }, @@ -3352,7 +3352,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"INTERVAL MONTH\"", @@ -3376,7 +3376,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"INTERVAL DAY\"", @@ -3400,7 +3400,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"INTERVAL DAY\"", @@ -3424,7 +3424,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"INTERVAL DAY\"", @@ -3448,7 +3448,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"STRING\"", @@ -3472,7 +3472,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"STRING\"", @@ -3496,7 +3496,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"STRING\"", @@ -3520,7 +3520,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"INTERVAL YEAR\"", @@ -3544,7 +3544,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"INTERVAL YEAR\"", @@ -3568,7 +3568,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"INTERVAL YEAR\"", @@ -3592,7 +3592,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"STRING\"", @@ -3616,7 +3616,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"STRING\"", @@ -3640,7 +3640,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"STRING\"", @@ -3680,7 +3680,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "(\"INTERVAL MONTH\" or \"INTERVAL DAY\")", @@ -3720,7 +3720,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "(\"INTERVAL MONTH\" or \"INTERVAL DAY\")", @@ -3792,7 +3792,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"INTERVAL MONTH\"", diff --git a/sql/core/src/test/resources/sql-tests/results/join-lateral.sql.out b/sql/core/src/test/resources/sql-tests/results/join-lateral.sql.out index 11bafb2cf63c..88ad44afac30 100644 --- a/sql/core/src/test/resources/sql-tests/results/join-lateral.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/join-lateral.sql.out @@ -131,7 +131,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "CANNOT_RESOLVE_STAR_EXPAND", + "condition" : "CANNOT_RESOLVE_STAR_EXPAND", "sqlState" : "42704", "messageParameters" : { "columns" : "`c1`, `c2`", @@ -180,7 +180,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INCOMPATIBLE_JOIN_TYPES", + "condition" : "INCOMPATIBLE_JOIN_TYPES", "sqlState" : "42613", "messageParameters" : { "joinType1" : "LATERAL", @@ -203,7 +203,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "UNSUPPORTED_FEATURE.LATERAL_JOIN_USING", + "condition" : "UNSUPPORTED_FEATURE.LATERAL_JOIN_USING", "sqlState" : "0A000", "queryContext" : [ { "objectType" : "", @@ -320,7 +320,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITHOUT_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITHOUT_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`t2`.`c1`" @@ -358,7 +358,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNSUPPORTED_SUBQUERY_EXPRESSION_CATEGORY.NON_DETERMINISTIC_LATERAL_SUBQUERIES", + "condition" : "UNSUPPORTED_SUBQUERY_EXPRESSION_CATEGORY.NON_DETERMINISTIC_LATERAL_SUBQUERIES", "sqlState" : "0A000", "messageParameters" : { "treeNode" : "LateralJoin lateral-subquery#x [c1#x && c2#x], Inner\n: +- SubqueryAlias __auto_generated_subquery_name\n: +- Project [(cast((outer(c1#x) + outer(c2#x)) as double) + rand(number)) AS c3#x]\n: +- OneRowRelation\n+- SubqueryAlias spark_catalog.default.t1\n +- View (`spark_catalog`.`default`.`t1`, [c1#x, c2#x])\n +- Project [cast(col1#x as int) AS c1#x, cast(col2#x as int) AS c2#x]\n +- LocalRelation [col1#x, col2#x]\n" @@ -380,7 +380,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNSUPPORTED_SUBQUERY_EXPRESSION_CATEGORY.NON_DETERMINISTIC_LATERAL_SUBQUERIES", + "condition" : "UNSUPPORTED_SUBQUERY_EXPRESSION_CATEGORY.NON_DETERMINISTIC_LATERAL_SUBQUERIES", "sqlState" : "0A000", "messageParameters" : { "treeNode" : "LateralJoin lateral-subquery#x [], Inner\n: +- SubqueryAlias __auto_generated_subquery_name\n: +- Project [rand(number) AS rand(number)#x]\n: +- SubqueryAlias spark_catalog.default.t2\n: +- View (`spark_catalog`.`default`.`t2`, [c1#x, c2#x])\n: +- Project [cast(col1#x as int) AS c1#x, cast(col2#x as int) AS c2#x]\n: +- LocalRelation [col1#x, col2#x]\n+- SubqueryAlias spark_catalog.default.t1\n +- View (`spark_catalog`.`default`.`t1`, [c1#x, c2#x])\n +- Project [cast(col1#x as int) AS c1#x, cast(col2#x as int) AS c2#x]\n +- LocalRelation [col1#x, col2#x]\n" @@ -402,7 +402,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNSUPPORTED_SUBQUERY_EXPRESSION_CATEGORY.LATERAL_JOIN_CONDITION_NON_DETERMINISTIC", + "condition" : "UNSUPPORTED_SUBQUERY_EXPRESSION_CATEGORY.LATERAL_JOIN_CONDITION_NON_DETERMINISTIC", "sqlState" : "0A000", "messageParameters" : { "condition" : "((CAST(spark_catalog.default.t1.c1 AS DOUBLE) + rand(0)) = CAST(s.c1 AS DOUBLE))" @@ -479,7 +479,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITHOUT_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITHOUT_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`t1`.`c1`" @@ -501,7 +501,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITHOUT_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITHOUT_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`c2`" @@ -541,7 +541,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`t1`.`c1`", @@ -1194,7 +1194,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNSUPPORTED_SUBQUERY_EXPRESSION_CATEGORY.CORRELATED_REFERENCE", + "condition" : "UNSUPPORTED_SUBQUERY_EXPRESSION_CATEGORY.CORRELATED_REFERENCE", "sqlState" : "0A000", "messageParameters" : { "sqlExprs" : "\"(c1 - c1)\",\"(c1 - c1) AS `(c1 - outer(spark_catalog.default.t1.c1))`\"" @@ -1368,7 +1368,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNSUPPORTED_SUBQUERY_EXPRESSION_CATEGORY.ACCESSING_OUTER_QUERY_COLUMN_IS_NOT_ALLOWED", + "condition" : "UNSUPPORTED_SUBQUERY_EXPRESSION_CATEGORY.ACCESSING_OUTER_QUERY_COLUMN_IS_NOT_ALLOWED", "sqlState" : "0A000", "messageParameters" : { "treeNode" : "Filter (c1#x > outer(c2#x))\n+- SubqueryAlias spark_catalog.default.t4\n +- View (`spark_catalog`.`default`.`t4`, [c1#x, c2#x])\n +- Project [cast(col1#x as int) AS c1#x, cast(col2#x as int) AS c2#x]\n +- LocalRelation [col1#x, col2#x]\n" @@ -1779,7 +1779,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.NON_FOLDABLE_INPUT", + "condition" : "DATATYPE_MISMATCH.NON_FOLDABLE_INPUT", "sqlState" : "42K09", "messageParameters" : { "inputExpr" : "\"outer(spark_catalog.default.t1.c1)\"", diff --git a/sql/core/src/test/resources/sql-tests/results/json-functions.sql.out b/sql/core/src/test/resources/sql-tests/results/json-functions.sql.out index 06b8ed88ff88..a38f660b7b7c 100644 --- a/sql/core/src/test/resources/sql-tests/results/json-functions.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/json-functions.sql.out @@ -70,7 +70,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_OPTIONS.NON_MAP_FUNCTION", + "condition" : "INVALID_OPTIONS.NON_MAP_FUNCTION", "sqlState" : "42K06", "queryContext" : [ { "objectType" : "", @@ -89,7 +89,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_OPTIONS.NON_STRING_TYPE", + "condition" : "INVALID_OPTIONS.NON_STRING_TYPE", "sqlState" : "42K06", "messageParameters" : { "mapType" : "\"MAP\"" @@ -111,7 +111,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", + "condition" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", "sqlState" : "42605", "messageParameters" : { "actualNum" : "0", @@ -152,7 +152,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_SCHEMA.NON_STRING_LITERAL", + "condition" : "INVALID_SCHEMA.NON_STRING_LITERAL", "sqlState" : "42K07", "messageParameters" : { "inputSchema" : "\"1\"" @@ -174,7 +174,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'InvalidType'", @@ -197,7 +197,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_OPTIONS.NON_MAP_FUNCTION", + "condition" : "INVALID_OPTIONS.NON_MAP_FUNCTION", "sqlState" : "42K06", "queryContext" : [ { "objectType" : "", @@ -216,7 +216,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_OPTIONS.NON_STRING_TYPE", + "condition" : "INVALID_OPTIONS.NON_STRING_TYPE", "sqlState" : "42K06", "messageParameters" : { "mapType" : "\"MAP\"" @@ -238,7 +238,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", + "condition" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", "sqlState" : "42605", "messageParameters" : { "actualNum" : "0", @@ -480,7 +480,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_NULL", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_NULL", "sqlState" : "42K09", "messageParameters" : { "exprName" : "json", @@ -511,7 +511,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.NON_FOLDABLE_INPUT", + "condition" : "DATATYPE_MISMATCH.NON_FOLDABLE_INPUT", "sqlState" : "42K09", "messageParameters" : { "inputExpr" : "\"jsonField\"", @@ -544,7 +544,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"2\"", @@ -570,7 +570,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", + "condition" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", "sqlState" : "42605", "messageParameters" : { "actualNum" : "0", @@ -659,7 +659,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", + "condition" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", "sqlState" : "42605", "messageParameters" : { "actualNum" : "0", @@ -692,7 +692,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"200\"", diff --git a/sql/core/src/test/resources/sql-tests/results/like-all.sql.out b/sql/core/src/test/resources/sql-tests/results/like-all.sql.out index ce3cc3a0c0f0..299645c77ae4 100644 --- a/sql/core/src/test/resources/sql-tests/results/like-all.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/like-all.sql.out @@ -130,7 +130,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0064", + "condition" : "_LEGACY_ERROR_TEMP_0064", "messageParameters" : { "msg" : "Expected something between '(' and ')'." }, diff --git a/sql/core/src/test/resources/sql-tests/results/like-any.sql.out b/sql/core/src/test/resources/sql-tests/results/like-any.sql.out index 8fcfbb607a35..079f84ca99a4 100644 --- a/sql/core/src/test/resources/sql-tests/results/like-any.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/like-any.sql.out @@ -136,7 +136,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0064", + "condition" : "_LEGACY_ERROR_TEMP_0064", "messageParameters" : { "msg" : "Expected something between '(' and ')'." }, diff --git a/sql/core/src/test/resources/sql-tests/results/limit.sql.out b/sql/core/src/test/resources/sql-tests/results/limit.sql.out index 750e6a6f5cbe..9ec030ac37d5 100644 --- a/sql/core/src/test/resources/sql-tests/results/limit.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/limit.sql.out @@ -51,7 +51,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INVALID_LIMIT_LIKE_EXPRESSION.IS_NEGATIVE", + "condition" : "INVALID_LIMIT_LIKE_EXPRESSION.IS_NEGATIVE", "sqlState" : "42K0E", "messageParameters" : { "expr" : "\"-1\"", @@ -75,7 +75,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INVALID_LIMIT_LIKE_EXPRESSION.IS_NEGATIVE", + "condition" : "INVALID_LIMIT_LIKE_EXPRESSION.IS_NEGATIVE", "sqlState" : "42K0E", "messageParameters" : { "expr" : "\"-1\"", @@ -107,7 +107,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INVALID_LIMIT_LIKE_EXPRESSION.IS_NULL", + "condition" : "INVALID_LIMIT_LIKE_EXPRESSION.IS_NULL", "sqlState" : "42K0E", "messageParameters" : { "expr" : "\"CAST(NULL AS INT)\"", @@ -130,7 +130,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INVALID_LIMIT_LIKE_EXPRESSION.IS_UNFOLDABLE", + "condition" : "INVALID_LIMIT_LIKE_EXPRESSION.IS_UNFOLDABLE", "sqlState" : "42K0E", "messageParameters" : { "expr" : "\"(key > 3)\"", @@ -153,7 +153,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INVALID_LIMIT_LIKE_EXPRESSION.DATA_TYPE", + "condition" : "INVALID_LIMIT_LIKE_EXPRESSION.DATA_TYPE", "sqlState" : "42K0E", "messageParameters" : { "dataType" : "\"BOOLEAN\"", @@ -170,7 +170,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INVALID_LIMIT_LIKE_EXPRESSION.DATA_TYPE", + "condition" : "INVALID_LIMIT_LIKE_EXPRESSION.DATA_TYPE", "sqlState" : "42K0E", "messageParameters" : { "dataType" : "\"STRING\"", diff --git a/sql/core/src/test/resources/sql-tests/results/listagg-collations.sql.out b/sql/core/src/test/resources/sql-tests/results/listagg-collations.sql.out index cf3bac04f09c..a21c0ced7a12 100644 --- a/sql/core/src/test/resources/sql-tests/results/listagg-collations.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/listagg-collations.sql.out @@ -72,7 +72,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INVALID_WITHIN_GROUP_EXPRESSION.MISMATCH_WITH_DISTINCT_INPUT", + "condition" : "INVALID_WITHIN_GROUP_EXPRESSION.MISMATCH_WITH_DISTINCT_INPUT", "sqlState" : "42K0K", "messageParameters" : { "funcArg" : "\"collate(c1, utf8_lcase)\"", diff --git a/sql/core/src/test/resources/sql-tests/results/listagg.sql.out b/sql/core/src/test/resources/sql-tests/results/listagg.sql.out index ef580704992c..4dce4cfc858d 100644 --- a/sql/core/src/test/resources/sql-tests/results/listagg.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/listagg.sql.out @@ -188,7 +188,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"c1\"", @@ -214,7 +214,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "(\"BINARY\" or \"STRING\")", @@ -238,7 +238,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.NON_FOLDABLE_INPUT", + "condition" : "DATATYPE_MISMATCH.NON_FOLDABLE_INPUT", "sqlState" : "42K09", "messageParameters" : { "inputExpr" : "\"a\"", @@ -275,7 +275,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INVALID_WINDOW_SPEC_FOR_AGGREGATION_FUNC", + "condition" : "INVALID_WINDOW_SPEC_FOR_AGGREGATION_FUNC", "sqlState" : "42601", "messageParameters" : { "aggFunc" : "\"listagg(a, NULL, a ASC NULLS FIRST)\"" @@ -297,7 +297,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INVALID_WINDOW_SPEC_FOR_AGGREGATION_FUNC", + "condition" : "INVALID_WINDOW_SPEC_FOR_AGGREGATION_FUNC", "sqlState" : "42601", "messageParameters" : { "aggFunc" : "\"listagg(a, NULL, a ASC NULLS FIRST)\"" @@ -319,7 +319,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DISTINCT_WINDOW_FUNCTION_UNSUPPORTED", + "condition" : "DISTINCT_WINDOW_FUNCTION_UNSUPPORTED", "sqlState" : "0A000", "messageParameters" : { "windowExpr" : "\"listagg(DISTINCT a, NULL) OVER (ORDER BY a ASC NULLS FIRST RANGE BETWEEN UNBOUNDED PRECEDING AND CURRENT ROW)\"" @@ -341,7 +341,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INVALID_WITHIN_GROUP_EXPRESSION.MISMATCH_WITH_DISTINCT_INPUT", + "condition" : "INVALID_WITHIN_GROUP_EXPRESSION.MISMATCH_WITH_DISTINCT_INPUT", "sqlState" : "42K0K", "messageParameters" : { "funcArg" : "\"a\"", @@ -358,7 +358,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INVALID_WITHIN_GROUP_EXPRESSION.MISMATCH_WITH_DISTINCT_INPUT", + "condition" : "INVALID_WITHIN_GROUP_EXPRESSION.MISMATCH_WITH_DISTINCT_INPUT", "sqlState" : "42K0K", "messageParameters" : { "funcArg" : "\"a\"", diff --git a/sql/core/src/test/resources/sql-tests/results/literals.sql.out b/sql/core/src/test/resources/sql-tests/results/literals.sql.out index 4e4c70cc333b..b27e1fee2d61 100644 --- a/sql/core/src/test/resources/sql-tests/results/literals.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/literals.sql.out @@ -38,7 +38,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_NUMERIC_LITERAL_RANGE", + "condition" : "INVALID_NUMERIC_LITERAL_RANGE", "sqlState" : "22003", "messageParameters" : { "maxValue" : "127", @@ -79,7 +79,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_NUMERIC_LITERAL_RANGE", + "condition" : "INVALID_NUMERIC_LITERAL_RANGE", "sqlState" : "22003", "messageParameters" : { "maxValue" : "32767", @@ -120,7 +120,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_NUMERIC_LITERAL_RANGE", + "condition" : "INVALID_NUMERIC_LITERAL_RANGE", "sqlState" : "22003", "messageParameters" : { "maxValue" : "9223372036854775807", @@ -177,7 +177,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "DECIMAL_PRECISION_EXCEEDS_MAX_PRECISION", + "condition" : "DECIMAL_PRECISION_EXCEEDS_MAX_PRECISION", "sqlState" : "22003", "messageParameters" : { "maxPrecision" : "38", @@ -193,7 +193,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "DECIMAL_PRECISION_EXCEEDS_MAX_PRECISION", + "condition" : "DECIMAL_PRECISION_EXCEEDS_MAX_PRECISION", "sqlState" : "22003", "messageParameters" : { "maxPrecision" : "38", @@ -225,7 +225,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_NUMERIC_LITERAL_RANGE", + "condition" : "INVALID_NUMERIC_LITERAL_RANGE", "sqlState" : "22003", "messageParameters" : { "maxValue" : "3.4028234663852886E+38", @@ -266,7 +266,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'.'", @@ -282,7 +282,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_NUMERIC_LITERAL_RANGE", + "condition" : "INVALID_NUMERIC_LITERAL_RANGE", "sqlState" : "22003", "messageParameters" : { "maxValue" : "1.7976931348623157E+308", @@ -397,7 +397,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_TYPED_LITERAL", + "condition" : "INVALID_TYPED_LITERAL", "sqlState" : "42604", "messageParameters" : { "value" : "'mar 11 2016'", @@ -428,7 +428,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_TYPED_LITERAL", + "condition" : "INVALID_TYPED_LITERAL", "sqlState" : "42604", "messageParameters" : { "value" : "'2016-33-11 20:54:00.000'", @@ -451,7 +451,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "UNSUPPORTED_TYPED_LITERAL", + "condition" : "UNSUPPORTED_TYPED_LITERAL", "sqlState" : "0A000", "messageParameters" : { "supportedTypes" : "\"DATE\", \"TIMESTAMP_NTZ\", \"TIMESTAMP_LTZ\", \"TIMESTAMP\", \"INTERVAL\", \"X\"", @@ -482,7 +482,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "DECIMAL_PRECISION_EXCEEDS_MAX_PRECISION", + "condition" : "DECIMAL_PRECISION_EXCEEDS_MAX_PRECISION", "sqlState" : "22003", "messageParameters" : { "maxPrecision" : "38", @@ -513,7 +513,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_TYPED_LITERAL", + "condition" : "INVALID_TYPED_LITERAL", "sqlState" : "42604", "messageParameters" : { "value" : "'XuZ'", @@ -544,7 +544,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"DATE '1999-01-01'\"", @@ -570,7 +570,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"TIMESTAMP '1999-01-01 00:00:00'\"", @@ -604,7 +604,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"map(1, 2)\"", @@ -630,7 +630,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"array(1, 2)\"", @@ -656,7 +656,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"named_struct(a, 1, b, spark)\"", @@ -682,7 +682,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"X'01'\"", @@ -708,7 +708,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"DATE '1999-01-01'\"", @@ -734,7 +734,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"TIMESTAMP '1999-01-01 00:00:00'\"", @@ -760,7 +760,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"X'2379ACFE'\"", diff --git a/sql/core/src/test/resources/sql-tests/results/map.sql.out b/sql/core/src/test/resources/sql-tests/results/map.sql.out index 71f5491287c3..c62a8d03267a 100644 --- a/sql/core/src/test/resources/sql-tests/results/map.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/map.sql.out @@ -70,7 +70,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.MAP_FUNCTION_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.MAP_FUNCTION_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "\"MAP\"", @@ -96,7 +96,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.MAP_FUNCTION_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.MAP_FUNCTION_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "\"MAP\"", diff --git a/sql/core/src/test/resources/sql-tests/results/mask-functions.sql.out b/sql/core/src/test/resources/sql-tests/results/mask-functions.sql.out index fdaad7168b84..d9942e8ece64 100644 --- a/sql/core/src/test/resources/sql-tests/results/mask-functions.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/mask-functions.sql.out @@ -46,7 +46,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.INPUT_SIZE_NOT_ONE", + "condition" : "DATATYPE_MISMATCH.INPUT_SIZE_NOT_ONE", "sqlState" : "42K09", "messageParameters" : { "exprName" : "upperChar", @@ -437,7 +437,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.NON_FOLDABLE_INPUT", + "condition" : "DATATYPE_MISMATCH.NON_FOLDABLE_INPUT", "sqlState" : "42K09", "messageParameters" : { "inputExpr" : "\"replaceArg\"", @@ -462,7 +462,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.NON_FOLDABLE_INPUT", + "condition" : "DATATYPE_MISMATCH.NON_FOLDABLE_INPUT", "sqlState" : "42K09", "messageParameters" : { "inputExpr" : "\"replaceArg\"", @@ -487,7 +487,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"replaceArg\"", @@ -513,7 +513,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.INPUT_SIZE_NOT_ONE", + "condition" : "DATATYPE_MISMATCH.INPUT_SIZE_NOT_ONE", "sqlState" : "42K09", "messageParameters" : { "exprName" : "digitChar", @@ -536,7 +536,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.INPUT_SIZE_NOT_ONE", + "condition" : "DATATYPE_MISMATCH.INPUT_SIZE_NOT_ONE", "sqlState" : "42K09", "messageParameters" : { "exprName" : "lowerChar", diff --git a/sql/core/src/test/resources/sql-tests/results/math.sql.out b/sql/core/src/test/resources/sql-tests/results/math.sql.out index fb60a920040e..ed9f0469ee6a 100644 --- a/sql/core/src/test/resources/sql-tests/results/math.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/math.sql.out @@ -46,7 +46,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "ARITHMETIC_OVERFLOW", + "condition" : "ARITHMETIC_OVERFLOW", "sqlState" : "22003", "messageParameters" : { "alternative" : "", @@ -70,7 +70,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "ARITHMETIC_OVERFLOW", + "condition" : "ARITHMETIC_OVERFLOW", "sqlState" : "22003", "messageParameters" : { "alternative" : "", @@ -134,7 +134,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "ARITHMETIC_OVERFLOW", + "condition" : "ARITHMETIC_OVERFLOW", "sqlState" : "22003", "messageParameters" : { "alternative" : "", @@ -158,7 +158,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "ARITHMETIC_OVERFLOW", + "condition" : "ARITHMETIC_OVERFLOW", "sqlState" : "22003", "messageParameters" : { "alternative" : "", @@ -222,7 +222,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "ARITHMETIC_OVERFLOW", + "condition" : "ARITHMETIC_OVERFLOW", "sqlState" : "22003", "messageParameters" : { "alternative" : "", @@ -246,7 +246,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "ARITHMETIC_OVERFLOW", + "condition" : "ARITHMETIC_OVERFLOW", "sqlState" : "22003", "messageParameters" : { "alternative" : "", @@ -310,7 +310,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "ARITHMETIC_OVERFLOW", + "condition" : "ARITHMETIC_OVERFLOW", "sqlState" : "22003", "messageParameters" : { "alternative" : "", @@ -334,7 +334,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "ARITHMETIC_OVERFLOW", + "condition" : "ARITHMETIC_OVERFLOW", "sqlState" : "22003", "messageParameters" : { "alternative" : "", @@ -398,7 +398,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "ARITHMETIC_OVERFLOW", + "condition" : "ARITHMETIC_OVERFLOW", "sqlState" : "22003", "messageParameters" : { "alternative" : "", @@ -422,7 +422,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "ARITHMETIC_OVERFLOW", + "condition" : "ARITHMETIC_OVERFLOW", "sqlState" : "22003", "messageParameters" : { "alternative" : "", @@ -486,7 +486,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "ARITHMETIC_OVERFLOW", + "condition" : "ARITHMETIC_OVERFLOW", "sqlState" : "22003", "messageParameters" : { "alternative" : "", @@ -510,7 +510,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "ARITHMETIC_OVERFLOW", + "condition" : "ARITHMETIC_OVERFLOW", "sqlState" : "22003", "messageParameters" : { "alternative" : "", @@ -574,7 +574,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "ARITHMETIC_OVERFLOW", + "condition" : "ARITHMETIC_OVERFLOW", "sqlState" : "22003", "messageParameters" : { "alternative" : "", @@ -598,7 +598,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "ARITHMETIC_OVERFLOW", + "condition" : "ARITHMETIC_OVERFLOW", "sqlState" : "22003", "messageParameters" : { "alternative" : "", @@ -662,7 +662,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "ARITHMETIC_OVERFLOW", + "condition" : "ARITHMETIC_OVERFLOW", "sqlState" : "22003", "messageParameters" : { "alternative" : "", @@ -686,7 +686,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "ARITHMETIC_OVERFLOW", + "condition" : "ARITHMETIC_OVERFLOW", "sqlState" : "22003", "messageParameters" : { "alternative" : "", @@ -734,7 +734,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "ARITHMETIC_OVERFLOW", + "condition" : "ARITHMETIC_OVERFLOW", "sqlState" : "22003", "messageParameters" : { "alternative" : "", @@ -758,7 +758,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "ARITHMETIC_OVERFLOW", + "condition" : "ARITHMETIC_OVERFLOW", "sqlState" : "22003", "messageParameters" : { "alternative" : "", @@ -782,7 +782,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "ARITHMETIC_OVERFLOW", + "condition" : "ARITHMETIC_OVERFLOW", "sqlState" : "22003", "messageParameters" : { "alternative" : "", @@ -878,7 +878,7 @@ struct<> -- !query output org.apache.spark.SparkNumberFormatException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'invalid'", diff --git a/sql/core/src/test/resources/sql-tests/results/misc-functions.sql.out b/sql/core/src/test/resources/sql-tests/results/misc-functions.sql.out index 7316c3136ce0..363d9b739e3d 100644 --- a/sql/core/src/test/resources/sql-tests/results/misc-functions.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/misc-functions.sql.out @@ -70,7 +70,7 @@ struct<> -- !query output org.apache.spark.SparkRuntimeException { - "errorClass" : "USER_RAISED_EXCEPTION", + "condition" : "USER_RAISED_EXCEPTION", "sqlState" : "P0001", "messageParameters" : { "errorMessage" : "'false' is not true!" @@ -85,7 +85,7 @@ struct<> -- !query output org.apache.spark.SparkRuntimeException { - "errorClass" : "USER_RAISED_EXCEPTION", + "condition" : "USER_RAISED_EXCEPTION", "sqlState" : "P0001", "messageParameters" : { "errorMessage" : "'cast(0 as boolean)' is not true!" @@ -100,7 +100,7 @@ struct<> -- !query output org.apache.spark.SparkRuntimeException { - "errorClass" : "USER_RAISED_EXCEPTION", + "condition" : "USER_RAISED_EXCEPTION", "sqlState" : "P0001", "messageParameters" : { "errorMessage" : "'null' is not true!" @@ -115,7 +115,7 @@ struct<> -- !query output org.apache.spark.SparkRuntimeException { - "errorClass" : "USER_RAISED_EXCEPTION", + "condition" : "USER_RAISED_EXCEPTION", "sqlState" : "P0001", "messageParameters" : { "errorMessage" : "'cast(null as boolean)' is not true!" @@ -130,7 +130,7 @@ struct<> -- !query output org.apache.spark.SparkRuntimeException { - "errorClass" : "USER_RAISED_EXCEPTION", + "condition" : "USER_RAISED_EXCEPTION", "sqlState" : "P0001", "messageParameters" : { "errorMessage" : "custom error message" @@ -153,7 +153,7 @@ struct<> -- !query output org.apache.spark.SparkRuntimeException { - "errorClass" : "USER_RAISED_EXCEPTION", + "condition" : "USER_RAISED_EXCEPTION", "sqlState" : "P0001", "messageParameters" : { "errorMessage" : "error message" @@ -168,7 +168,7 @@ struct<> -- !query output org.apache.spark.SparkRuntimeException { - "errorClass" : "USER_RAISED_EXCEPTION", + "condition" : "USER_RAISED_EXCEPTION", "sqlState" : "P0001", "messageParameters" : { "errorMessage" : "too big: 8" @@ -183,7 +183,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", + "condition" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", "sqlState" : "42605", "messageParameters" : { "actualNum" : "2", @@ -208,7 +208,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", + "condition" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", "sqlState" : "42605", "messageParameters" : { "actualNum" : "2", @@ -233,7 +233,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", + "condition" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", "sqlState" : "42605", "messageParameters" : { "actualNum" : "0", @@ -258,7 +258,7 @@ struct<> -- !query output org.apache.spark.SparkRuntimeException { - "errorClass" : "USER_RAISED_EXCEPTION", + "condition" : "USER_RAISED_EXCEPTION", "sqlState" : "P0001", "messageParameters" : { "errorMessage" : "null" @@ -273,7 +273,7 @@ struct<> -- !query output org.apache.spark.SparkRuntimeException { - "errorClass" : "USER_RAISED_EXCEPTION", + "condition" : "USER_RAISED_EXCEPTION", "sqlState" : "P0001", "messageParameters" : { "errorMessage" : "1" diff --git a/sql/core/src/test/resources/sql-tests/results/mode.sql.out b/sql/core/src/test/resources/sql-tests/results/mode.sql.out index 70f253066d4f..d5ab4509102b 100644 --- a/sql/core/src/test/resources/sql-tests/results/mode.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/mode.sql.out @@ -51,7 +51,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_WITHIN_GROUP_EXPRESSION.DISTINCT_UNSUPPORTED", + "condition" : "INVALID_WITHIN_GROUP_EXPRESSION.DISTINCT_UNSUPPORTED", "sqlState" : "42K0K", "messageParameters" : { "funcName" : "`mode`" @@ -97,7 +97,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "UNEXPECTED_INPUT_TYPE", + "condition" : "UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "functionName" : "`mode`", @@ -123,7 +123,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_NULL", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_NULL", "sqlState" : "42K09", "messageParameters" : { "exprName" : "`deterministic`", @@ -146,7 +146,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "NON_FOLDABLE_ARGUMENT", + "condition" : "NON_FOLDABLE_ARGUMENT", "sqlState" : "42K08", "messageParameters" : { "funcName" : "`mode`", @@ -373,7 +373,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_WITHIN_GROUP_EXPRESSION.DISTINCT_UNSUPPORTED", + "condition" : "INVALID_WITHIN_GROUP_EXPRESSION.DISTINCT_UNSUPPORTED", "sqlState" : "42K0K", "messageParameters" : { "funcName" : "`mode`" @@ -397,7 +397,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_WITHIN_GROUP_EXPRESSION.WITHIN_GROUP_MISSING", + "condition" : "INVALID_WITHIN_GROUP_EXPRESSION.WITHIN_GROUP_MISSING", "sqlState" : "42K0K", "messageParameters" : { "funcName" : "`mode`" @@ -421,7 +421,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_WITHIN_GROUP_EXPRESSION.WRONG_NUM_ORDERINGS", + "condition" : "INVALID_WITHIN_GROUP_EXPRESSION.WRONG_NUM_ORDERINGS", "sqlState" : "42K0K", "messageParameters" : { "actualNum" : "1", diff --git a/sql/core/src/test/resources/sql-tests/results/named-function-arguments.sql.out b/sql/core/src/test/resources/sql-tests/results/named-function-arguments.sql.out index e5063dc0cf31..7e9c402443f1 100644 --- a/sql/core/src/test/resources/sql-tests/results/named-function-arguments.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/named-function-arguments.sql.out @@ -229,7 +229,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "UNSUPPORTED_GENERATOR.NESTED_IN_EXPRESSIONS", + "condition" : "UNSUPPORTED_GENERATOR.NESTED_IN_EXPRESSIONS", "sqlState" : "42K0E", "messageParameters" : { "expression" : "\"explode(explode(array(1)))\"" @@ -251,7 +251,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "UNSUPPORTED_GENERATOR.NESTED_IN_EXPRESSIONS", + "condition" : "UNSUPPORTED_GENERATOR.NESTED_IN_EXPRESSIONS", "sqlState" : "42K0E", "messageParameters" : { "expression" : "\"explode(explode(array(1)))\"" @@ -281,7 +281,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"functiontablesubqueryargumentexpression()\"", @@ -307,7 +307,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "UNEXPECTED_POSITIONAL_ARGUMENT", + "condition" : "UNEXPECTED_POSITIONAL_ARGUMENT", "sqlState" : "4274K", "messageParameters" : { "parameterName" : "`lowerChar`", @@ -330,7 +330,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "DUPLICATE_ROUTINE_PARAMETER_ASSIGNMENT.DOUBLE_NAMED_ARGUMENT_REFERENCE", + "condition" : "DUPLICATE_ROUTINE_PARAMETER_ASSIGNMENT.DOUBLE_NAMED_ARGUMENT_REFERENCE", "sqlState" : "4274K", "messageParameters" : { "parameterName" : "`digitChar`", @@ -353,7 +353,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "DUPLICATE_ROUTINE_PARAMETER_ASSIGNMENT.BOTH_POSITIONAL_AND_NAMED", + "condition" : "DUPLICATE_ROUTINE_PARAMETER_ASSIGNMENT.BOTH_POSITIONAL_AND_NAMED", "sqlState" : "4274K", "messageParameters" : { "parameterName" : "`str`", @@ -376,7 +376,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "REQUIRED_PARAMETER_NOT_FOUND", + "condition" : "REQUIRED_PARAMETER_NOT_FOUND", "sqlState" : "4274K", "messageParameters" : { "index" : "0", @@ -400,7 +400,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "UNRECOGNIZED_PARAMETER_NAME", + "condition" : "UNRECOGNIZED_PARAMETER_NAME", "sqlState" : "4274K", "messageParameters" : { "argumentName" : "`cellular`", @@ -424,7 +424,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "NAMED_PARAMETERS_NOT_SUPPORTED", + "condition" : "NAMED_PARAMETERS_NOT_SUPPORTED", "sqlState" : "4274K", "messageParameters" : { "functionName" : "`encode`" @@ -446,7 +446,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", + "condition" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", "sqlState" : "42605", "messageParameters" : { "actualNum" : "6", diff --git a/sql/core/src/test/resources/sql-tests/results/natural-join.sql.out b/sql/core/src/test/resources/sql-tests/results/natural-join.sql.out index 38791e6741fc..f0c87470a382 100644 --- a/sql/core/src/test/resources/sql-tests/results/natural-join.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/natural-join.sql.out @@ -250,7 +250,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`nt2`.`k`", diff --git a/sql/core/src/test/resources/sql-tests/results/nonansi/array.sql.out b/sql/core/src/test/resources/sql-tests/results/nonansi/array.sql.out index c1330c620acf..ba2730748d0a 100644 --- a/sql/core/src/test/resources/sql-tests/results/nonansi/array.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/nonansi/array.sql.out @@ -129,7 +129,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"1\"", @@ -198,7 +198,7 @@ struct<> -- !query output org.apache.spark.SparkRuntimeException { - "errorClass" : "INVALID_INDEX_OF_ZERO", + "condition" : "INVALID_INDEX_OF_ZERO", "sqlState" : "22003" } @@ -314,7 +314,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"map(a, 1, b, 2)\"", @@ -428,7 +428,7 @@ struct<> -- !query output org.apache.spark.SparkRuntimeException { - "errorClass" : "INVALID_INDEX_OF_ZERO", + "condition" : "INVALID_INDEX_OF_ZERO", "sqlState" : "22003", "queryContext" : [ { "objectType" : "", @@ -463,7 +463,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.ARRAY_FUNCTION_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.ARRAY_FUNCTION_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "\"ARRAY\"", @@ -569,7 +569,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"id\"", diff --git a/sql/core/src/test/resources/sql-tests/results/nonansi/cast.sql.out b/sql/core/src/test/resources/sql-tests/results/nonansi/cast.sql.out index 738697c63883..77dd6363fb41 100644 --- a/sql/core/src/test/resources/sql-tests/results/nonansi/cast.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/nonansi/cast.sql.out @@ -351,7 +351,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0029", + "condition" : "_LEGACY_ERROR_TEMP_0029", "messageParameters" : { "literal" : "interval 3 month 1 hour" }, @@ -639,7 +639,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "CAST_OVERFLOW", + "condition" : "CAST_OVERFLOW", "sqlState" : "22003", "messageParameters" : { "sourceType" : "\"INTERVAL HOUR TO SECOND\"", @@ -672,7 +672,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "CAST_OVERFLOW", + "condition" : "CAST_OVERFLOW", "sqlState" : "22003", "messageParameters" : { "sourceType" : "\"INTERVAL MONTH\"", @@ -689,7 +689,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "CAST_OVERFLOW", + "condition" : "CAST_OVERFLOW", "sqlState" : "22003", "messageParameters" : { "sourceType" : "\"INTERVAL SECOND\"", @@ -778,7 +778,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "CAST_OVERFLOW", + "condition" : "CAST_OVERFLOW", "sqlState" : "22003", "messageParameters" : { "sourceType" : "\"INT\"", @@ -795,7 +795,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "CAST_OVERFLOW", + "condition" : "CAST_OVERFLOW", "sqlState" : "22003", "messageParameters" : { "sourceType" : "\"BIGINT\"", @@ -868,7 +868,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "NUMERIC_VALUE_OUT_OF_RANGE.WITH_SUGGESTION", + "condition" : "NUMERIC_VALUE_OUT_OF_RANGE.WITH_SUGGESTION", "sqlState" : "22003", "messageParameters" : { "config" : "\"spark.sql.ansi.enabled\"", @@ -1103,7 +1103,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "UNSUPPORTED_DATATYPE", + "condition" : "UNSUPPORTED_DATATYPE", "sqlState" : "0A000", "messageParameters" : { "typeName" : "\"BINT\"" @@ -1125,7 +1125,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "UNSUPPORTED_DATATYPE", + "condition" : "UNSUPPORTED_DATATYPE", "sqlState" : "0A000", "messageParameters" : { "typeName" : "\"SELECT\"" @@ -1147,7 +1147,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'::'", diff --git a/sql/core/src/test/resources/sql-tests/results/nonansi/date.sql.out b/sql/core/src/test/resources/sql-tests/results/nonansi/date.sql.out index c46c200ff026..2f24f4ce1848 100644 --- a/sql/core/src/test/resources/sql-tests/results/nonansi/date.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/nonansi/date.sql.out @@ -22,7 +22,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_TYPED_LITERAL", + "condition" : "INVALID_TYPED_LITERAL", "sqlState" : "42604", "messageParameters" : { "value" : "'2020-01-01中文'", @@ -69,7 +69,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_TYPED_LITERAL", + "condition" : "INVALID_TYPED_LITERAL", "sqlState" : "42604", "messageParameters" : { "value" : "'015'", @@ -92,7 +92,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_TYPED_LITERAL", + "condition" : "INVALID_TYPED_LITERAL", "sqlState" : "42604", "messageParameters" : { "value" : "'2021-4294967297-11'", @@ -131,7 +131,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", + "condition" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", "sqlState" : "42605", "messageParameters" : { "actualNum" : "1", @@ -334,7 +334,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"1\"", @@ -360,7 +360,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"1.0\"", @@ -386,7 +386,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"10.0\"", @@ -420,7 +420,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "SECOND_FUNCTION_ARGUMENT_NOT_INTEGER", + "condition" : "SECOND_FUNCTION_ARGUMENT_NOT_INTEGER", "sqlState" : "22023", "messageParameters" : { "functionName" : "date_add" @@ -499,7 +499,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"1\"", @@ -525,7 +525,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"1.0\"", @@ -551,7 +551,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"10.0\"", @@ -585,7 +585,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "SECOND_FUNCTION_ARGUMENT_NOT_INTEGER", + "condition" : "SECOND_FUNCTION_ARGUMENT_NOT_INTEGER", "sqlState" : "22023", "messageParameters" : { "functionName" : "date_sub" @@ -632,7 +632,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"int_str\"", @@ -658,7 +658,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"int_str\"", @@ -700,7 +700,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"10.0\"", @@ -758,7 +758,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"2001-09-28\"", @@ -816,7 +816,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"date_str\"", @@ -842,7 +842,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"1\"", @@ -868,7 +868,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"1\"", @@ -923,7 +923,7 @@ struct<> -- !query output org.apache.spark.SparkUpgradeException { - "errorClass" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION.DATETIME_PATTERN_RECOGNITION", + "condition" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION.DATETIME_PATTERN_RECOGNITION", "sqlState" : "42K0B", "messageParameters" : { "config" : "\"spark.sql.legacy.timeParserPolicy\"", @@ -940,7 +940,7 @@ struct<> -- !query output org.apache.spark.SparkUpgradeException { - "errorClass" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION.DATETIME_PATTERN_RECOGNITION", + "condition" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION.DATETIME_PATTERN_RECOGNITION", "sqlState" : "42K0B", "messageParameters" : { "config" : "\"spark.sql.legacy.timeParserPolicy\"", @@ -957,7 +957,7 @@ struct<> -- !query output org.apache.spark.SparkUpgradeException { - "errorClass" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION.DATETIME_PATTERN_RECOGNITION", + "condition" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION.DATETIME_PATTERN_RECOGNITION", "sqlState" : "42K0B", "messageParameters" : { "config" : "\"spark.sql.legacy.timeParserPolicy\"", @@ -1054,7 +1054,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", + "condition" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", "sqlState" : "22023", "messageParameters" : { "functionName" : "`dateadd`", @@ -1078,7 +1078,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", + "condition" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", "sqlState" : "22023", "messageParameters" : { "functionName" : "`date_add`", @@ -1182,7 +1182,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", + "condition" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", "sqlState" : "22023", "messageParameters" : { "functionName" : "`date_diff`", @@ -1206,7 +1206,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", + "condition" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", "sqlState" : "22023", "messageParameters" : { "functionName" : "`datediff`", diff --git a/sql/core/src/test/resources/sql-tests/results/nonansi/datetime-parsing-invalid.sql.out b/sql/core/src/test/resources/sql-tests/results/nonansi/datetime-parsing-invalid.sql.out index fffbb2a4e017..b86e7e326e98 100644 --- a/sql/core/src/test/resources/sql-tests/results/nonansi/datetime-parsing-invalid.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/nonansi/datetime-parsing-invalid.sql.out @@ -47,7 +47,7 @@ struct<> -- !query output org.apache.spark.SparkUpgradeException { - "errorClass" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION.DATETIME_PATTERN_RECOGNITION", + "condition" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION.DATETIME_PATTERN_RECOGNITION", "sqlState" : "42K0B", "messageParameters" : { "config" : "\"spark.sql.legacy.timeParserPolicy\"", diff --git a/sql/core/src/test/resources/sql-tests/results/nonansi/double-quoted-identifiers.sql.out b/sql/core/src/test/resources/sql-tests/results/nonansi/double-quoted-identifiers.sql.out index 81a98a60590f..db6096044dce 100644 --- a/sql/core/src/test/resources/sql-tests/results/nonansi/double-quoted-identifiers.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/nonansi/double-quoted-identifiers.sql.out @@ -6,7 +6,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'\"not_exist\"'", @@ -22,7 +22,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'\"not_exist\"'", @@ -38,7 +38,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'\"not_exist\"'", @@ -54,7 +54,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'\"not_exist\"'", @@ -70,7 +70,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'\"not_exist\"'", @@ -86,7 +86,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'\"hello\"'", @@ -102,7 +102,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'\"not_exist\"'", @@ -118,7 +118,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'\"not_exist\"'", @@ -134,7 +134,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "TABLE_OR_VIEW_NOT_FOUND", + "condition" : "TABLE_OR_VIEW_NOT_FOUND", "sqlState" : "42P01", "messageParameters" : { "relationName" : "`hello`" @@ -156,7 +156,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.analysis.NoSuchNamespaceException { - "errorClass" : "SCHEMA_NOT_FOUND", + "condition" : "SCHEMA_NOT_FOUND", "sqlState" : "42704", "messageParameters" : { "schemaName" : "`spark_catalog`.`not_exist`" @@ -171,7 +171,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "TABLE_OR_VIEW_NOT_FOUND", + "condition" : "TABLE_OR_VIEW_NOT_FOUND", "sqlState" : "42P01", "messageParameters" : { "relationName" : "`not_exist`" @@ -193,7 +193,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "TABLE_OR_VIEW_NOT_FOUND", + "condition" : "TABLE_OR_VIEW_NOT_FOUND", "sqlState" : "42P01", "messageParameters" : { "relationName" : "`not_exist`" @@ -215,7 +215,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "TABLE_OR_VIEW_NOT_FOUND", + "condition" : "TABLE_OR_VIEW_NOT_FOUND", "sqlState" : "42P01", "messageParameters" : { "relationName" : "`not_exist`" @@ -237,7 +237,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "TABLE_OR_VIEW_NOT_FOUND", + "condition" : "TABLE_OR_VIEW_NOT_FOUND", "sqlState" : "42P01", "messageParameters" : { "relationName" : "`not_exist`" @@ -259,7 +259,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "UNRESOLVED_ROUTINE", + "condition" : "UNRESOLVED_ROUTINE", "sqlState" : "42883", "messageParameters" : { "routineName" : "`not_exist`", @@ -282,7 +282,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "UNRESOLVED_ROUTINE", + "condition" : "UNRESOLVED_ROUTINE", "sqlState" : "42883", "messageParameters" : { "routineName" : "`not_exist`.`not_exist`", @@ -369,7 +369,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'\"myschema\"'", @@ -386,7 +386,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'\"myview\"'", @@ -402,7 +402,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'\"a2\"'", @@ -418,7 +418,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'\"myview\"'", @@ -434,7 +434,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'\"myschema\"'", diff --git a/sql/core/src/test/resources/sql-tests/results/nonansi/higher-order-functions.sql.out b/sql/core/src/test/resources/sql-tests/results/nonansi/higher-order-functions.sql.out index 7bfc35a61e09..1ac44e7447de 100644 --- a/sql/core/src/test/resources/sql-tests/results/nonansi/higher-order-functions.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/nonansi/higher-order-functions.sql.out @@ -18,7 +18,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_LAMBDA_FUNCTION_CALL.NON_HIGHER_ORDER_FUNCTION", + "condition" : "INVALID_LAMBDA_FUNCTION_CALL.NON_HIGHER_ORDER_FUNCTION", "sqlState" : "42K0D", "messageParameters" : { "class" : "org.apache.spark.sql.catalyst.expressions.Upper" @@ -40,7 +40,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_LAMBDA_FUNCTION_CALL.NON_HIGHER_ORDER_FUNCTION", + "condition" : "INVALID_LAMBDA_FUNCTION_CALL.NON_HIGHER_ORDER_FUNCTION", "sqlState" : "42K0D", "messageParameters" : { "class" : "org.apache.spark.sql.catalyst.expressions.CeilExpressionBuilder$" diff --git a/sql/core/src/test/resources/sql-tests/results/nonansi/interval.sql.out b/sql/core/src/test/resources/sql-tests/results/nonansi/interval.sql.out index a8a0423bdb3e..4b8191da0927 100644 --- a/sql/core/src/test/resources/sql-tests/results/nonansi/interval.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/nonansi/interval.sql.out @@ -14,7 +14,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0029", + "condition" : "_LEGACY_ERROR_TEMP_0029", "messageParameters" : { "literal" : "interval 4 month 2 weeks 3 microseconds" }, @@ -191,7 +191,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"STRING\"", @@ -215,7 +215,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"STRING\"", @@ -239,7 +239,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "INTERVAL_DIVIDED_BY_ZERO", + "condition" : "INTERVAL_DIVIDED_BY_ZERO", "sqlState" : "22012", "queryContext" : [ { "objectType" : "", @@ -282,7 +282,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "INTERVAL_DIVIDED_BY_ZERO", + "condition" : "INTERVAL_DIVIDED_BY_ZERO", "sqlState" : "22012", "queryContext" : [ { "objectType" : "", @@ -325,7 +325,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"INT\"", @@ -349,7 +349,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"INT\"", @@ -373,7 +373,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"VOID\"", @@ -397,7 +397,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"VOID\"", @@ -421,7 +421,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0029", + "condition" : "_LEGACY_ERROR_TEMP_0029", "messageParameters" : { "literal" : "interval '-1 month 1 day -1 second'" }, @@ -458,7 +458,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0029", + "condition" : "_LEGACY_ERROR_TEMP_0029", "messageParameters" : { "literal" : "interval -1 month 1 day -1 second" }, @@ -495,7 +495,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0029", + "condition" : "_LEGACY_ERROR_TEMP_0029", "messageParameters" : { "literal" : "interval '-1 month 1 day -1 second'" }, @@ -532,7 +532,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0029", + "condition" : "_LEGACY_ERROR_TEMP_0029", "messageParameters" : { "literal" : "interval -1 month 1 day -1 second" }, @@ -825,7 +825,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "INTERVAL_ARITHMETIC_OVERFLOW.WITHOUT_SUGGESTION", + "condition" : "INTERVAL_ARITHMETIC_OVERFLOW.WITHOUT_SUGGESTION", "sqlState" : "22015", "queryContext" : [ { "objectType" : "", @@ -876,7 +876,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "INTERVAL_ARITHMETIC_OVERFLOW.WITHOUT_SUGGESTION", + "condition" : "INTERVAL_ARITHMETIC_OVERFLOW.WITHOUT_SUGGESTION", "sqlState" : "22015", "queryContext" : [ { "objectType" : "", @@ -903,7 +903,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "INTERVAL_ARITHMETIC_OVERFLOW.WITHOUT_SUGGESTION", + "condition" : "INTERVAL_ARITHMETIC_OVERFLOW.WITHOUT_SUGGESTION", "sqlState" : "22015", "queryContext" : [ { "objectType" : "", @@ -986,7 +986,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0029", + "condition" : "_LEGACY_ERROR_TEMP_0029", "messageParameters" : { "literal" : "interval 1 year 2 month 3 week 4 day 5 hour 6 minute 7 seconds 8 millisecond 9 microsecond" }, @@ -1031,7 +1031,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0029", + "condition" : "_LEGACY_ERROR_TEMP_0029", "messageParameters" : { "literal" : "interval '30' year '25' month '-100' day '40' hour '80' minute '299.889987299' second" }, @@ -1172,7 +1172,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_INTERVAL_FORMAT.UNMATCHED_FORMAT_STRING_WITH_NOTICE", + "condition" : "INVALID_INTERVAL_FORMAT.UNMATCHED_FORMAT_STRING_WITH_NOTICE", "sqlState" : "22006", "messageParameters" : { "input" : "20 15:40:32.99899999", @@ -1197,7 +1197,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_INTERVAL_FORMAT.UNMATCHED_FORMAT_STRING_WITH_NOTICE", + "condition" : "INVALID_INTERVAL_FORMAT.UNMATCHED_FORMAT_STRING_WITH_NOTICE", "sqlState" : "22006", "messageParameters" : { "input" : "20 15:40:32.99899999", @@ -1222,7 +1222,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_INTERVAL_FORMAT.UNMATCHED_FORMAT_STRING_WITH_NOTICE", + "condition" : "INVALID_INTERVAL_FORMAT.UNMATCHED_FORMAT_STRING_WITH_NOTICE", "sqlState" : "22006", "messageParameters" : { "input" : "15:40:32.99899999", @@ -1247,7 +1247,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_INTERVAL_FORMAT.UNMATCHED_FORMAT_STRING_WITH_NOTICE", + "condition" : "INVALID_INTERVAL_FORMAT.UNMATCHED_FORMAT_STRING_WITH_NOTICE", "sqlState" : "22006", "messageParameters" : { "input" : "15:40.99899999", @@ -1272,7 +1272,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_INTERVAL_FORMAT.UNMATCHED_FORMAT_STRING_WITH_NOTICE", + "condition" : "INVALID_INTERVAL_FORMAT.UNMATCHED_FORMAT_STRING_WITH_NOTICE", "sqlState" : "22006", "messageParameters" : { "input" : "15:40", @@ -1297,7 +1297,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_INTERVAL_FORMAT.UNMATCHED_FORMAT_STRING_WITH_NOTICE", + "condition" : "INVALID_INTERVAL_FORMAT.UNMATCHED_FORMAT_STRING_WITH_NOTICE", "sqlState" : "22006", "messageParameters" : { "input" : "20 40:32.99899999", @@ -1322,7 +1322,7 @@ struct<> -- !query output org.apache.spark.SparkIllegalArgumentException { - "errorClass" : "INVALID_INTERVAL_FORMAT.INVALID_UNIT", + "condition" : "INVALID_INTERVAL_FORMAT.INVALID_UNIT", "sqlState" : "22006", "messageParameters" : { "input" : " 10 nanoseconds", @@ -1418,7 +1418,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'fake_unit'", @@ -1434,7 +1434,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0027", + "condition" : "_LEGACY_ERROR_TEMP_0027", "queryContext" : [ { "objectType" : "", "objectName" : "", @@ -1452,7 +1452,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_INTERVAL_FORMAT.UNSUPPORTED_FROM_TO_EXPRESSION", + "condition" : "INVALID_INTERVAL_FORMAT.UNSUPPORTED_FROM_TO_EXPRESSION", "sqlState" : "22006", "messageParameters" : { "from" : "year", @@ -1476,7 +1476,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0024", + "condition" : "_LEGACY_ERROR_TEMP_0024", "queryContext" : [ { "objectType" : "", "objectName" : "", @@ -1494,7 +1494,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0024", + "condition" : "_LEGACY_ERROR_TEMP_0024", "queryContext" : [ { "objectType" : "", "objectName" : "", @@ -1512,7 +1512,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0024", + "condition" : "_LEGACY_ERROR_TEMP_0024", "queryContext" : [ { "objectType" : "", "objectName" : "", @@ -1530,7 +1530,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0024", + "condition" : "_LEGACY_ERROR_TEMP_0024", "queryContext" : [ { "objectType" : "", "objectName" : "", @@ -1548,7 +1548,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0024", + "condition" : "_LEGACY_ERROR_TEMP_0024", "queryContext" : [ { "objectType" : "", "objectName" : "", @@ -1566,7 +1566,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0024", + "condition" : "_LEGACY_ERROR_TEMP_0024", "queryContext" : [ { "objectType" : "", "objectName" : "", @@ -1584,7 +1584,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0024", + "condition" : "_LEGACY_ERROR_TEMP_0024", "queryContext" : [ { "objectType" : "", "objectName" : "", @@ -1602,7 +1602,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "UNRESOLVED_ROUTINE", + "condition" : "UNRESOLVED_ROUTINE", "sqlState" : "42883", "messageParameters" : { "routineName" : "`interval`", @@ -1625,7 +1625,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "UNRESOLVED_ROUTINE", + "condition" : "UNRESOLVED_ROUTINE", "sqlState" : "42883", "messageParameters" : { "routineName" : "`interval`", @@ -1648,7 +1648,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'day'", @@ -1664,7 +1664,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "UNRESOLVED_ROUTINE", + "condition" : "UNRESOLVED_ROUTINE", "sqlState" : "42883", "messageParameters" : { "routineName" : "`interval`", @@ -1687,7 +1687,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "UNRESOLVED_ROUTINE", + "condition" : "UNRESOLVED_ROUTINE", "sqlState" : "42883", "messageParameters" : { "routineName" : "`interval`", @@ -1710,7 +1710,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'days'", @@ -1734,7 +1734,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_INTERVAL_FORMAT.INTERVAL_PARSING", + "condition" : "INVALID_INTERVAL_FORMAT.INTERVAL_PARSING", "sqlState" : "22006", "messageParameters" : { "input" : "178956970-8", @@ -1793,7 +1793,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"INTERVAL YEAR\"", @@ -1833,7 +1833,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"INTERVAL YEAR\"", @@ -1857,7 +1857,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"INTERVAL YEAR\"", @@ -1905,7 +1905,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"INTERVAL YEAR\"", @@ -1929,7 +1929,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"INTERVAL YEAR\"", @@ -1969,7 +1969,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"INTERVAL '2-2' YEAR TO MONTH\"", @@ -1995,7 +1995,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"INTERVAL '2-2' YEAR TO MONTH\"", @@ -2021,7 +2021,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"INTERVAL '2-2' YEAR TO MONTH\"", @@ -2047,7 +2047,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"INTERVAL DAY\"", @@ -2071,7 +2071,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"1\"", @@ -2097,7 +2097,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"INT\"", @@ -2121,7 +2121,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"1\"", @@ -2147,7 +2147,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"INTERVAL MONTH\"", @@ -2219,7 +2219,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_INTERVAL_FORMAT.UNMATCHED_FORMAT_STRING", + "condition" : "INVALID_INTERVAL_FORMAT.UNMATCHED_FORMAT_STRING", "sqlState" : "22006", "messageParameters" : { "input" : "-\t2-2\t", @@ -2252,7 +2252,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_INTERVAL_FORMAT.UNMATCHED_FORMAT_STRING_WITH_NOTICE", + "condition" : "INVALID_INTERVAL_FORMAT.UNMATCHED_FORMAT_STRING_WITH_NOTICE", "sqlState" : "22006", "messageParameters" : { "input" : "\n-\t10\t 12:34:46.789\t", @@ -2277,7 +2277,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_TYPED_LITERAL", + "condition" : "INVALID_TYPED_LITERAL", "sqlState" : "42604", "messageParameters" : { "value" : "'中文 interval 1 day'", @@ -2300,7 +2300,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_TYPED_LITERAL", + "condition" : "INVALID_TYPED_LITERAL", "sqlState" : "42604", "messageParameters" : { "value" : "'interval中文 1 day'", @@ -2323,7 +2323,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_TYPED_LITERAL", + "condition" : "INVALID_TYPED_LITERAL", "sqlState" : "42604", "messageParameters" : { "value" : "'interval 1中文day'", @@ -2346,7 +2346,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "INTERVAL_ARITHMETIC_OVERFLOW.WITHOUT_SUGGESTION", + "condition" : "INTERVAL_ARITHMETIC_OVERFLOW.WITHOUT_SUGGESTION", "sqlState" : "22015" } @@ -2358,7 +2358,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "INTERVAL_ARITHMETIC_OVERFLOW.WITH_SUGGESTION", + "condition" : "INTERVAL_ARITHMETIC_OVERFLOW.WITH_SUGGESTION", "sqlState" : "22015", "messageParameters" : { "functionName" : "`try_subtract`" @@ -2373,7 +2373,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "INTERVAL_ARITHMETIC_OVERFLOW.WITH_SUGGESTION", + "condition" : "INTERVAL_ARITHMETIC_OVERFLOW.WITH_SUGGESTION", "sqlState" : "22015", "messageParameters" : { "functionName" : "`try_add`" @@ -2458,7 +2458,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_TYPED_LITERAL", + "condition" : "INVALID_TYPED_LITERAL", "sqlState" : "42604", "messageParameters" : { "value" : "'+'", @@ -2481,7 +2481,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_TYPED_LITERAL", + "condition" : "INVALID_TYPED_LITERAL", "sqlState" : "42604", "messageParameters" : { "value" : "'+.'", @@ -2504,7 +2504,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_TYPED_LITERAL", + "condition" : "INVALID_TYPED_LITERAL", "sqlState" : "42604", "messageParameters" : { "value" : "'1'", @@ -2527,7 +2527,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_TYPED_LITERAL", + "condition" : "INVALID_TYPED_LITERAL", "sqlState" : "42604", "messageParameters" : { "value" : "'1.2'", @@ -2550,7 +2550,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_TYPED_LITERAL", + "condition" : "INVALID_TYPED_LITERAL", "sqlState" : "42604", "messageParameters" : { "value" : "'- 2'", @@ -2573,7 +2573,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_TYPED_LITERAL", + "condition" : "INVALID_TYPED_LITERAL", "sqlState" : "42604", "messageParameters" : { "value" : "'1 day -'", @@ -2596,7 +2596,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_TYPED_LITERAL", + "condition" : "INVALID_TYPED_LITERAL", "sqlState" : "42604", "messageParameters" : { "value" : "'1 day 1'", @@ -2619,7 +2619,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0026", + "condition" : "_LEGACY_ERROR_TEMP_0026", "messageParameters" : { "value" : "1 day 2" }, @@ -2640,7 +2640,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0026", + "condition" : "_LEGACY_ERROR_TEMP_0026", "messageParameters" : { "value" : "interval 1" }, @@ -2685,7 +2685,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "INTERVAL_ARITHMETIC_OVERFLOW.WITH_SUGGESTION", + "condition" : "INTERVAL_ARITHMETIC_OVERFLOW.WITH_SUGGESTION", "sqlState" : "22015", "messageParameters" : { "functionName" : "`try_divide`" @@ -2707,7 +2707,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "INTERVAL_ARITHMETIC_OVERFLOW.WITH_SUGGESTION", + "condition" : "INTERVAL_ARITHMETIC_OVERFLOW.WITH_SUGGESTION", "sqlState" : "22015", "messageParameters" : { "functionName" : "`try_divide`" @@ -2763,7 +2763,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "INTERVAL_ARITHMETIC_OVERFLOW.WITH_SUGGESTION", + "condition" : "INTERVAL_ARITHMETIC_OVERFLOW.WITH_SUGGESTION", "sqlState" : "22015", "messageParameters" : { "functionName" : "`try_divide`" @@ -2785,7 +2785,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "INTERVAL_ARITHMETIC_OVERFLOW.WITH_SUGGESTION", + "condition" : "INTERVAL_ARITHMETIC_OVERFLOW.WITH_SUGGESTION", "sqlState" : "22015", "messageParameters" : { "functionName" : "`try_divide`" @@ -2921,7 +2921,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0063", + "condition" : "_LEGACY_ERROR_TEMP_0063", "messageParameters" : { "msg" : "requirement failed: day 106751992 outside range [0, 106751991]" }, @@ -2942,7 +2942,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0063", + "condition" : "_LEGACY_ERROR_TEMP_0063", "messageParameters" : { "msg" : "requirement failed: day 106751992 outside range [0, 106751991]" }, @@ -2963,7 +2963,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0063", + "condition" : "_LEGACY_ERROR_TEMP_0063", "messageParameters" : { "msg" : "requirement failed: hour 2562047789 outside range [0, 2562047788]" }, @@ -2984,7 +2984,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0063", + "condition" : "_LEGACY_ERROR_TEMP_0063", "messageParameters" : { "msg" : "requirement failed: hour 2562047789 outside range [0, 2562047788]" }, @@ -3005,7 +3005,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0063", + "condition" : "_LEGACY_ERROR_TEMP_0063", "messageParameters" : { "msg" : "requirement failed: minute 153722867281 outside range [0, 153722867280]" }, @@ -3026,7 +3026,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0063", + "condition" : "_LEGACY_ERROR_TEMP_0063", "messageParameters" : { "msg" : "requirement failed: minute 153722867281 outside range [0, 153722867280]" }, @@ -3175,7 +3175,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"INTERVAL MONTH\"", @@ -3199,7 +3199,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"INTERVAL DAY\"", @@ -3223,7 +3223,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"INTERVAL DAY\"", @@ -3247,7 +3247,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"INTERVAL DAY\"", @@ -3271,7 +3271,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"STRING\"", @@ -3295,7 +3295,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"STRING\"", @@ -3319,7 +3319,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"STRING\"", @@ -3343,7 +3343,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"INTERVAL YEAR\"", @@ -3367,7 +3367,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"INTERVAL YEAR\"", @@ -3391,7 +3391,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"INTERVAL YEAR\"", @@ -3415,7 +3415,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"STRING\"", @@ -3439,7 +3439,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"STRING\"", @@ -3463,7 +3463,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"STRING\"", @@ -3503,7 +3503,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "(\"INTERVAL MONTH\" or \"INTERVAL DAY\")", @@ -3543,7 +3543,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "(\"INTERVAL MONTH\" or \"INTERVAL DAY\")", @@ -3615,7 +3615,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"INTERVAL MONTH\"", diff --git a/sql/core/src/test/resources/sql-tests/results/nonansi/literals.sql.out b/sql/core/src/test/resources/sql-tests/results/nonansi/literals.sql.out index 4e4c70cc333b..b27e1fee2d61 100644 --- a/sql/core/src/test/resources/sql-tests/results/nonansi/literals.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/nonansi/literals.sql.out @@ -38,7 +38,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_NUMERIC_LITERAL_RANGE", + "condition" : "INVALID_NUMERIC_LITERAL_RANGE", "sqlState" : "22003", "messageParameters" : { "maxValue" : "127", @@ -79,7 +79,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_NUMERIC_LITERAL_RANGE", + "condition" : "INVALID_NUMERIC_LITERAL_RANGE", "sqlState" : "22003", "messageParameters" : { "maxValue" : "32767", @@ -120,7 +120,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_NUMERIC_LITERAL_RANGE", + "condition" : "INVALID_NUMERIC_LITERAL_RANGE", "sqlState" : "22003", "messageParameters" : { "maxValue" : "9223372036854775807", @@ -177,7 +177,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "DECIMAL_PRECISION_EXCEEDS_MAX_PRECISION", + "condition" : "DECIMAL_PRECISION_EXCEEDS_MAX_PRECISION", "sqlState" : "22003", "messageParameters" : { "maxPrecision" : "38", @@ -193,7 +193,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "DECIMAL_PRECISION_EXCEEDS_MAX_PRECISION", + "condition" : "DECIMAL_PRECISION_EXCEEDS_MAX_PRECISION", "sqlState" : "22003", "messageParameters" : { "maxPrecision" : "38", @@ -225,7 +225,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_NUMERIC_LITERAL_RANGE", + "condition" : "INVALID_NUMERIC_LITERAL_RANGE", "sqlState" : "22003", "messageParameters" : { "maxValue" : "3.4028234663852886E+38", @@ -266,7 +266,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'.'", @@ -282,7 +282,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_NUMERIC_LITERAL_RANGE", + "condition" : "INVALID_NUMERIC_LITERAL_RANGE", "sqlState" : "22003", "messageParameters" : { "maxValue" : "1.7976931348623157E+308", @@ -397,7 +397,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_TYPED_LITERAL", + "condition" : "INVALID_TYPED_LITERAL", "sqlState" : "42604", "messageParameters" : { "value" : "'mar 11 2016'", @@ -428,7 +428,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_TYPED_LITERAL", + "condition" : "INVALID_TYPED_LITERAL", "sqlState" : "42604", "messageParameters" : { "value" : "'2016-33-11 20:54:00.000'", @@ -451,7 +451,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "UNSUPPORTED_TYPED_LITERAL", + "condition" : "UNSUPPORTED_TYPED_LITERAL", "sqlState" : "0A000", "messageParameters" : { "supportedTypes" : "\"DATE\", \"TIMESTAMP_NTZ\", \"TIMESTAMP_LTZ\", \"TIMESTAMP\", \"INTERVAL\", \"X\"", @@ -482,7 +482,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "DECIMAL_PRECISION_EXCEEDS_MAX_PRECISION", + "condition" : "DECIMAL_PRECISION_EXCEEDS_MAX_PRECISION", "sqlState" : "22003", "messageParameters" : { "maxPrecision" : "38", @@ -513,7 +513,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_TYPED_LITERAL", + "condition" : "INVALID_TYPED_LITERAL", "sqlState" : "42604", "messageParameters" : { "value" : "'XuZ'", @@ -544,7 +544,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"DATE '1999-01-01'\"", @@ -570,7 +570,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"TIMESTAMP '1999-01-01 00:00:00'\"", @@ -604,7 +604,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"map(1, 2)\"", @@ -630,7 +630,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"array(1, 2)\"", @@ -656,7 +656,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"named_struct(a, 1, b, spark)\"", @@ -682,7 +682,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"X'01'\"", @@ -708,7 +708,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"DATE '1999-01-01'\"", @@ -734,7 +734,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"TIMESTAMP '1999-01-01 00:00:00'\"", @@ -760,7 +760,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"X'2379ACFE'\"", diff --git a/sql/core/src/test/resources/sql-tests/results/nonansi/map.sql.out b/sql/core/src/test/resources/sql-tests/results/nonansi/map.sql.out index 71f5491287c3..c62a8d03267a 100644 --- a/sql/core/src/test/resources/sql-tests/results/nonansi/map.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/nonansi/map.sql.out @@ -70,7 +70,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.MAP_FUNCTION_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.MAP_FUNCTION_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "\"MAP\"", @@ -96,7 +96,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.MAP_FUNCTION_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.MAP_FUNCTION_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "\"MAP\"", diff --git a/sql/core/src/test/resources/sql-tests/results/nonansi/string-functions.sql.out b/sql/core/src/test/resources/sql-tests/results/nonansi/string-functions.sql.out index 3f9f24f817f2..961aab410048 100644 --- a/sql/core/src/test/resources/sql-tests/results/nonansi/string-functions.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/nonansi/string-functions.sql.out @@ -6,7 +6,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", + "condition" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", "sqlState" : "42605", "messageParameters" : { "actualNum" : "0", @@ -24,7 +24,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", + "condition" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", "sqlState" : "42605", "messageParameters" : { "actualNum" : "0", @@ -218,7 +218,7 @@ struct<> -- !query output org.apache.spark.SparkRuntimeException { - "errorClass" : "INVALID_INDEX_OF_ZERO", + "condition" : "INVALID_INDEX_OF_ZERO", "sqlState" : "22003" } @@ -774,7 +774,7 @@ struct<> -- !query output org.apache.spark.SparkIllegalArgumentException { - "errorClass" : "INVALID_PARAMETER_VALUE.CHARSET", + "condition" : "INVALID_PARAMETER_VALUE.CHARSET", "sqlState" : "22023", "messageParameters" : { "charset" : "WINDOWS-1252", @@ -792,7 +792,7 @@ struct<> -- !query output org.apache.spark.SparkIllegalArgumentException { - "errorClass" : "INVALID_PARAMETER_VALUE.CHARSET", + "condition" : "INVALID_PARAMETER_VALUE.CHARSET", "sqlState" : "22023", "messageParameters" : { "charset" : "WINDOWS-1252", @@ -810,7 +810,7 @@ struct<> -- !query output org.apache.spark.SparkIllegalArgumentException { - "errorClass" : "INVALID_PARAMETER_VALUE.CHARSET", + "condition" : "INVALID_PARAMETER_VALUE.CHARSET", "sqlState" : "22023", "messageParameters" : { "charset" : "Windows-xxx", @@ -828,7 +828,7 @@ struct<> -- !query output org.apache.spark.SparkIllegalArgumentException { - "errorClass" : "INVALID_PARAMETER_VALUE.CHARSET", + "condition" : "INVALID_PARAMETER_VALUE.CHARSET", "sqlState" : "22023", "messageParameters" : { "charset" : "Windows-xxx", @@ -878,7 +878,7 @@ struct<> -- !query output org.apache.spark.SparkRuntimeException { - "errorClass" : "MALFORMED_CHARACTER_CODING", + "condition" : "MALFORMED_CHARACTER_CODING", "sqlState" : "22000", "messageParameters" : { "charset" : "US-ASCII", @@ -894,7 +894,7 @@ struct<> -- !query output org.apache.spark.SparkRuntimeException { - "errorClass" : "MALFORMED_CHARACTER_CODING", + "condition" : "MALFORMED_CHARACTER_CODING", "sqlState" : "22000", "messageParameters" : { "charset" : "US-ASCII", @@ -950,7 +950,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", + "condition" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", "sqlState" : "42605", "messageParameters" : { "actualNum" : "0", @@ -975,7 +975,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", + "condition" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", "sqlState" : "42605", "messageParameters" : { "actualNum" : "1", @@ -1072,7 +1072,7 @@ struct<> -- !query output org.apache.spark.SparkIllegalArgumentException { - "errorClass" : "INVALID_PARAMETER_VALUE.CHARSET", + "condition" : "INVALID_PARAMETER_VALUE.CHARSET", "sqlState" : "22023", "messageParameters" : { "charset" : "Windows-xxx", @@ -1090,7 +1090,7 @@ struct<> -- !query output org.apache.spark.SparkIllegalArgumentException { - "errorClass" : "INVALID_PARAMETER_VALUE.CHARSET", + "condition" : "INVALID_PARAMETER_VALUE.CHARSET", "sqlState" : "22023", "messageParameters" : { "charset" : "Windows-xxx", @@ -1140,7 +1140,7 @@ struct<> -- !query output org.apache.spark.SparkIllegalArgumentException { - "errorClass" : "INVALID_PARAMETER_VALUE.CHARSET", + "condition" : "INVALID_PARAMETER_VALUE.CHARSET", "sqlState" : "22023", "messageParameters" : { "charset" : "WINDOWS-1252", @@ -1158,7 +1158,7 @@ struct<> -- !query output org.apache.spark.SparkIllegalArgumentException { - "errorClass" : "INVALID_PARAMETER_VALUE.CHARSET", + "condition" : "INVALID_PARAMETER_VALUE.CHARSET", "sqlState" : "22023", "messageParameters" : { "charset" : "WINDOWS-1252", @@ -1208,7 +1208,7 @@ struct<> -- !query output org.apache.spark.SparkRuntimeException { - "errorClass" : "MALFORMED_CHARACTER_CODING", + "condition" : "MALFORMED_CHARACTER_CODING", "sqlState" : "22000", "messageParameters" : { "charset" : "US-ASCII", @@ -1224,7 +1224,7 @@ struct<> -- !query output org.apache.spark.SparkRuntimeException { - "errorClass" : "MALFORMED_CHARACTER_CODING", + "condition" : "MALFORMED_CHARACTER_CODING", "sqlState" : "22000", "messageParameters" : { "charset" : "US-ASCII", @@ -1600,7 +1600,7 @@ struct<> -- !query output org.apache.spark.SparkIllegalArgumentException { - "errorClass" : "CONVERSION_INVALID_INPUT", + "condition" : "CONVERSION_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "fmt" : "'BASE64'", @@ -1628,7 +1628,7 @@ struct<> -- !query output org.apache.spark.SparkIllegalArgumentException { - "errorClass" : "CONVERSION_INVALID_INPUT", + "condition" : "CONVERSION_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "fmt" : "'BASE64'", @@ -1646,7 +1646,7 @@ struct<> -- !query output org.apache.spark.SparkIllegalArgumentException { - "errorClass" : "CONVERSION_INVALID_INPUT", + "condition" : "CONVERSION_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "fmt" : "'BASE64'", @@ -1664,7 +1664,7 @@ struct<> -- !query output org.apache.spark.SparkIllegalArgumentException { - "errorClass" : "CONVERSION_INVALID_INPUT", + "condition" : "CONVERSION_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "fmt" : "'BASE64'", @@ -1682,7 +1682,7 @@ struct<> -- !query output org.apache.spark.SparkIllegalArgumentException { - "errorClass" : "CONVERSION_INVALID_INPUT", + "condition" : "CONVERSION_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "fmt" : "'BASE64'", @@ -1700,7 +1700,7 @@ struct<> -- !query output org.apache.spark.SparkIllegalArgumentException { - "errorClass" : "CONVERSION_INVALID_INPUT", + "condition" : "CONVERSION_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "fmt" : "'BASE64'", @@ -1718,7 +1718,7 @@ struct<> -- !query output org.apache.spark.SparkIllegalArgumentException { - "errorClass" : "CONVERSION_INVALID_INPUT", + "condition" : "CONVERSION_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "fmt" : "'BASE64'", @@ -1825,7 +1825,7 @@ struct<> -- !query output org.apache.spark.SparkIllegalArgumentException { - "errorClass" : "CONVERSION_INVALID_INPUT", + "condition" : "CONVERSION_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "fmt" : "'HEX'", @@ -1843,7 +1843,7 @@ struct<> -- !query output org.apache.spark.SparkIllegalArgumentException { - "errorClass" : "CONVERSION_INVALID_INPUT", + "condition" : "CONVERSION_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "fmt" : "'HEX'", @@ -1869,7 +1869,7 @@ struct<> -- !query output org.apache.spark.SparkIllegalArgumentException { - "errorClass" : "CONVERSION_INVALID_INPUT", + "condition" : "CONVERSION_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "fmt" : "'BASE64'", @@ -1887,7 +1887,7 @@ struct<> -- !query output org.apache.spark.SparkIllegalArgumentException { - "errorClass" : "CONVERSION_INVALID_INPUT", + "condition" : "CONVERSION_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "fmt" : "'HEX'", @@ -1946,7 +1946,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.INVALID_ARG_VALUE", + "condition" : "DATATYPE_MISMATCH.INVALID_ARG_VALUE", "sqlState" : "42K09", "messageParameters" : { "inputName" : "fmt", @@ -1972,7 +1972,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.INVALID_ARG_VALUE", + "condition" : "DATATYPE_MISMATCH.INVALID_ARG_VALUE", "sqlState" : "42K09", "messageParameters" : { "inputName" : "fmt", @@ -2006,7 +2006,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.NON_FOLDABLE_INPUT", + "condition" : "DATATYPE_MISMATCH.NON_FOLDABLE_INPUT", "sqlState" : "42K09", "messageParameters" : { "inputExpr" : "\"fmtField\"", @@ -2247,7 +2247,7 @@ struct<> -- !query output org.apache.spark.SparkIllegalArgumentException { - "errorClass" : "INVALID_UTF8_STRING", + "condition" : "INVALID_UTF8_STRING", "sqlState" : "22029", "messageParameters" : { "str" : "\\x80" diff --git a/sql/core/src/test/resources/sql-tests/results/nonansi/timestamp.sql.out b/sql/core/src/test/resources/sql-tests/results/nonansi/timestamp.sql.out index 0e0b014a3b16..d3792fb8da75 100644 --- a/sql/core/src/test/resources/sql-tests/results/nonansi/timestamp.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/nonansi/timestamp.sql.out @@ -14,7 +14,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_TYPED_LITERAL", + "condition" : "INVALID_TYPED_LITERAL", "sqlState" : "42604", "messageParameters" : { "value" : "'2019-01-01中文'", @@ -37,7 +37,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_TYPED_LITERAL", + "condition" : "INVALID_TYPED_LITERAL", "sqlState" : "42604", "messageParameters" : { "value" : "'4294967297'", @@ -60,7 +60,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_TYPED_LITERAL", + "condition" : "INVALID_TYPED_LITERAL", "sqlState" : "42604", "messageParameters" : { "value" : "'2021-01-01T12:30:4294967297.123456'", @@ -633,7 +633,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"2011-11-11 11:11:10\"", @@ -659,7 +659,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"2011-11-11 11:11:11\"", @@ -709,7 +709,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"str\"", @@ -735,7 +735,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"str\"", @@ -761,7 +761,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP\"", @@ -785,7 +785,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DOUBLE\"", @@ -809,7 +809,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP\"", @@ -833,7 +833,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"VOID\"", @@ -880,7 +880,7 @@ struct<> -- !query output org.apache.spark.SparkUpgradeException { - "errorClass" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION.DATETIME_PATTERN_RECOGNITION", + "condition" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION.DATETIME_PATTERN_RECOGNITION", "sqlState" : "42K0B", "messageParameters" : { "config" : "\"spark.sql.legacy.timeParserPolicy\"", @@ -897,7 +897,7 @@ struct<> -- !query output org.apache.spark.SparkUpgradeException { - "errorClass" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION.DATETIME_PATTERN_RECOGNITION", + "condition" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION.DATETIME_PATTERN_RECOGNITION", "sqlState" : "42K0B", "messageParameters" : { "config" : "\"spark.sql.legacy.timeParserPolicy\"", @@ -914,7 +914,7 @@ struct<> -- !query output org.apache.spark.SparkUpgradeException { - "errorClass" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION.DATETIME_PATTERN_RECOGNITION", + "condition" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION.DATETIME_PATTERN_RECOGNITION", "sqlState" : "42K0B", "messageParameters" : { "config" : "\"spark.sql.legacy.timeParserPolicy\"", @@ -931,7 +931,7 @@ struct<> -- !query output org.apache.spark.SparkUpgradeException { - "errorClass" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION.DATETIME_PATTERN_RECOGNITION", + "condition" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION.DATETIME_PATTERN_RECOGNITION", "sqlState" : "42K0B", "messageParameters" : { "config" : "\"spark.sql.legacy.timeParserPolicy\"", @@ -948,7 +948,7 @@ struct<> -- !query output org.apache.spark.SparkUpgradeException { - "errorClass" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION.DATETIME_PATTERN_RECOGNITION", + "condition" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION.DATETIME_PATTERN_RECOGNITION", "sqlState" : "42K0B", "messageParameters" : { "config" : "\"spark.sql.legacy.timeParserPolicy\"", @@ -965,7 +965,7 @@ struct<> -- !query output org.apache.spark.SparkUpgradeException { - "errorClass" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION.DATETIME_PATTERN_RECOGNITION", + "condition" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION.DATETIME_PATTERN_RECOGNITION", "sqlState" : "42K0B", "messageParameters" : { "config" : "\"spark.sql.legacy.timeParserPolicy\"", @@ -1014,7 +1014,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", + "condition" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", "sqlState" : "22023", "messageParameters" : { "functionName" : "`timestampadd`", @@ -1038,7 +1038,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", + "condition" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", "sqlState" : "22023", "messageParameters" : { "functionName" : "`timestampadd`", @@ -1094,7 +1094,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", + "condition" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", "sqlState" : "22023", "messageParameters" : { "functionName" : "`timestampdiff`", @@ -1118,7 +1118,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", + "condition" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", "sqlState" : "22023", "messageParameters" : { "functionName" : "`timestampdiff`", @@ -1174,7 +1174,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", + "condition" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", "sqlState" : "22023", "messageParameters" : { "functionName" : "`timediff`", @@ -1198,7 +1198,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", + "condition" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", "sqlState" : "22023", "messageParameters" : { "functionName" : "`timediff`", diff --git a/sql/core/src/test/resources/sql-tests/results/nonansi/try_aggregates.sql.out b/sql/core/src/test/resources/sql-tests/results/nonansi/try_aggregates.sql.out index df1fe996781a..dca7773ba2ff 100644 --- a/sql/core/src/test/resources/sql-tests/results/nonansi/try_aggregates.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/nonansi/try_aggregates.sql.out @@ -118,7 +118,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "INTERVAL_DIVIDED_BY_ZERO", + "condition" : "INTERVAL_DIVIDED_BY_ZERO", "sqlState" : "22012", "queryContext" : [ { "objectType" : "", @@ -137,7 +137,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "INTERVAL_DIVIDED_BY_ZERO", + "condition" : "INTERVAL_DIVIDED_BY_ZERO", "sqlState" : "22012", "queryContext" : [ { "objectType" : "", @@ -268,7 +268,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "INTERVAL_DIVIDED_BY_ZERO", + "condition" : "INTERVAL_DIVIDED_BY_ZERO", "sqlState" : "22012", "queryContext" : [ { "objectType" : "", @@ -287,7 +287,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "INTERVAL_DIVIDED_BY_ZERO", + "condition" : "INTERVAL_DIVIDED_BY_ZERO", "sqlState" : "22012", "queryContext" : [ { "objectType" : "", diff --git a/sql/core/src/test/resources/sql-tests/results/nonansi/try_aggregates.sql.out.java21 b/sql/core/src/test/resources/sql-tests/results/nonansi/try_aggregates.sql.out.java21 index 7affe568234f..1f8233864569 100644 --- a/sql/core/src/test/resources/sql-tests/results/nonansi/try_aggregates.sql.out.java21 +++ b/sql/core/src/test/resources/sql-tests/results/nonansi/try_aggregates.sql.out.java21 @@ -84,7 +84,25 @@ SELECT try_sum(col / 0) FROM VALUES (5), (10), (15) AS tab(col) -- !query schema struct -- !query output +<<<<<<< HEAD:sql/core/src/test/resources/sql-tests/results/ansi/try_aggregates.sql.out.java21 +org.apache.spark.SparkArithmeticException +{ + "condition" : "DIVIDE_BY_ZERO", + "sqlState" : "22012", + "messageParameters" : { + "config" : "\"spark.sql.ansi.enabled\"" + }, + "queryContext" : [ { + "objectType" : "", + "objectName" : "", + "startIndex" : 16, + "stopIndex" : 22, + "fragment" : "col / 0" + } ] +} +======= NULL +>>>>>>> origin/master:sql/core/src/test/resources/sql-tests/results/nonansi/try_aggregates.sql.out.java21 -- !query @@ -92,7 +110,25 @@ SELECT try_sum(col / 0) FROM VALUES (5.0), (10.0), (15.0) AS tab(col) -- !query schema struct -- !query output +<<<<<<< HEAD:sql/core/src/test/resources/sql-tests/results/ansi/try_aggregates.sql.out.java21 +org.apache.spark.SparkArithmeticException +{ + "condition" : "DIVIDE_BY_ZERO", + "sqlState" : "22012", + "messageParameters" : { + "config" : "\"spark.sql.ansi.enabled\"" + }, + "queryContext" : [ { + "objectType" : "", + "objectName" : "", + "startIndex" : 16, + "stopIndex" : 22, + "fragment" : "col / 0" + } ] +} +======= NULL +>>>>>>> origin/master:sql/core/src/test/resources/sql-tests/results/nonansi/try_aggregates.sql.out.java21 -- !query @@ -100,7 +136,25 @@ SELECT try_sum(col / 0) FROM VALUES (NULL), (10), (15) AS tab(col) -- !query schema struct -- !query output +<<<<<<< HEAD:sql/core/src/test/resources/sql-tests/results/ansi/try_aggregates.sql.out.java21 +org.apache.spark.SparkArithmeticException +{ + "condition" : "DIVIDE_BY_ZERO", + "sqlState" : "22012", + "messageParameters" : { + "config" : "\"spark.sql.ansi.enabled\"" + }, + "queryContext" : [ { + "objectType" : "", + "objectName" : "", + "startIndex" : 16, + "stopIndex" : 22, + "fragment" : "col / 0" + } ] +} +======= NULL +>>>>>>> origin/master:sql/core/src/test/resources/sql-tests/results/nonansi/try_aggregates.sql.out.java21 -- !query @@ -108,7 +162,27 @@ SELECT try_sum(col + 1L) FROM VALUES (9223372036854775807L), (1L) AS tab(col) -- !query schema struct -- !query output +<<<<<<< HEAD:sql/core/src/test/resources/sql-tests/results/ansi/try_aggregates.sql.out.java21 +org.apache.spark.SparkArithmeticException +{ + "condition" : "ARITHMETIC_OVERFLOW", + "sqlState" : "22003", + "messageParameters" : { + "alternative" : " Use 'try_add' to tolerate overflow and return NULL instead.", + "config" : "\"spark.sql.ansi.enabled\"", + "message" : "long overflow" + }, + "queryContext" : [ { + "objectType" : "", + "objectName" : "", + "startIndex" : 16, + "stopIndex" : 23, + "fragment" : "col + 1L" + } ] +} +======= -9223372036854775806 +>>>>>>> origin/master:sql/core/src/test/resources/sql-tests/results/nonansi/try_aggregates.sql.out.java21 -- !query @@ -118,7 +192,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "INTERVAL_DIVIDED_BY_ZERO", + "condition" : "INTERVAL_DIVIDED_BY_ZERO", "sqlState" : "22012", "queryContext" : [ { "objectType" : "", @@ -137,7 +211,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "INTERVAL_DIVIDED_BY_ZERO", + "condition" : "INTERVAL_DIVIDED_BY_ZERO", "sqlState" : "22012", "queryContext" : [ { "objectType" : "", @@ -234,7 +308,25 @@ SELECT try_avg(col / 0) FROM VALUES (5), (10), (15) AS tab(col) -- !query schema struct -- !query output +<<<<<<< HEAD:sql/core/src/test/resources/sql-tests/results/ansi/try_aggregates.sql.out.java21 +org.apache.spark.SparkArithmeticException +{ + "condition" : "DIVIDE_BY_ZERO", + "sqlState" : "22012", + "messageParameters" : { + "config" : "\"spark.sql.ansi.enabled\"" + }, + "queryContext" : [ { + "objectType" : "", + "objectName" : "", + "startIndex" : 16, + "stopIndex" : 22, + "fragment" : "col / 0" + } ] +} +======= NULL +>>>>>>> origin/master:sql/core/src/test/resources/sql-tests/results/nonansi/try_aggregates.sql.out.java21 -- !query @@ -242,7 +334,25 @@ SELECT try_avg(col / 0) FROM VALUES (5.0), (10.0), (15.0) AS tab(col) -- !query schema struct -- !query output +<<<<<<< HEAD:sql/core/src/test/resources/sql-tests/results/ansi/try_aggregates.sql.out.java21 +org.apache.spark.SparkArithmeticException +{ + "condition" : "DIVIDE_BY_ZERO", + "sqlState" : "22012", + "messageParameters" : { + "config" : "\"spark.sql.ansi.enabled\"" + }, + "queryContext" : [ { + "objectType" : "", + "objectName" : "", + "startIndex" : 16, + "stopIndex" : 22, + "fragment" : "col / 0" + } ] +} +======= NULL +>>>>>>> origin/master:sql/core/src/test/resources/sql-tests/results/nonansi/try_aggregates.sql.out.java21 -- !query @@ -250,7 +360,25 @@ SELECT try_avg(col / 0) FROM VALUES (NULL), (10), (15) AS tab(col) -- !query schema struct -- !query output +<<<<<<< HEAD:sql/core/src/test/resources/sql-tests/results/ansi/try_aggregates.sql.out.java21 +org.apache.spark.SparkArithmeticException +{ + "condition" : "DIVIDE_BY_ZERO", + "sqlState" : "22012", + "messageParameters" : { + "config" : "\"spark.sql.ansi.enabled\"" + }, + "queryContext" : [ { + "objectType" : "", + "objectName" : "", + "startIndex" : 16, + "stopIndex" : 22, + "fragment" : "col / 0" + } ] +} +======= NULL +>>>>>>> origin/master:sql/core/src/test/resources/sql-tests/results/nonansi/try_aggregates.sql.out.java21 -- !query @@ -258,7 +386,27 @@ SELECT try_avg(col + 1L) FROM VALUES (9223372036854775807L), (1L) AS tab(col) -- !query schema struct -- !query output +<<<<<<< HEAD:sql/core/src/test/resources/sql-tests/results/ansi/try_aggregates.sql.out.java21 +org.apache.spark.SparkArithmeticException +{ + "condition" : "ARITHMETIC_OVERFLOW", + "sqlState" : "22003", + "messageParameters" : { + "alternative" : " Use 'try_add' to tolerate overflow and return NULL instead.", + "config" : "\"spark.sql.ansi.enabled\"", + "message" : "long overflow" + }, + "queryContext" : [ { + "objectType" : "", + "objectName" : "", + "startIndex" : 16, + "stopIndex" : 23, + "fragment" : "col + 1L" + } ] +} +======= -4.611686018427388E18 +>>>>>>> origin/master:sql/core/src/test/resources/sql-tests/results/nonansi/try_aggregates.sql.out.java21 -- !query @@ -268,7 +416,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "INTERVAL_DIVIDED_BY_ZERO", + "condition" : "INTERVAL_DIVIDED_BY_ZERO", "sqlState" : "22012", "queryContext" : [ { "objectType" : "", @@ -287,7 +435,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "INTERVAL_DIVIDED_BY_ZERO", + "condition" : "INTERVAL_DIVIDED_BY_ZERO", "sqlState" : "22012", "queryContext" : [ { "objectType" : "", diff --git a/sql/core/src/test/resources/sql-tests/results/nonansi/try_arithmetic.sql.out b/sql/core/src/test/resources/sql-tests/results/nonansi/try_arithmetic.sql.out index b12680c2a675..b92a2ba2312c 100644 --- a/sql/core/src/test/resources/sql-tests/results/nonansi/try_arithmetic.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/nonansi/try_arithmetic.sql.out @@ -182,7 +182,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"INTERVAL '2' YEAR\"", diff --git a/sql/core/src/test/resources/sql-tests/results/nonansi/try_arithmetic.sql.out.java21 b/sql/core/src/test/resources/sql-tests/results/nonansi/try_arithmetic.sql.out.java21 index 002a0dfcf37e..32855840543d 100644 --- a/sql/core/src/test/resources/sql-tests/results/nonansi/try_arithmetic.sql.out.java21 +++ b/sql/core/src/test/resources/sql-tests/results/nonansi/try_arithmetic.sql.out.java21 @@ -182,7 +182,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"INTERVAL '2' YEAR\"", diff --git a/sql/core/src/test/resources/sql-tests/results/nonansi/try_datetime_functions.sql.out b/sql/core/src/test/resources/sql-tests/results/nonansi/try_datetime_functions.sql.out index 75a6f15bd363..a2383e7081d3 100644 --- a/sql/core/src/test/resources/sql-tests/results/nonansi/try_datetime_functions.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/nonansi/try_datetime_functions.sql.out @@ -46,7 +46,7 @@ struct<> -- !query output org.apache.spark.SparkUpgradeException { - "errorClass" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION.DATETIME_PATTERN_RECOGNITION", + "condition" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION.DATETIME_PATTERN_RECOGNITION", "sqlState" : "42K0B", "messageParameters" : { "config" : "\"spark.sql.legacy.timeParserPolicy\"", diff --git a/sql/core/src/test/resources/sql-tests/results/nonansi/try_element_at.sql.out b/sql/core/src/test/resources/sql-tests/results/nonansi/try_element_at.sql.out index 0437f9d6dd9e..ea0c68680c18 100644 --- a/sql/core/src/test/resources/sql-tests/results/nonansi/try_element_at.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/nonansi/try_element_at.sql.out @@ -6,7 +6,7 @@ struct<> -- !query output org.apache.spark.SparkRuntimeException { - "errorClass" : "INVALID_INDEX_OF_ZERO", + "condition" : "INVALID_INDEX_OF_ZERO", "sqlState" : "22003" } diff --git a/sql/core/src/test/resources/sql-tests/results/operators.sql.out b/sql/core/src/test/resources/sql-tests/results/operators.sql.out index 356e5eca5feb..cf0a3a9a35f3 100644 --- a/sql/core/src/test/resources/sql-tests/results/operators.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/operators.sql.out @@ -134,7 +134,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "DIVIDE_BY_ZERO", + "condition" : "DIVIDE_BY_ZERO", "sqlState" : "22012", "messageParameters" : { "config" : "\"spark.sql.ansi.enabled\"" @@ -180,7 +180,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "DIVIDE_BY_ZERO", + "condition" : "DIVIDE_BY_ZERO", "sqlState" : "22012", "messageParameters" : { "config" : "\"spark.sql.ansi.enabled\"" @@ -226,7 +226,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "DIVIDE_BY_ZERO", + "condition" : "DIVIDE_BY_ZERO", "sqlState" : "22012", "messageParameters" : { "config" : "\"spark.sql.ansi.enabled\"" @@ -496,7 +496,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "DIVIDE_BY_ZERO", + "condition" : "DIVIDE_BY_ZERO", "sqlState" : "22012", "messageParameters" : { "config" : "\"spark.sql.ansi.enabled\"" @@ -566,7 +566,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "DIVIDE_BY_ZERO", + "condition" : "DIVIDE_BY_ZERO", "sqlState" : "22012", "messageParameters" : { "config" : "\"spark.sql.ansi.enabled\"" @@ -588,7 +588,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "DIVIDE_BY_ZERO", + "condition" : "DIVIDE_BY_ZERO", "sqlState" : "22012", "messageParameters" : { "config" : "\"spark.sql.ansi.enabled\"" diff --git a/sql/core/src/test/resources/sql-tests/results/order-by-all.sql.out b/sql/core/src/test/resources/sql-tests/results/order-by-all.sql.out index b0fcfbcb633a..a75c27702a31 100644 --- a/sql/core/src/test/resources/sql-tests/results/order-by-all.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/order-by-all.sql.out @@ -186,7 +186,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`all`", diff --git a/sql/core/src/test/resources/sql-tests/results/order-by-ordinal.sql.out b/sql/core/src/test/resources/sql-tests/results/order-by-ordinal.sql.out index f975715c12bc..50e049b87c20 100644 --- a/sql/core/src/test/resources/sql-tests/results/order-by-ordinal.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/order-by-ordinal.sql.out @@ -73,7 +73,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "ORDER_BY_POS_OUT_OF_RANGE", + "condition" : "ORDER_BY_POS_OUT_OF_RANGE", "sqlState" : "42805", "messageParameters" : { "index" : "0", @@ -96,7 +96,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "ORDER_BY_POS_OUT_OF_RANGE", + "condition" : "ORDER_BY_POS_OUT_OF_RANGE", "sqlState" : "42805", "messageParameters" : { "index" : "-1", @@ -119,7 +119,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "ORDER_BY_POS_OUT_OF_RANGE", + "condition" : "ORDER_BY_POS_OUT_OF_RANGE", "sqlState" : "42805", "messageParameters" : { "index" : "3", diff --git a/sql/core/src/test/resources/sql-tests/results/percentiles.sql.out b/sql/core/src/test/resources/sql-tests/results/percentiles.sql.out index 55aaa8ee7378..5f052c8ff22c 100644 --- a/sql/core/src/test/resources/sql-tests/results/percentiles.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/percentiles.sql.out @@ -122,7 +122,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_SQL_SYNTAX.FUNCTION_WITH_UNSUPPORTED_SYNTAX", + "condition" : "INVALID_SQL_SYNTAX.FUNCTION_WITH_UNSUPPORTED_SYNTAX", "sqlState" : "42000", "messageParameters" : { "prettyName" : "`round`", @@ -147,7 +147,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_SQL_SYNTAX.FUNCTION_WITH_UNSUPPORTED_SYNTAX", + "condition" : "INVALID_SQL_SYNTAX.FUNCTION_WITH_UNSUPPORTED_SYNTAX", "sqlState" : "42000", "messageParameters" : { "prettyName" : "`round`", @@ -172,7 +172,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_SQL_SYNTAX.FUNCTION_WITH_UNSUPPORTED_SYNTAX", + "condition" : "INVALID_SQL_SYNTAX.FUNCTION_WITH_UNSUPPORTED_SYNTAX", "sqlState" : "42000", "messageParameters" : { "prettyName" : "`percentile`", @@ -197,7 +197,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_SQL_SYNTAX.FUNCTION_WITH_UNSUPPORTED_SYNTAX", + "condition" : "INVALID_SQL_SYNTAX.FUNCTION_WITH_UNSUPPORTED_SYNTAX", "sqlState" : "42000", "messageParameters" : { "prettyName" : "`percentile`", @@ -222,7 +222,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_WITHIN_GROUP_EXPRESSION.DISTINCT_UNSUPPORTED", + "condition" : "INVALID_WITHIN_GROUP_EXPRESSION.DISTINCT_UNSUPPORTED", "sqlState" : "42K0K", "messageParameters" : { "funcName" : "`percentile_cont`" @@ -246,7 +246,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_WITHIN_GROUP_EXPRESSION.DISTINCT_UNSUPPORTED", + "condition" : "INVALID_WITHIN_GROUP_EXPRESSION.DISTINCT_UNSUPPORTED", "sqlState" : "42K0K", "messageParameters" : { "funcName" : "`percentile_cont`" @@ -270,7 +270,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", + "condition" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", "sqlState" : "42605", "messageParameters" : { "actualNum" : "0", @@ -297,7 +297,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", + "condition" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", "sqlState" : "42605", "messageParameters" : { "actualNum" : "0", @@ -324,7 +324,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_WITHIN_GROUP_EXPRESSION.WITHIN_GROUP_MISSING", + "condition" : "INVALID_WITHIN_GROUP_EXPRESSION.WITHIN_GROUP_MISSING", "sqlState" : "42K0K", "messageParameters" : { "funcName" : "`percentile_cont`" @@ -348,7 +348,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_WITHIN_GROUP_EXPRESSION.WITHIN_GROUP_MISSING", + "condition" : "INVALID_WITHIN_GROUP_EXPRESSION.WITHIN_GROUP_MISSING", "sqlState" : "42K0K", "messageParameters" : { "funcName" : "`percentile_cont`" @@ -372,7 +372,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_WITHIN_GROUP_EXPRESSION.WRONG_NUM_ORDERINGS", + "condition" : "INVALID_WITHIN_GROUP_EXPRESSION.WRONG_NUM_ORDERINGS", "sqlState" : "42K0K", "messageParameters" : { "actualNum" : "2", @@ -398,7 +398,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", + "condition" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", "sqlState" : "42605", "messageParameters" : { "actualNum" : "2", @@ -482,7 +482,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INVALID_WINDOW_SPEC_FOR_AGGREGATION_FUNC", + "condition" : "INVALID_WINDOW_SPEC_FOR_AGGREGATION_FUNC", "sqlState" : "42601", "messageParameters" : { "aggFunc" : "\"percentile_cont(salary, 0.25)\"" @@ -511,7 +511,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INVALID_WINDOW_SPEC_FOR_AGGREGATION_FUNC", + "condition" : "INVALID_WINDOW_SPEC_FOR_AGGREGATION_FUNC", "sqlState" : "42601", "messageParameters" : { "aggFunc" : "\"percentile_disc(salary, 0.25)\"" @@ -539,7 +539,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INVALID_WINDOW_SPEC_FOR_AGGREGATION_FUNC", + "condition" : "INVALID_WINDOW_SPEC_FOR_AGGREGATION_FUNC", "sqlState" : "42601", "messageParameters" : { "aggFunc" : "\"median(salary)\"" @@ -568,7 +568,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INVALID_WINDOW_SPEC_FOR_AGGREGATION_FUNC", + "condition" : "INVALID_WINDOW_SPEC_FOR_AGGREGATION_FUNC", "sqlState" : "42601", "messageParameters" : { "aggFunc" : "\"percentile_cont(salary, 0.25)\"" @@ -597,7 +597,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INVALID_WINDOW_SPEC_FOR_AGGREGATION_FUNC", + "condition" : "INVALID_WINDOW_SPEC_FOR_AGGREGATION_FUNC", "sqlState" : "42601", "messageParameters" : { "aggFunc" : "\"percentile_disc(salary, 0.25)\"" @@ -625,7 +625,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INVALID_WINDOW_SPEC_FOR_AGGREGATION_FUNC", + "condition" : "INVALID_WINDOW_SPEC_FOR_AGGREGATION_FUNC", "sqlState" : "42601", "messageParameters" : { "aggFunc" : "\"median(salary)\"" @@ -718,7 +718,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INVALID_WINDOW_SPEC_FOR_AGGREGATION_FUNC", + "condition" : "INVALID_WINDOW_SPEC_FOR_AGGREGATION_FUNC", "sqlState" : "42601", "messageParameters" : { "aggFunc" : "\"percentile_cont(salary, 0.25)\"" @@ -748,7 +748,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INVALID_WINDOW_SPEC_FOR_AGGREGATION_FUNC", + "condition" : "INVALID_WINDOW_SPEC_FOR_AGGREGATION_FUNC", "sqlState" : "42601", "messageParameters" : { "aggFunc" : "\"percentile_disc(salary, 0.25)\"" @@ -777,7 +777,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INVALID_WINDOW_SPEC_FOR_AGGREGATION_FUNC", + "condition" : "INVALID_WINDOW_SPEC_FOR_AGGREGATION_FUNC", "sqlState" : "42601", "messageParameters" : { "aggFunc" : "\"median(salary)\"" @@ -807,7 +807,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INVALID_WINDOW_SPEC_FOR_AGGREGATION_FUNC", + "condition" : "INVALID_WINDOW_SPEC_FOR_AGGREGATION_FUNC", "sqlState" : "42601", "messageParameters" : { "aggFunc" : "\"percentile_cont(salary, 0.25)\"" @@ -837,7 +837,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INVALID_WINDOW_SPEC_FOR_AGGREGATION_FUNC", + "condition" : "INVALID_WINDOW_SPEC_FOR_AGGREGATION_FUNC", "sqlState" : "42601", "messageParameters" : { "aggFunc" : "\"percentile_disc(salary, 0.25)\"" @@ -866,7 +866,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INVALID_WINDOW_SPEC_FOR_AGGREGATION_FUNC", + "condition" : "INVALID_WINDOW_SPEC_FOR_AGGREGATION_FUNC", "sqlState" : "42601", "messageParameters" : { "aggFunc" : "\"median(salary)\"" @@ -1164,7 +1164,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.NON_FOLDABLE_INPUT", + "condition" : "DATATYPE_MISMATCH.NON_FOLDABLE_INPUT", "sqlState" : "42K09", "messageParameters" : { "inputExpr" : "\"b\"", diff --git a/sql/core/src/test/resources/sql-tests/results/pipe-operators.sql.out b/sql/core/src/test/resources/sql-tests/results/pipe-operators.sql.out index fe1a263c0644..8473fe0cec8c 100644 --- a/sql/core/src/test/resources/sql-tests/results/pipe-operators.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/pipe-operators.sql.out @@ -353,7 +353,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'from'", @@ -587,7 +587,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "PIPE_OPERATOR_CONTAINS_AGGREGATE_FUNCTION", + "condition" : "PIPE_OPERATOR_CONTAINS_AGGREGATE_FUNCTION", "sqlState" : "0A000", "messageParameters" : { "clause" : "SELECT", @@ -611,7 +611,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "PIPE_OPERATOR_CONTAINS_AGGREGATE_FUNCTION", + "condition" : "PIPE_OPERATOR_CONTAINS_AGGREGATE_FUNCTION", "sqlState" : "0A000", "messageParameters" : { "clause" : "SELECT", @@ -758,7 +758,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "PIPE_OPERATOR_CONTAINS_AGGREGATE_FUNCTION", + "condition" : "PIPE_OPERATOR_CONTAINS_AGGREGATE_FUNCTION", "sqlState" : "0A000", "messageParameters" : { "clause" : "EXTEND", @@ -782,7 +782,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'as'", @@ -799,7 +799,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_USAGE_OF_STAR_OR_REGEX", + "condition" : "INVALID_USAGE_OF_STAR_OR_REGEX", "sqlState" : "42000", "messageParameters" : { "elem" : "'*'", @@ -965,7 +965,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`z`", @@ -989,7 +989,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'as'", @@ -1006,7 +1006,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0035", + "condition" : "_LEGACY_ERROR_TEMP_0035", "messageParameters" : { "message" : "SQL pipe syntax |> SET operator with multi-part assignment key (only single-part keys are allowed)" }, @@ -1077,7 +1077,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`z`", @@ -1101,7 +1101,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'.'", @@ -1118,7 +1118,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`col.i1`", @@ -1142,7 +1142,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "EXCEPT_OVERLAPPING_COLUMNS", + "condition" : "EXCEPT_OVERLAPPING_COLUMNS", "sqlState" : "42702", "messageParameters" : { "columns" : "z, y, z" @@ -1219,7 +1219,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "','", @@ -1236,7 +1236,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'1'", @@ -1253,7 +1253,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_IDENTIFIER", + "condition" : "INVALID_IDENTIFIER", "sqlState" : "42602", "messageParameters" : { "ident" : "u-v" @@ -1269,7 +1269,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'@'", @@ -1286,7 +1286,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'#'", @@ -1395,7 +1395,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INVALID_WHERE_CONDITION", + "condition" : "INVALID_WHERE_CONDITION", "sqlState" : "42903", "messageParameters" : { "condition" : "\"(sum(x) = 1)\"", @@ -1419,7 +1419,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INVALID_WHERE_CONDITION", + "condition" : "INVALID_WHERE_CONDITION", "sqlState" : "42903", "messageParameters" : { "condition" : "\"((y = abc) OR ((length(y) + sum(x)) = 1))\"", @@ -1443,7 +1443,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "_LEGACY_ERROR_TEMP_1034", + "condition" : "_LEGACY_ERROR_TEMP_1034", "messageParameters" : { "clauseName" : "WHERE" } @@ -1459,7 +1459,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "NOT_ALLOWED_IN_PIPE_OPERATOR_WHERE.WINDOW_CLAUSE", + "condition" : "NOT_ALLOWED_IN_PIPE_OPERATOR_WHERE.WINDOW_CLAUSE", "sqlState" : "42601", "queryContext" : [ { "objectType" : "", @@ -1478,7 +1478,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "_LEGACY_ERROR_TEMP_1034", + "condition" : "_LEGACY_ERROR_TEMP_1034", "messageParameters" : { "clauseName" : "WHERE" } @@ -1494,7 +1494,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`y`", @@ -1520,7 +1520,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`y`", @@ -1546,7 +1546,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`y`", @@ -1572,7 +1572,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`y`", @@ -1596,7 +1596,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`y`", @@ -1732,7 +1732,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`year`", @@ -1759,7 +1759,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "NON_LITERAL_PIVOT_VALUES", + "condition" : "NON_LITERAL_PIVOT_VALUES", "sqlState" : "42K08", "messageParameters" : { "expression" : "\"course\"" @@ -1782,7 +1782,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "NOT_ALLOWED_IN_FROM.UNPIVOT_WITH_PIVOT", + "condition" : "NOT_ALLOWED_IN_FROM.UNPIVOT_WITH_PIVOT", "sqlState" : "42601", "queryContext" : [ { "objectType" : "", @@ -1809,7 +1809,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "NOT_ALLOWED_IN_FROM.UNPIVOT_WITH_PIVOT", + "condition" : "NOT_ALLOWED_IN_FROM.UNPIVOT_WITH_PIVOT", "sqlState" : "42601", "queryContext" : [ { "objectType" : "", @@ -1837,7 +1837,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'pivot'", @@ -1864,7 +1864,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'unpivot'", @@ -1923,7 +1923,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0014", + "condition" : "_LEGACY_ERROR_TEMP_0014", "queryContext" : [ { "objectType" : "", "objectName" : "", @@ -1942,7 +1942,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0064", + "condition" : "_LEGACY_ERROR_TEMP_0064", "messageParameters" : { "msg" : "Sampling fraction (-1.0) must be on interval [0, 1]" }, @@ -1964,7 +1964,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INVALID_LIMIT_LIKE_EXPRESSION.IS_NEGATIVE", + "condition" : "INVALID_LIMIT_LIKE_EXPRESSION.IS_NEGATIVE", "sqlState" : "42K0E", "messageParameters" : { "expr" : "\"-5\"", @@ -1989,7 +1989,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INVALID_LIMIT_LIKE_EXPRESSION.IS_UNFOLDABLE", + "condition" : "INVALID_LIMIT_LIKE_EXPRESSION.IS_UNFOLDABLE", "sqlState" : "42K0E", "messageParameters" : { "expr" : "\"x\"", @@ -2013,7 +2013,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0064", + "condition" : "_LEGACY_ERROR_TEMP_0064", "messageParameters" : { "msg" : "Sampling fraction (2.0) must be on interval [0, 1]" }, @@ -2035,7 +2035,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0015", + "condition" : "_LEGACY_ERROR_TEMP_0015", "messageParameters" : { "msg" : "byteLengthLiteral" }, @@ -2057,7 +2057,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0016", + "condition" : "_LEGACY_ERROR_TEMP_0016", "messageParameters" : { "bytesStr" : "200" }, @@ -2350,7 +2350,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'inner'", @@ -2368,7 +2368,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`join_test_t1`.`a`", @@ -2392,7 +2392,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'jt'", @@ -2555,7 +2555,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "NUM_COLUMNS_MISMATCH", + "condition" : "NUM_COLUMNS_MISMATCH", "sqlState" : "42826", "messageParameters" : { "firstNumColumns" : "1", @@ -2581,7 +2581,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "second", @@ -2722,7 +2722,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'order'", @@ -2740,7 +2740,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`x`", @@ -2765,7 +2765,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`x`", @@ -2789,7 +2789,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "MULTIPLE_QUERY_RESULT_CLAUSES_WITH_PIPE_OPERATORS", + "condition" : "MULTIPLE_QUERY_RESULT_CLAUSES_WITH_PIPE_OPERATORS", "sqlState" : "42000", "messageParameters" : { "clause1" : "ORDER BY", @@ -2813,7 +2813,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "UNSUPPORTED_FEATURE.COMBINATION_QUERY_RESULT_CLAUSES", + "condition" : "UNSUPPORTED_FEATURE.COMBINATION_QUERY_RESULT_CLAUSES", "sqlState" : "0A000", "queryContext" : [ { "objectType" : "", @@ -3060,7 +3060,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "PIPE_OPERATOR_AGGREGATE_EXPRESSION_CONTAINS_NO_AGGREGATE_FUNCTION", + "condition" : "PIPE_OPERATOR_AGGREGATE_EXPRESSION_CONTAINS_NO_AGGREGATE_FUNCTION", "sqlState" : "0A000", "messageParameters" : { "expr" : "a#x" @@ -3083,7 +3083,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "UNSUPPORTED_FEATURE.PIPE_OPERATOR_AGGREGATE_UNSUPPORTED_CASE", + "condition" : "UNSUPPORTED_FEATURE.PIPE_OPERATOR_AGGREGATE_UNSUPPORTED_CASE", "sqlState" : "0A000", "messageParameters" : { "case" : "GROUP BY ALL" @@ -3107,7 +3107,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "UNSUPPORTED_FEATURE.PIPE_OPERATOR_AGGREGATE_UNSUPPORTED_CASE", + "condition" : "UNSUPPORTED_FEATURE.PIPE_OPERATOR_AGGREGATE_UNSUPPORTED_CASE", "sqlState" : "0A000", "messageParameters" : { "case" : "GROUP BY ROLLUP" @@ -3131,7 +3131,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "UNSUPPORTED_FEATURE.PIPE_OPERATOR_AGGREGATE_UNSUPPORTED_CASE", + "condition" : "UNSUPPORTED_FEATURE.PIPE_OPERATOR_AGGREGATE_UNSUPPORTED_CASE", "sqlState" : "0A000", "messageParameters" : { "case" : "GROUP BY CUBE" @@ -3155,7 +3155,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "UNSUPPORTED_FEATURE.PIPE_OPERATOR_AGGREGATE_UNSUPPORTED_CASE", + "condition" : "UNSUPPORTED_FEATURE.PIPE_OPERATOR_AGGREGATE_UNSUPPORTED_CASE", "sqlState" : "0A000", "messageParameters" : { "case" : "GROUPING SETS" @@ -3180,7 +3180,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "UNSUPPORTED_FEATURE.PIPE_OPERATOR_AGGREGATE_UNSUPPORTED_CASE", + "condition" : "UNSUPPORTED_FEATURE.PIPE_OPERATOR_AGGREGATE_UNSUPPORTED_CASE", "sqlState" : "0A000", "messageParameters" : { "case" : "GROUPING" @@ -3205,7 +3205,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "UNSUPPORTED_FEATURE.PIPE_OPERATOR_AGGREGATE_UNSUPPORTED_CASE", + "condition" : "UNSUPPORTED_FEATURE.PIPE_OPERATOR_AGGREGATE_UNSUPPORTED_CASE", "sqlState" : "0A000", "messageParameters" : { "case" : "GROUPING_ID" @@ -3228,7 +3228,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "')'", @@ -3245,7 +3245,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "PIPE_OPERATOR_AGGREGATE_EXPRESSION_CONTAINS_NO_AGGREGATE_FUNCTION", + "condition" : "PIPE_OPERATOR_AGGREGATE_EXPRESSION_CONTAINS_NO_AGGREGATE_FUNCTION", "sqlState" : "0A000", "messageParameters" : { "expr" : "a#x" @@ -3268,7 +3268,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "PIPE_OPERATOR_CONTAINS_AGGREGATE_FUNCTION", + "condition" : "PIPE_OPERATOR_CONTAINS_AGGREGATE_FUNCTION", "sqlState" : "0A000", "messageParameters" : { "clause" : "SELECT", @@ -3292,7 +3292,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0035", + "condition" : "_LEGACY_ERROR_TEMP_0035", "messageParameters" : { "message" : "The AGGREGATE clause requires a list of aggregate expressions or a list of grouping expressions, or both" }, @@ -3314,7 +3314,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`group`", @@ -3338,7 +3338,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'group'", @@ -3355,7 +3355,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "UNSUPPORTED_FEATURE.PIPE_OPERATOR_AGGREGATE_UNSUPPORTED_CASE", + "condition" : "UNSUPPORTED_FEATURE.PIPE_OPERATOR_AGGREGATE_UNSUPPORTED_CASE", "sqlState" : "0A000", "messageParameters" : { "case" : "window functions; please update the query to move the window functions to a subsequent |> SELECT operator instead" @@ -3380,7 +3380,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "AMBIGUOUS_REFERENCE", + "condition" : "AMBIGUOUS_REFERENCE", "sqlState" : "42704", "messageParameters" : { "name" : "`x`", @@ -3497,7 +3497,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "MISSING_WINDOW_SPECIFICATION", + "condition" : "MISSING_WINDOW_SPECIFICATION", "sqlState" : "42P20", "messageParameters" : { "docroot" : "https://spark.apache.org/docs/latest", @@ -3562,7 +3562,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "MISSING_WINDOW_SPECIFICATION", + "condition" : "MISSING_WINDOW_SPECIFICATION", "sqlState" : "42P20", "messageParameters" : { "docroot" : "https://spark.apache.org/docs/latest", @@ -3581,7 +3581,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "MISSING_WINDOW_SPECIFICATION", + "condition" : "MISSING_WINDOW_SPECIFICATION", "sqlState" : "42P20", "messageParameters" : { "docroot" : "https://spark.apache.org/docs/latest", @@ -3600,7 +3600,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "MISSING_WINDOW_SPECIFICATION", + "condition" : "MISSING_WINDOW_SPECIFICATION", "sqlState" : "42P20", "messageParameters" : { "docroot" : "https://spark.apache.org/docs/latest", @@ -3619,7 +3619,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "MISSING_WINDOW_SPECIFICATION", + "condition" : "MISSING_WINDOW_SPECIFICATION", "sqlState" : "42P20", "messageParameters" : { "docroot" : "https://spark.apache.org/docs/latest", diff --git a/sql/core/src/test/resources/sql-tests/results/pivot.sql.out b/sql/core/src/test/resources/sql-tests/results/pivot.sql.out index aaafb19cb979..f604d3ea75b4 100644 --- a/sql/core/src/test/resources/sql-tests/results/pivot.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/pivot.sql.out @@ -200,7 +200,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "_LEGACY_ERROR_TEMP_1006", + "condition" : "_LEGACY_ERROR_TEMP_1006", "messageParameters" : { "sql" : "coursesales.earnings" } @@ -220,7 +220,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "_LEGACY_ERROR_TEMP_1006", + "condition" : "_LEGACY_ERROR_TEMP_1006", "messageParameters" : { "sql" : "__auto_generated_subquery_name.year" } @@ -240,7 +240,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`year`", @@ -284,7 +284,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "NESTED_AGGREGATE_FUNCTION", + "condition" : "NESTED_AGGREGATE_FUNCTION", "sqlState" : "42607", "queryContext" : [ { "objectType" : "", @@ -345,7 +345,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "PIVOT_VALUE_DATA_TYPE_MISMATCH", + "condition" : "PIVOT_VALUE_DATA_TYPE_MISMATCH", "sqlState" : "42K09", "messageParameters" : { "pivotType" : "struct", @@ -366,7 +366,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`s`", @@ -393,7 +393,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "NON_LITERAL_PIVOT_VALUES", + "condition" : "NON_LITERAL_PIVOT_VALUES", "sqlState" : "42K08", "messageParameters" : { "expression" : "\"course\"" @@ -518,7 +518,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "INCOMPARABLE_PIVOT_COLUMN", + "condition" : "INCOMPARABLE_PIVOT_COLUMN", "sqlState" : "42818", "messageParameters" : { "columnName" : "`m`" @@ -541,7 +541,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "INCOMPARABLE_PIVOT_COLUMN", + "condition" : "INCOMPARABLE_PIVOT_COLUMN", "sqlState" : "42818", "messageParameters" : { "columnName" : "`named_struct('course', __auto_generated_subquery_name`.`course, 'm', __auto_generated_subquery_name`.`m)`" diff --git a/sql/core/src/test/resources/sql-tests/results/postgreSQL/aggregates_part1.sql.out b/sql/core/src/test/resources/sql-tests/results/postgreSQL/aggregates_part1.sql.out index 42cf942add48..c65d72ce19f1 100644 --- a/sql/core/src/test/resources/sql-tests/results/postgreSQL/aggregates_part1.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/postgreSQL/aggregates_part1.sql.out @@ -485,7 +485,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "UNSUPPORTED_SUBQUERY_EXPRESSION_CATEGORY.AGGREGATE_FUNCTION_MIXED_OUTER_LOCAL_REFERENCES", + "condition" : "UNSUPPORTED_SUBQUERY_EXPRESSION_CATEGORY.AGGREGATE_FUNCTION_MIXED_OUTER_LOCAL_REFERENCES", "sqlState" : "0A000", "messageParameters" : { "function" : "sum(DISTINCT (outer(a.four) + b.four))" @@ -509,7 +509,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`o`.`unique1`", diff --git a/sql/core/src/test/resources/sql-tests/results/postgreSQL/aggregates_part3.sql.out b/sql/core/src/test/resources/sql-tests/results/postgreSQL/aggregates_part3.sql.out index 038bcc0283a6..8e0685b8cb05 100644 --- a/sql/core/src/test/resources/sql-tests/results/postgreSQL/aggregates_part3.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/postgreSQL/aggregates_part3.sql.out @@ -6,7 +6,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "NESTED_AGGREGATE_FUNCTION", + "condition" : "NESTED_AGGREGATE_FUNCTION", "sqlState" : "42607", "queryContext" : [ { "objectType" : "", diff --git a/sql/core/src/test/resources/sql-tests/results/postgreSQL/boolean.sql.out b/sql/core/src/test/resources/sql-tests/results/postgreSQL/boolean.sql.out index 052e7b4f2522..fd6fac196ad8 100644 --- a/sql/core/src/test/resources/sql-tests/results/postgreSQL/boolean.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/postgreSQL/boolean.sql.out @@ -54,7 +54,7 @@ struct<> -- !query output org.apache.spark.SparkRuntimeException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'test'", @@ -86,7 +86,7 @@ struct<> -- !query output org.apache.spark.SparkRuntimeException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'foo'", @@ -126,7 +126,7 @@ struct<> -- !query output org.apache.spark.SparkRuntimeException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'yeah'", @@ -166,7 +166,7 @@ struct<> -- !query output org.apache.spark.SparkRuntimeException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'nay'", @@ -190,7 +190,7 @@ struct<> -- !query output org.apache.spark.SparkRuntimeException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'on'", @@ -214,7 +214,7 @@ struct<> -- !query output org.apache.spark.SparkRuntimeException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'off'", @@ -238,7 +238,7 @@ struct<> -- !query output org.apache.spark.SparkRuntimeException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'of'", @@ -262,7 +262,7 @@ struct<> -- !query output org.apache.spark.SparkRuntimeException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'o'", @@ -286,7 +286,7 @@ struct<> -- !query output org.apache.spark.SparkRuntimeException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'on_'", @@ -310,7 +310,7 @@ struct<> -- !query output org.apache.spark.SparkRuntimeException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'off_'", @@ -342,7 +342,7 @@ struct<> -- !query output org.apache.spark.SparkRuntimeException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'11'", @@ -374,7 +374,7 @@ struct<> -- !query output org.apache.spark.SparkRuntimeException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'000'", @@ -398,7 +398,7 @@ struct<> -- !query output org.apache.spark.SparkRuntimeException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "''", @@ -519,7 +519,7 @@ struct<> -- !query output org.apache.spark.SparkRuntimeException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "' tru e '", @@ -543,7 +543,7 @@ struct<> -- !query output org.apache.spark.SparkRuntimeException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "''", @@ -702,7 +702,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_INLINE_TABLE.FAILED_SQL_EXPRESSION_EVALUATION", + "condition" : "INVALID_INLINE_TABLE.FAILED_SQL_EXPRESSION_EVALUATION", "sqlState" : "42000", "messageParameters" : { "sqlExpr" : "\"XXX\"" diff --git a/sql/core/src/test/resources/sql-tests/results/postgreSQL/create_view.sql.out b/sql/core/src/test/resources/sql-tests/results/postgreSQL/create_view.sql.out index 90c26b858292..2583d14b512b 100644 --- a/sql/core/src/test/resources/sql-tests/results/postgreSQL/create_view.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/postgreSQL/create_view.sql.out @@ -54,7 +54,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "MISSING_AGGREGATION", + "condition" : "MISSING_AGGREGATION", "sqlState" : "42803", "messageParameters" : { "expression" : "\"data\"", @@ -71,7 +71,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITHOUT_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITHOUT_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`FROM`" @@ -279,7 +279,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_TEMP_OBJ_REFERENCE", + "condition" : "INVALID_TEMP_OBJ_REFERENCE", "sqlState" : "42K0F", "messageParameters" : { "obj" : "VIEW", @@ -345,7 +345,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_TEMP_OBJ_REFERENCE", + "condition" : "INVALID_TEMP_OBJ_REFERENCE", "sqlState" : "42K0F", "messageParameters" : { "obj" : "VIEW", @@ -404,7 +404,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_TEMP_OBJ_REFERENCE", + "condition" : "INVALID_TEMP_OBJ_REFERENCE", "sqlState" : "42K0F", "messageParameters" : { "obj" : "VIEW", @@ -425,7 +425,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_TEMP_OBJ_REFERENCE", + "condition" : "INVALID_TEMP_OBJ_REFERENCE", "sqlState" : "42K0F", "messageParameters" : { "obj" : "VIEW", @@ -598,7 +598,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_TEMP_OBJ_REFERENCE", + "condition" : "INVALID_TEMP_OBJ_REFERENCE", "sqlState" : "42K0F", "messageParameters" : { "obj" : "VIEW", @@ -616,7 +616,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_TEMP_OBJ_REFERENCE", + "condition" : "INVALID_TEMP_OBJ_REFERENCE", "sqlState" : "42K0F", "messageParameters" : { "obj" : "VIEW", @@ -634,7 +634,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_TEMP_OBJ_REFERENCE", + "condition" : "INVALID_TEMP_OBJ_REFERENCE", "sqlState" : "42K0F", "messageParameters" : { "obj" : "VIEW", @@ -652,7 +652,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_TEMP_OBJ_REFERENCE", + "condition" : "INVALID_TEMP_OBJ_REFERENCE", "sqlState" : "42K0F", "messageParameters" : { "obj" : "VIEW", @@ -670,7 +670,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "TABLE_OR_VIEW_NOT_FOUND", + "condition" : "TABLE_OR_VIEW_NOT_FOUND", "sqlState" : "42P01", "messageParameters" : { "relationName" : "`v7_temp`" @@ -692,7 +692,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "TABLE_OR_VIEW_NOT_FOUND", + "condition" : "TABLE_OR_VIEW_NOT_FOUND", "sqlState" : "42P01", "messageParameters" : { "relationName" : "`v10_temp`" @@ -714,7 +714,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "TABLE_OR_VIEW_NOT_FOUND", + "condition" : "TABLE_OR_VIEW_NOT_FOUND", "sqlState" : "42P01", "messageParameters" : { "relationName" : "`v11_temp`" @@ -810,7 +810,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_TEMP_OBJ_REFERENCE", + "condition" : "INVALID_TEMP_OBJ_REFERENCE", "sqlState" : "42K0F", "messageParameters" : { "obj" : "VIEW", @@ -861,7 +861,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_TEMP_OBJ_REFERENCE", + "condition" : "INVALID_TEMP_OBJ_REFERENCE", "sqlState" : "42K0F", "messageParameters" : { "obj" : "VIEW", @@ -912,7 +912,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_TEMP_OBJ_REFERENCE", + "condition" : "INVALID_TEMP_OBJ_REFERENCE", "sqlState" : "42K0F", "messageParameters" : { "obj" : "VIEW", @@ -963,7 +963,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_TEMP_OBJ_REFERENCE", + "condition" : "INVALID_TEMP_OBJ_REFERENCE", "sqlState" : "42K0F", "messageParameters" : { "obj" : "VIEW", @@ -981,7 +981,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_TEMP_OBJ_REFERENCE", + "condition" : "INVALID_TEMP_OBJ_REFERENCE", "sqlState" : "42K0F", "messageParameters" : { "obj" : "VIEW", diff --git a/sql/core/src/test/resources/sql-tests/results/postgreSQL/date.sql.out b/sql/core/src/test/resources/sql-tests/results/postgreSQL/date.sql.out index d9f4301dd0e8..98b21c590245 100755 --- a/sql/core/src/test/resources/sql-tests/results/postgreSQL/date.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/postgreSQL/date.sql.out @@ -199,7 +199,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_TYPED_LITERAL", + "condition" : "INVALID_TYPED_LITERAL", "sqlState" : "42604", "messageParameters" : { "value" : "'1999 Jan 08'", @@ -222,7 +222,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_TYPED_LITERAL", + "condition" : "INVALID_TYPED_LITERAL", "sqlState" : "42604", "messageParameters" : { "value" : "'1999 08 Jan'", @@ -261,7 +261,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_TYPED_LITERAL", + "condition" : "INVALID_TYPED_LITERAL", "sqlState" : "42604", "messageParameters" : { "value" : "'1999 01 08'", @@ -284,7 +284,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_TYPED_LITERAL", + "condition" : "INVALID_TYPED_LITERAL", "sqlState" : "42604", "messageParameters" : { "value" : "'1999 08 01'", @@ -315,7 +315,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_TYPED_LITERAL", + "condition" : "INVALID_TYPED_LITERAL", "sqlState" : "42604", "messageParameters" : { "value" : "'1999 Jan 08'", @@ -338,7 +338,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_TYPED_LITERAL", + "condition" : "INVALID_TYPED_LITERAL", "sqlState" : "42604", "messageParameters" : { "value" : "'1999 08 Jan'", @@ -377,7 +377,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_TYPED_LITERAL", + "condition" : "INVALID_TYPED_LITERAL", "sqlState" : "42604", "messageParameters" : { "value" : "'1999 01 08'", @@ -400,7 +400,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_TYPED_LITERAL", + "condition" : "INVALID_TYPED_LITERAL", "sqlState" : "42604", "messageParameters" : { "value" : "'1999 08 01'", @@ -439,7 +439,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_TYPED_LITERAL", + "condition" : "INVALID_TYPED_LITERAL", "sqlState" : "42604", "messageParameters" : { "value" : "'1999 Jan 08'", @@ -462,7 +462,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_TYPED_LITERAL", + "condition" : "INVALID_TYPED_LITERAL", "sqlState" : "42604", "messageParameters" : { "value" : "'1999 08 Jan'", @@ -501,7 +501,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_TYPED_LITERAL", + "condition" : "INVALID_TYPED_LITERAL", "sqlState" : "42604", "messageParameters" : { "value" : "'1999 01 08'", @@ -524,7 +524,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_TYPED_LITERAL", + "condition" : "INVALID_TYPED_LITERAL", "sqlState" : "42604", "messageParameters" : { "value" : "'1999 08 01'", @@ -687,7 +687,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "DATETIME_FIELD_OUT_OF_BOUNDS", + "condition" : "DATETIME_FIELD_OUT_OF_BOUNDS", "sqlState" : "22023", "messageParameters" : { "ansiConfig" : "\"spark.sql.ansi.enabled\"", @@ -703,7 +703,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "DATETIME_FIELD_OUT_OF_BOUNDS", + "condition" : "DATETIME_FIELD_OUT_OF_BOUNDS", "sqlState" : "22023", "messageParameters" : { "ansiConfig" : "\"spark.sql.ansi.enabled\"", @@ -719,7 +719,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "DATETIME_FIELD_OUT_OF_BOUNDS", + "condition" : "DATETIME_FIELD_OUT_OF_BOUNDS", "sqlState" : "22023", "messageParameters" : { "ansiConfig" : "\"spark.sql.ansi.enabled\"", diff --git a/sql/core/src/test/resources/sql-tests/results/postgreSQL/float4.sql.out b/sql/core/src/test/resources/sql-tests/results/postgreSQL/float4.sql.out index 1a15610b4ded..9f5eaf98bc98 100644 --- a/sql/core/src/test/resources/sql-tests/results/postgreSQL/float4.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/postgreSQL/float4.sql.out @@ -94,7 +94,7 @@ struct<> -- !query output org.apache.spark.SparkNumberFormatException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'N A N'", @@ -118,7 +118,7 @@ struct<> -- !query output org.apache.spark.SparkNumberFormatException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'NaN x'", @@ -142,7 +142,7 @@ struct<> -- !query output org.apache.spark.SparkNumberFormatException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "' INFINITY x'", @@ -190,7 +190,7 @@ struct<> -- !query output org.apache.spark.SparkNumberFormatException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'nan'", @@ -386,7 +386,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "CAST_OVERFLOW", + "condition" : "CAST_OVERFLOW", "sqlState" : "22003", "messageParameters" : { "sourceType" : "\"FLOAT\"", @@ -411,7 +411,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "CAST_OVERFLOW", + "condition" : "CAST_OVERFLOW", "sqlState" : "22003", "messageParameters" : { "sourceType" : "\"FLOAT\"", @@ -452,7 +452,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "CAST_OVERFLOW", + "condition" : "CAST_OVERFLOW", "sqlState" : "22003", "messageParameters" : { "sourceType" : "\"FLOAT\"", diff --git a/sql/core/src/test/resources/sql-tests/results/postgreSQL/float4.sql.out.java21 b/sql/core/src/test/resources/sql-tests/results/postgreSQL/float4.sql.out.java21 index 3c2189c39963..e44345dd0f4c 100644 --- a/sql/core/src/test/resources/sql-tests/results/postgreSQL/float4.sql.out.java21 +++ b/sql/core/src/test/resources/sql-tests/results/postgreSQL/float4.sql.out.java21 @@ -94,7 +94,7 @@ struct<> -- !query output org.apache.spark.SparkNumberFormatException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'N A N'", @@ -118,7 +118,7 @@ struct<> -- !query output org.apache.spark.SparkNumberFormatException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'NaN x'", @@ -142,7 +142,7 @@ struct<> -- !query output org.apache.spark.SparkNumberFormatException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "' INFINITY x'", @@ -190,7 +190,7 @@ struct<> -- !query output org.apache.spark.SparkNumberFormatException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'nan'", @@ -386,7 +386,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "CAST_OVERFLOW", + "condition" : "CAST_OVERFLOW", "sqlState" : "22003", "messageParameters" : { "sourceType" : "\"FLOAT\"", @@ -411,7 +411,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "CAST_OVERFLOW", + "condition" : "CAST_OVERFLOW", "sqlState" : "22003", "messageParameters" : { "sourceType" : "\"FLOAT\"", @@ -452,7 +452,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "CAST_OVERFLOW", + "condition" : "CAST_OVERFLOW", "sqlState" : "22003", "messageParameters" : { "sourceType" : "\"FLOAT\"", diff --git a/sql/core/src/test/resources/sql-tests/results/postgreSQL/float8.sql.out b/sql/core/src/test/resources/sql-tests/results/postgreSQL/float8.sql.out index b1a114bea30e..7f779829bd83 100644 --- a/sql/core/src/test/resources/sql-tests/results/postgreSQL/float8.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/postgreSQL/float8.sql.out @@ -126,7 +126,7 @@ struct<> -- !query output org.apache.spark.SparkNumberFormatException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'N A N'", @@ -150,7 +150,7 @@ struct<> -- !query output org.apache.spark.SparkNumberFormatException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'NaN x'", @@ -174,7 +174,7 @@ struct<> -- !query output org.apache.spark.SparkNumberFormatException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "' INFINITY x'", @@ -222,7 +222,7 @@ struct<> -- !query output org.apache.spark.SparkNumberFormatException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'nan'", @@ -891,7 +891,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "CAST_OVERFLOW", + "condition" : "CAST_OVERFLOW", "sqlState" : "22003", "messageParameters" : { "sourceType" : "\"DOUBLE\"", diff --git a/sql/core/src/test/resources/sql-tests/results/postgreSQL/int4.sql.out b/sql/core/src/test/resources/sql-tests/results/postgreSQL/int4.sql.out index 16c18c86f291..de0a5eeef593 100755 --- a/sql/core/src/test/resources/sql-tests/results/postgreSQL/int4.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/postgreSQL/int4.sql.out @@ -198,7 +198,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "ARITHMETIC_OVERFLOW", + "condition" : "ARITHMETIC_OVERFLOW", "sqlState" : "22003", "messageParameters" : { "alternative" : " Use 'try_multiply' to tolerate overflow and return NULL instead.", @@ -233,7 +233,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "ARITHMETIC_OVERFLOW", + "condition" : "ARITHMETIC_OVERFLOW", "sqlState" : "22003", "messageParameters" : { "alternative" : " Use 'try_multiply' to tolerate overflow and return NULL instead.", @@ -268,7 +268,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "ARITHMETIC_OVERFLOW", + "condition" : "ARITHMETIC_OVERFLOW", "sqlState" : "22003", "messageParameters" : { "alternative" : " Use 'try_add' to tolerate overflow and return NULL instead.", @@ -304,7 +304,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "ARITHMETIC_OVERFLOW", + "condition" : "ARITHMETIC_OVERFLOW", "sqlState" : "22003", "messageParameters" : { "alternative" : " Use 'try_add' to tolerate overflow and return NULL instead.", @@ -340,7 +340,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "ARITHMETIC_OVERFLOW", + "condition" : "ARITHMETIC_OVERFLOW", "sqlState" : "22003", "messageParameters" : { "alternative" : " Use 'try_subtract' to tolerate overflow and return NULL instead.", @@ -376,7 +376,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "ARITHMETIC_OVERFLOW", + "condition" : "ARITHMETIC_OVERFLOW", "sqlState" : "22003", "messageParameters" : { "alternative" : " Use 'try_subtract' to tolerate overflow and return NULL instead.", diff --git a/sql/core/src/test/resources/sql-tests/results/postgreSQL/int8.sql.out b/sql/core/src/test/resources/sql-tests/results/postgreSQL/int8.sql.out index 5e8abc273b12..a8345e33e7ff 100755 --- a/sql/core/src/test/resources/sql-tests/results/postgreSQL/int8.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/postgreSQL/int8.sql.out @@ -390,7 +390,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "ARITHMETIC_OVERFLOW", + "condition" : "ARITHMETIC_OVERFLOW", "sqlState" : "22003", "messageParameters" : { "alternative" : " Use 'try_multiply' to tolerate overflow and return NULL instead.", @@ -660,7 +660,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "DIVIDE_BY_ZERO", + "condition" : "DIVIDE_BY_ZERO", "sqlState" : "22012", "messageParameters" : { "config" : "\"spark.sql.ansi.enabled\"" @@ -682,7 +682,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "DIVIDE_BY_ZERO", + "condition" : "DIVIDE_BY_ZERO", "sqlState" : "22012", "messageParameters" : { "config" : "\"spark.sql.ansi.enabled\"" @@ -704,7 +704,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "DIVIDE_BY_ZERO", + "condition" : "DIVIDE_BY_ZERO", "sqlState" : "22012", "messageParameters" : { "config" : "\"spark.sql.ansi.enabled\"" @@ -734,7 +734,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "CAST_OVERFLOW", + "condition" : "CAST_OVERFLOW", "sqlState" : "22003", "messageParameters" : { "sourceType" : "\"BIGINT\"", @@ -759,7 +759,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "CAST_OVERFLOW", + "condition" : "CAST_OVERFLOW", "sqlState" : "22003", "messageParameters" : { "sourceType" : "\"BIGINT\"", @@ -804,7 +804,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "CAST_OVERFLOW", + "condition" : "CAST_OVERFLOW", "sqlState" : "22003", "messageParameters" : { "sourceType" : "\"DOUBLE\"", @@ -850,7 +850,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "FAILED_FUNCTION_CALL", + "condition" : "FAILED_FUNCTION_CALL", "sqlState" : "38000", "messageParameters" : { "funcName" : "`range`" @@ -892,7 +892,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "CAST_OVERFLOW", + "condition" : "CAST_OVERFLOW", "sqlState" : "22003", "messageParameters" : { "sourceType" : "\"BIGINT\"", @@ -909,7 +909,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "ARITHMETIC_OVERFLOW", + "condition" : "ARITHMETIC_OVERFLOW", "sqlState" : "22003", "messageParameters" : { "alternative" : " Use 'try_multiply' to tolerate overflow and return NULL instead.", @@ -949,7 +949,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "ARITHMETIC_OVERFLOW", + "condition" : "ARITHMETIC_OVERFLOW", "sqlState" : "22003", "messageParameters" : { "alternative" : " Use 'try_multiply' to tolerate overflow and return NULL instead.", @@ -989,7 +989,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "ARITHMETIC_OVERFLOW", + "condition" : "ARITHMETIC_OVERFLOW", "sqlState" : "22003", "messageParameters" : { "alternative" : " Use 'try_multiply' to tolerate overflow and return NULL instead.", diff --git a/sql/core/src/test/resources/sql-tests/results/postgreSQL/int8.sql.out.java21 b/sql/core/src/test/resources/sql-tests/results/postgreSQL/int8.sql.out.java21 index e7df03dc8cad..828560520895 100755 --- a/sql/core/src/test/resources/sql-tests/results/postgreSQL/int8.sql.out.java21 +++ b/sql/core/src/test/resources/sql-tests/results/postgreSQL/int8.sql.out.java21 @@ -390,7 +390,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "ARITHMETIC_OVERFLOW", + "condition" : "ARITHMETIC_OVERFLOW", "sqlState" : "22003", "messageParameters" : { "alternative" : " Use 'try_multiply' to tolerate overflow and return NULL instead.", @@ -660,7 +660,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "DIVIDE_BY_ZERO", + "condition" : "DIVIDE_BY_ZERO", "sqlState" : "22012", "messageParameters" : { "config" : "\"spark.sql.ansi.enabled\"" @@ -682,7 +682,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "DIVIDE_BY_ZERO", + "condition" : "DIVIDE_BY_ZERO", "sqlState" : "22012", "messageParameters" : { "config" : "\"spark.sql.ansi.enabled\"" @@ -704,7 +704,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "DIVIDE_BY_ZERO", + "condition" : "DIVIDE_BY_ZERO", "sqlState" : "22012", "messageParameters" : { "config" : "\"spark.sql.ansi.enabled\"" @@ -734,7 +734,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "CAST_OVERFLOW", + "condition" : "CAST_OVERFLOW", "sqlState" : "22003", "messageParameters" : { "sourceType" : "\"BIGINT\"", @@ -759,7 +759,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "CAST_OVERFLOW", + "condition" : "CAST_OVERFLOW", "sqlState" : "22003", "messageParameters" : { "sourceType" : "\"BIGINT\"", @@ -804,7 +804,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "CAST_OVERFLOW", + "condition" : "CAST_OVERFLOW", "sqlState" : "22003", "messageParameters" : { "sourceType" : "\"DOUBLE\"", @@ -850,7 +850,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "FAILED_FUNCTION_CALL", + "condition" : "FAILED_FUNCTION_CALL", "sqlState" : "38000", "messageParameters" : { "funcName" : "`range`" @@ -892,7 +892,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "CAST_OVERFLOW", + "condition" : "CAST_OVERFLOW", "sqlState" : "22003", "messageParameters" : { "sourceType" : "\"BIGINT\"", @@ -909,7 +909,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "ARITHMETIC_OVERFLOW", + "condition" : "ARITHMETIC_OVERFLOW", "sqlState" : "22003", "messageParameters" : { "alternative" : " Use 'try_multiply' to tolerate overflow and return NULL instead.", @@ -949,7 +949,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "ARITHMETIC_OVERFLOW", + "condition" : "ARITHMETIC_OVERFLOW", "sqlState" : "22003", "messageParameters" : { "alternative" : " Use 'try_multiply' to tolerate overflow and return NULL instead.", @@ -989,7 +989,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "ARITHMETIC_OVERFLOW", + "condition" : "ARITHMETIC_OVERFLOW", "sqlState" : "22003", "messageParameters" : { "alternative" : " Use 'try_multiply' to tolerate overflow and return NULL instead.", diff --git a/sql/core/src/test/resources/sql-tests/results/postgreSQL/interval.sql.out b/sql/core/src/test/resources/sql-tests/results/postgreSQL/interval.sql.out index 3855d922361b..42e943ef7def 100644 --- a/sql/core/src/test/resources/sql-tests/results/postgreSQL/interval.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/postgreSQL/interval.sql.out @@ -102,7 +102,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_INTERVAL_FORMAT.UNMATCHED_FORMAT_STRING_WITH_NOTICE", + "condition" : "INVALID_INTERVAL_FORMAT.UNMATCHED_FORMAT_STRING_WITH_NOTICE", "sqlState" : "22006", "messageParameters" : { "input" : "1 2:03", @@ -127,7 +127,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_INTERVAL_FORMAT.UNMATCHED_FORMAT_STRING_WITH_NOTICE", + "condition" : "INVALID_INTERVAL_FORMAT.UNMATCHED_FORMAT_STRING_WITH_NOTICE", "sqlState" : "22006", "messageParameters" : { "input" : "1 2:03:04", @@ -160,7 +160,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_INTERVAL_FORMAT.UNMATCHED_FORMAT_STRING_WITH_NOTICE", + "condition" : "INVALID_INTERVAL_FORMAT.UNMATCHED_FORMAT_STRING_WITH_NOTICE", "sqlState" : "22006", "messageParameters" : { "input" : "1 2:03:04", @@ -185,7 +185,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_INTERVAL_FORMAT.UNMATCHED_FORMAT_STRING_WITH_NOTICE", + "condition" : "INVALID_INTERVAL_FORMAT.UNMATCHED_FORMAT_STRING_WITH_NOTICE", "sqlState" : "22006", "messageParameters" : { "input" : "1 2:03", @@ -218,7 +218,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_INTERVAL_FORMAT.UNMATCHED_FORMAT_STRING_WITH_NOTICE", + "condition" : "INVALID_INTERVAL_FORMAT.UNMATCHED_FORMAT_STRING_WITH_NOTICE", "sqlState" : "22006", "messageParameters" : { "input" : "1 2:03", @@ -243,7 +243,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_INTERVAL_FORMAT.UNMATCHED_FORMAT_STRING_WITH_NOTICE", + "condition" : "INVALID_INTERVAL_FORMAT.UNMATCHED_FORMAT_STRING_WITH_NOTICE", "sqlState" : "22006", "messageParameters" : { "input" : "1 2:03:04", @@ -268,7 +268,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_INTERVAL_FORMAT.UNMATCHED_FORMAT_STRING_WITH_NOTICE", + "condition" : "INVALID_INTERVAL_FORMAT.UNMATCHED_FORMAT_STRING_WITH_NOTICE", "sqlState" : "22006", "messageParameters" : { "input" : "1 2:03", @@ -293,7 +293,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_INTERVAL_FORMAT.UNMATCHED_FORMAT_STRING_WITH_NOTICE", + "condition" : "INVALID_INTERVAL_FORMAT.UNMATCHED_FORMAT_STRING_WITH_NOTICE", "sqlState" : "22006", "messageParameters" : { "input" : "1 2:03:04", @@ -318,7 +318,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_INTERVAL_FORMAT.UNMATCHED_FORMAT_STRING_WITH_NOTICE", + "condition" : "INVALID_INTERVAL_FORMAT.UNMATCHED_FORMAT_STRING_WITH_NOTICE", "sqlState" : "22006", "messageParameters" : { "input" : "1 2:03", @@ -343,7 +343,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_INTERVAL_FORMAT.UNMATCHED_FORMAT_STRING_WITH_NOTICE", + "condition" : "INVALID_INTERVAL_FORMAT.UNMATCHED_FORMAT_STRING_WITH_NOTICE", "sqlState" : "22006", "messageParameters" : { "input" : "1 2:03:04", diff --git a/sql/core/src/test/resources/sql-tests/results/postgreSQL/join.sql.out b/sql/core/src/test/resources/sql-tests/results/postgreSQL/join.sql.out index 55aaff818ea9..aeee2f2667ea 100644 --- a/sql/core/src/test/resources/sql-tests/results/postgreSQL/join.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/postgreSQL/join.sql.out @@ -547,7 +547,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "AMBIGUOUS_REFERENCE", + "condition" : "AMBIGUOUS_REFERENCE", "sqlState" : "42704", "messageParameters" : { "name" : "`i`", @@ -3332,7 +3332,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "AMBIGUOUS_REFERENCE", + "condition" : "AMBIGUOUS_REFERENCE", "sqlState" : "42704", "messageParameters" : { "name" : "`f1`", @@ -3356,7 +3356,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`y`.`f1`", @@ -3389,7 +3389,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`t1`.`uunique1`", @@ -3413,7 +3413,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`t2`.`uunique1`", @@ -3437,7 +3437,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`uunique1`", @@ -3651,7 +3651,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITHOUT_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITHOUT_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`f1`" @@ -3673,7 +3673,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITHOUT_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITHOUT_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`a`.`f1`" @@ -3695,7 +3695,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITHOUT_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITHOUT_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`f1`" @@ -3717,7 +3717,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITHOUT_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITHOUT_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`a`.`f1`" diff --git a/sql/core/src/test/resources/sql-tests/results/postgreSQL/limit.sql.out b/sql/core/src/test/resources/sql-tests/results/postgreSQL/limit.sql.out index 1addd56b8f0e..18f0bd332834 100644 --- a/sql/core/src/test/resources/sql-tests/results/postgreSQL/limit.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/postgreSQL/limit.sql.out @@ -132,7 +132,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INVALID_LIMIT_LIKE_EXPRESSION.IS_UNFOLDABLE", + "condition" : "INVALID_LIMIT_LIKE_EXPRESSION.IS_UNFOLDABLE", "sqlState" : "42K0E", "messageParameters" : { "expr" : "\"CASE WHEN (_nondeterministic < 0.5) THEN NULL END\"", @@ -155,7 +155,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INVALID_LIMIT_LIKE_EXPRESSION.IS_UNFOLDABLE", + "condition" : "INVALID_LIMIT_LIKE_EXPRESSION.IS_UNFOLDABLE", "sqlState" : "42K0E", "messageParameters" : { "expr" : "\"CASE WHEN (_nondeterministic < 0.5) THEN NULL END\"", diff --git a/sql/core/src/test/resources/sql-tests/results/postgreSQL/numeric.sql.out b/sql/core/src/test/resources/sql-tests/results/postgreSQL/numeric.sql.out index 7b40f0bcaf23..1047ce45f6f2 100644 --- a/sql/core/src/test/resources/sql-tests/results/postgreSQL/numeric.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/postgreSQL/numeric.sql.out @@ -3582,7 +3582,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "DECIMAL_PRECISION_EXCEEDS_MAX_PRECISION", + "condition" : "DECIMAL_PRECISION_EXCEEDS_MAX_PRECISION", "sqlState" : "22003", "messageParameters" : { "maxPrecision" : "38", @@ -3832,7 +3832,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "INSERT_COLUMN_ARITY_MISMATCH.TOO_MANY_DATA_COLUMNS", + "condition" : "INSERT_COLUMN_ARITY_MISMATCH.TOO_MANY_DATA_COLUMNS", "sqlState" : "21S01", "messageParameters" : { "dataColumns" : "`id`, `id`, `val`, `val`, `(val * val)`", @@ -4697,7 +4697,7 @@ struct<> -- !query output org.apache.spark.SparkIllegalArgumentException { - "errorClass" : "INVALID_FORMAT.MISMATCH_INPUT", + "condition" : "INVALID_FORMAT.MISMATCH_INPUT", "sqlState" : "42601", "messageParameters" : { "format" : "99G999G999", @@ -4714,7 +4714,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INVALID_FORMAT.THOUSANDS_SEPS_MUST_BEFORE_DEC", + "condition" : "INVALID_FORMAT.THOUSANDS_SEPS_MUST_BEFORE_DEC", "sqlState" : "42601", "messageParameters" : { "format" : "'99G999G999D999G999'" @@ -4776,7 +4776,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INVALID_FORMAT.CONT_THOUSANDS_SEPS", + "condition" : "INVALID_FORMAT.CONT_THOUSANDS_SEPS", "sqlState" : "42601", "messageParameters" : { "format" : "'999G'" @@ -4798,7 +4798,7 @@ struct<> -- !query output org.apache.spark.SparkIllegalArgumentException { - "errorClass" : "INVALID_FORMAT.MISMATCH_INPUT", + "condition" : "INVALID_FORMAT.MISMATCH_INPUT", "sqlState" : "42601", "messageParameters" : { "format" : "999G999", diff --git a/sql/core/src/test/resources/sql-tests/results/postgreSQL/numeric.sql.out.java21 b/sql/core/src/test/resources/sql-tests/results/postgreSQL/numeric.sql.out.java21 index 7f89381c18eb..de4363fddf38 100644 --- a/sql/core/src/test/resources/sql-tests/results/postgreSQL/numeric.sql.out.java21 +++ b/sql/core/src/test/resources/sql-tests/results/postgreSQL/numeric.sql.out.java21 @@ -3582,7 +3582,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "DECIMAL_PRECISION_EXCEEDS_MAX_PRECISION", + "condition" : "DECIMAL_PRECISION_EXCEEDS_MAX_PRECISION", "sqlState" : "22003", "messageParameters" : { "maxPrecision" : "38", @@ -3832,7 +3832,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "INSERT_COLUMN_ARITY_MISMATCH.TOO_MANY_DATA_COLUMNS", + "condition" : "INSERT_COLUMN_ARITY_MISMATCH.TOO_MANY_DATA_COLUMNS", "sqlState" : "21S01", "messageParameters" : { "dataColumns" : "`id`, `id`, `val`, `val`, `(val * val)`", @@ -4697,7 +4697,7 @@ struct<> -- !query output org.apache.spark.SparkIllegalArgumentException { - "errorClass" : "INVALID_FORMAT.MISMATCH_INPUT", + "condition" : "INVALID_FORMAT.MISMATCH_INPUT", "sqlState" : "42601", "messageParameters" : { "format" : "99G999G999", @@ -4714,7 +4714,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INVALID_FORMAT.THOUSANDS_SEPS_MUST_BEFORE_DEC", + "condition" : "INVALID_FORMAT.THOUSANDS_SEPS_MUST_BEFORE_DEC", "sqlState" : "42601", "messageParameters" : { "format" : "'99G999G999D999G999'" @@ -4776,7 +4776,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INVALID_FORMAT.CONT_THOUSANDS_SEPS", + "condition" : "INVALID_FORMAT.CONT_THOUSANDS_SEPS", "sqlState" : "42601", "messageParameters" : { "format" : "'999G'" @@ -4798,7 +4798,7 @@ struct<> -- !query output org.apache.spark.SparkIllegalArgumentException { - "errorClass" : "INVALID_FORMAT.MISMATCH_INPUT", + "condition" : "INVALID_FORMAT.MISMATCH_INPUT", "sqlState" : "42601", "messageParameters" : { "format" : "999G999", diff --git a/sql/core/src/test/resources/sql-tests/results/postgreSQL/select_having.sql.out b/sql/core/src/test/resources/sql-tests/results/postgreSQL/select_having.sql.out index dc295b0578b0..eaed5ada03ac 100644 --- a/sql/core/src/test/resources/sql-tests/results/postgreSQL/select_having.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/postgreSQL/select_having.sql.out @@ -141,7 +141,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "MISSING_GROUP_BY", + "condition" : "MISSING_GROUP_BY", "sqlState" : "42803", "queryContext" : [ { "objectType" : "", @@ -160,7 +160,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`a`", @@ -199,7 +199,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "DIVIDE_BY_ZERO", + "condition" : "DIVIDE_BY_ZERO", "sqlState" : "22012", "messageParameters" : { "config" : "\"spark.sql.ansi.enabled\"" diff --git a/sql/core/src/test/resources/sql-tests/results/postgreSQL/select_implicit.sql.out b/sql/core/src/test/resources/sql-tests/results/postgreSQL/select_implicit.sql.out index f0c283cb4036..c57b1ea77720 100755 --- a/sql/core/src/test/resources/sql-tests/results/postgreSQL/select_implicit.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/postgreSQL/select_implicit.sql.out @@ -120,7 +120,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`b`", @@ -215,7 +215,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "GROUP_BY_POS_OUT_OF_RANGE", + "condition" : "GROUP_BY_POS_OUT_OF_RANGE", "sqlState" : "42805", "messageParameters" : { "index" : "3", @@ -240,7 +240,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "AMBIGUOUS_REFERENCE", + "condition" : "AMBIGUOUS_REFERENCE", "sqlState" : "42704", "messageParameters" : { "name" : "`b`", @@ -367,7 +367,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`b`", @@ -444,7 +444,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "AMBIGUOUS_REFERENCE", + "condition" : "AMBIGUOUS_REFERENCE", "sqlState" : "42704", "messageParameters" : { "name" : "`b`", @@ -482,7 +482,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "AMBIGUOUS_REFERENCE", + "condition" : "AMBIGUOUS_REFERENCE", "sqlState" : "42704", "messageParameters" : { "name" : "`b`", diff --git a/sql/core/src/test/resources/sql-tests/results/postgreSQL/strings.sql.out b/sql/core/src/test/resources/sql-tests/results/postgreSQL/strings.sql.out index b4442a767079..dc95fc986a34 100644 --- a/sql/core/src/test/resources/sql-tests/results/postgreSQL/strings.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/postgreSQL/strings.sql.out @@ -444,7 +444,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_FORMAT.ESC_IN_THE_MIDDLE", + "condition" : "INVALID_FORMAT.ESC_IN_THE_MIDDLE", "sqlState" : "42601", "messageParameters" : { "char" : "'a'", @@ -460,7 +460,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_FORMAT.ESC_IN_THE_MIDDLE", + "condition" : "INVALID_FORMAT.ESC_IN_THE_MIDDLE", "sqlState" : "42601", "messageParameters" : { "char" : "'a'", @@ -476,7 +476,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_FORMAT.ESC_IN_THE_MIDDLE", + "condition" : "INVALID_FORMAT.ESC_IN_THE_MIDDLE", "sqlState" : "42601", "messageParameters" : { "char" : "'a'", @@ -492,7 +492,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_FORMAT.ESC_IN_THE_MIDDLE", + "condition" : "INVALID_FORMAT.ESC_IN_THE_MIDDLE", "sqlState" : "42601", "messageParameters" : { "char" : "'a'", @@ -508,7 +508,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_FORMAT.ESC_IN_THE_MIDDLE", + "condition" : "INVALID_FORMAT.ESC_IN_THE_MIDDLE", "sqlState" : "42601", "messageParameters" : { "char" : "'e'", @@ -524,7 +524,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_FORMAT.ESC_IN_THE_MIDDLE", + "condition" : "INVALID_FORMAT.ESC_IN_THE_MIDDLE", "sqlState" : "42601", "messageParameters" : { "char" : "'e'", @@ -540,7 +540,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_FORMAT.ESC_IN_THE_MIDDLE", + "condition" : "INVALID_FORMAT.ESC_IN_THE_MIDDLE", "sqlState" : "42601", "messageParameters" : { "char" : "'e'", @@ -556,7 +556,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_FORMAT.ESC_IN_THE_MIDDLE", + "condition" : "INVALID_FORMAT.ESC_IN_THE_MIDDLE", "sqlState" : "42601", "messageParameters" : { "char" : "'e'", diff --git a/sql/core/src/test/resources/sql-tests/results/postgreSQL/text.sql.out b/sql/core/src/test/resources/sql-tests/results/postgreSQL/text.sql.out index 0a940f5f3c74..ffd1a4f0b692 100755 --- a/sql/core/src/test/resources/sql-tests/results/postgreSQL/text.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/postgreSQL/text.sql.out @@ -63,7 +63,7 @@ struct<> -- !query output org.apache.spark.SparkNumberFormatException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'four: 2'", @@ -87,7 +87,7 @@ struct<> -- !query output org.apache.spark.SparkNumberFormatException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'four: 2'", @@ -307,7 +307,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_PARAMETER_VALUE.ZERO_INDEX", + "condition" : "INVALID_PARAMETER_VALUE.ZERO_INDEX", "sqlState" : "22023", "messageParameters" : { "functionName" : "`format_string`", diff --git a/sql/core/src/test/resources/sql-tests/results/postgreSQL/union.sql.out b/sql/core/src/test/resources/sql-tests/results/postgreSQL/union.sql.out index 94692a57300f..24ef5834b6f4 100644 --- a/sql/core/src/test/resources/sql-tests/results/postgreSQL/union.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/postgreSQL/union.sql.out @@ -524,7 +524,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`q2`", @@ -697,7 +697,7 @@ struct<> -- !query output org.apache.spark.SparkNumberFormatException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'foo'", diff --git a/sql/core/src/test/resources/sql-tests/results/postgreSQL/window_part2.sql.out b/sql/core/src/test/resources/sql-tests/results/postgreSQL/window_part2.sql.out index 352c5f05cb06..014cb23a57fc 100644 --- a/sql/core/src/test/resources/sql-tests/results/postgreSQL/window_part2.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/postgreSQL/window_part2.sql.out @@ -223,7 +223,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "ARITHMETIC_OVERFLOW", + "condition" : "ARITHMETIC_OVERFLOW", "sqlState" : "22003", "messageParameters" : { "alternative" : " Use 'try_add' to tolerate overflow and return NULL instead.", @@ -246,7 +246,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "ARITHMETIC_OVERFLOW", + "condition" : "ARITHMETIC_OVERFLOW", "sqlState" : "22003", "messageParameters" : { "alternative" : " Use 'try_add' to tolerate overflow and return NULL instead.", @@ -486,7 +486,7 @@ struct<> -- !query output org.apache.spark.SparkNumberFormatException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'NaN'", diff --git a/sql/core/src/test/resources/sql-tests/results/postgreSQL/window_part3.sql.out b/sql/core/src/test/resources/sql-tests/results/postgreSQL/window_part3.sql.out index 6cfb2cb4b451..81c6e91f52bb 100644 --- a/sql/core/src/test/resources/sql-tests/results/postgreSQL/window_part3.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/postgreSQL/window_part3.sql.out @@ -70,7 +70,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_INLINE_TABLE.FAILED_SQL_EXPRESSION_EVALUATION", + "condition" : "INVALID_INLINE_TABLE.FAILED_SQL_EXPRESSION_EVALUATION", "sqlState" : "42000", "messageParameters" : { "sqlExpr" : "\"CAST(11:00 BST AS TIMESTAMP)\"" @@ -256,7 +256,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.RANGE_FRAME_WITHOUT_ORDER", + "condition" : "DATATYPE_MISMATCH.RANGE_FRAME_WITHOUT_ORDER", "sqlState" : "42K09", "messageParameters" : { "sqlExpr" : "\"(PARTITION BY f1 RANGE BETWEEN 1 PRECEDING AND 1 FOLLOWING)\"" @@ -319,7 +319,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "_LEGACY_ERROR_TEMP_1034", + "condition" : "_LEGACY_ERROR_TEMP_1034", "messageParameters" : { "clauseName" : "WHERE" } @@ -333,7 +333,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNSUPPORTED_EXPR_FOR_OPERATOR", + "condition" : "UNSUPPORTED_EXPR_FOR_OPERATOR", "sqlState" : "42K0E", "messageParameters" : { "invalidExprSqls" : "\"row_number() OVER (ORDER BY salary ASC NULLS FIRST ROWS BETWEEN UNBOUNDED PRECEDING AND CURRENT ROW)\"" @@ -355,7 +355,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNSUPPORTED_EXPR_FOR_OPERATOR", + "condition" : "UNSUPPORTED_EXPR_FOR_OPERATOR", "sqlState" : "42K0E", "messageParameters" : { "invalidExprSqls" : "\"RANK() OVER (ORDER BY 1 ASC NULLS FIRST ROWS BETWEEN UNBOUNDED PRECEDING AND CURRENT ROW)\"" @@ -377,7 +377,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'BY'", @@ -393,7 +393,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "_LEGACY_ERROR_TEMP_1034", + "condition" : "_LEGACY_ERROR_TEMP_1034", "messageParameters" : { "clauseName" : "WHERE" } @@ -407,7 +407,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "_LEGACY_ERROR_TEMP_1034", + "condition" : "_LEGACY_ERROR_TEMP_1034", "messageParameters" : { "clauseName" : "WHERE" } @@ -421,7 +421,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'BY'", @@ -437,7 +437,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "UNRESOLVED_ROUTINE", + "condition" : "UNRESOLVED_ROUTINE", "sqlState" : "42883", "messageParameters" : { "routineName" : "`range`", @@ -460,7 +460,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.VALUE_OUT_OF_RANGE", + "condition" : "DATATYPE_MISMATCH.VALUE_OUT_OF_RANGE", "sqlState" : "42K09", "messageParameters" : { "currentValue" : "0", @@ -485,7 +485,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.VALUE_OUT_OF_RANGE", + "condition" : "DATATYPE_MISMATCH.VALUE_OUT_OF_RANGE", "sqlState" : "42K09", "messageParameters" : { "currentValue" : "0L", diff --git a/sql/core/src/test/resources/sql-tests/results/postgreSQL/window_part4.sql.out b/sql/core/src/test/resources/sql-tests/results/postgreSQL/window_part4.sql.out index 2d539725b2a7..7ef30c7df2f8 100644 --- a/sql/core/src/test/resources/sql-tests/results/postgreSQL/window_part4.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/postgreSQL/window_part4.sql.out @@ -499,7 +499,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_INLINE_TABLE.FAILED_SQL_EXPRESSION_EVALUATION", + "condition" : "INVALID_INLINE_TABLE.FAILED_SQL_EXPRESSION_EVALUATION", "sqlState" : "42000", "messageParameters" : { "sqlExpr" : "\"CAST(nan AS INT)\"" diff --git a/sql/core/src/test/resources/sql-tests/results/postgreSQL/with.sql.out b/sql/core/src/test/resources/sql-tests/results/postgreSQL/with.sql.out index af1c33023a91..37ba104b9fa1 100644 --- a/sql/core/src/test/resources/sql-tests/results/postgreSQL/with.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/postgreSQL/with.sql.out @@ -217,7 +217,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "TABLE_OR_VIEW_NOT_FOUND", + "condition" : "TABLE_OR_VIEW_NOT_FOUND", "sqlState" : "42P01", "messageParameters" : { "relationName" : "`outermost`" @@ -360,7 +360,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "UNSUPPORTED_DATATYPE", + "condition" : "UNSUPPORTED_DATATYPE", "sqlState" : "0A000", "messageParameters" : { "typeName" : "\"BAZ\"" @@ -382,7 +382,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "UNSUPPORTED_DATATYPE", + "condition" : "UNSUPPORTED_DATATYPE", "sqlState" : "0A000", "messageParameters" : { "typeName" : "\"ORDINALITY\"" @@ -412,7 +412,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "TABLE_OR_VIEW_NOT_FOUND", + "condition" : "TABLE_OR_VIEW_NOT_FOUND", "sqlState" : "42P01", "messageParameters" : { "relationName" : "`test`" diff --git a/sql/core/src/test/resources/sql-tests/results/predicate-functions.sql.out b/sql/core/src/test/resources/sql-tests/results/predicate-functions.sql.out index e2d0563a0c45..43e5d0713947 100644 --- a/sql/core/src/test/resources/sql-tests/results/predicate-functions.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/predicate-functions.sql.out @@ -214,7 +214,7 @@ struct<> -- !query output org.apache.spark.SparkNumberFormatException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'1.0'", @@ -238,7 +238,7 @@ struct<> -- !query output org.apache.spark.SparkNumberFormatException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'2.0'", @@ -262,7 +262,7 @@ struct<> -- !query output org.apache.spark.SparkNumberFormatException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'2.2'", @@ -318,7 +318,7 @@ struct<> -- !query output org.apache.spark.SparkNumberFormatException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'1.0'", @@ -342,7 +342,7 @@ struct<> -- !query output org.apache.spark.SparkNumberFormatException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'2.0'", @@ -406,7 +406,7 @@ struct<> -- !query output org.apache.spark.SparkNumberFormatException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'1.0'", @@ -430,7 +430,7 @@ struct<> -- !query output org.apache.spark.SparkNumberFormatException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'2.0'", @@ -494,7 +494,7 @@ struct<> -- !query output org.apache.spark.SparkNumberFormatException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'1.0'", @@ -518,7 +518,7 @@ struct<> -- !query output org.apache.spark.SparkNumberFormatException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'2.0'", diff --git a/sql/core/src/test/resources/sql-tests/results/query_regex_column.sql.out b/sql/core/src/test/resources/sql-tests/results/query_regex_column.sql.out index 8df97c723e10..c72a608fcc38 100644 --- a/sql/core/src/test/resources/sql-tests/results/query_regex_column.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/query_regex_column.sql.out @@ -34,7 +34,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`(a)?+.+`", @@ -57,7 +57,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`t`.`(a)?+.+`", @@ -80,7 +80,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`(a|b)`", @@ -103,7 +103,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`(a|b)?+.+`", @@ -126,7 +126,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`(a|b)?+.+`", @@ -149,7 +149,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`(a)`", @@ -383,7 +383,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`(a)`", @@ -406,7 +406,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`(a)?+.+`", diff --git a/sql/core/src/test/resources/sql-tests/results/random.sql.out b/sql/core/src/test/resources/sql-tests/results/random.sql.out index 049d134003cd..0f6f8dcb4756 100644 --- a/sql/core/src/test/resources/sql-tests/results/random.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/random.sql.out @@ -38,7 +38,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"1.0\"", @@ -96,7 +96,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"1\"", @@ -364,7 +364,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"CAST(NULL AS FLOAT)\"", @@ -390,7 +390,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.NON_FOLDABLE_INPUT", + "condition" : "DATATYPE_MISMATCH.NON_FOLDABLE_INPUT", "sqlState" : "42K09", "messageParameters" : { "inputExpr" : "\"col\"", @@ -415,7 +415,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.NON_FOLDABLE_INPUT", + "condition" : "DATATYPE_MISMATCH.NON_FOLDABLE_INPUT", "sqlState" : "42K09", "messageParameters" : { "inputExpr" : "\"col\"", @@ -440,7 +440,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", + "condition" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", "sqlState" : "42605", "messageParameters" : { "actualNum" : "1", @@ -465,7 +465,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", + "condition" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", "sqlState" : "42605", "messageParameters" : { "actualNum" : "4", @@ -490,7 +490,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"0.0\"", @@ -516,7 +516,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"0.0\"", @@ -542,7 +542,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"CAST(0 AS DECIMAL(10,3))\"", @@ -568,7 +568,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"abc\"", @@ -594,7 +594,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"def\"", @@ -620,7 +620,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"ghi\"", @@ -784,7 +784,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.NON_FOLDABLE_INPUT", + "condition" : "DATATYPE_MISMATCH.NON_FOLDABLE_INPUT", "sqlState" : "42K09", "messageParameters" : { "inputExpr" : "\"col\"", @@ -809,7 +809,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.NON_FOLDABLE_INPUT", + "condition" : "DATATYPE_MISMATCH.NON_FOLDABLE_INPUT", "sqlState" : "42K09", "messageParameters" : { "inputExpr" : "\"col\"", @@ -834,7 +834,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", + "condition" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", "sqlState" : "42605", "messageParameters" : { "actualNum" : "3", @@ -859,7 +859,7 @@ struct<> -- !query output org.apache.spark.SparkRuntimeException { - "errorClass" : "INVALID_PARAMETER_VALUE.LENGTH", + "condition" : "INVALID_PARAMETER_VALUE.LENGTH", "sqlState" : "22023", "messageParameters" : { "functionName" : "`randstr`", @@ -876,7 +876,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"a\"", @@ -902,7 +902,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"1.5\"", diff --git a/sql/core/src/test/resources/sql-tests/results/regexp-functions.sql.out b/sql/core/src/test/resources/sql-tests/results/regexp-functions.sql.out index 62a0e5ff917c..bef4f137ee67 100644 --- a/sql/core/src/test/resources/sql-tests/results/regexp-functions.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/regexp-functions.sql.out @@ -6,7 +6,7 @@ struct<> -- !query output org.apache.spark.SparkRuntimeException { - "errorClass" : "INVALID_PARAMETER_VALUE.REGEX_GROUP_INDEX", + "condition" : "INVALID_PARAMETER_VALUE.REGEX_GROUP_INDEX", "sqlState" : "22023", "messageParameters" : { "functionName" : "`regexp_extract`", @@ -32,7 +32,7 @@ struct<> -- !query output org.apache.spark.SparkRuntimeException { - "errorClass" : "INVALID_PARAMETER_VALUE.REGEX_GROUP_INDEX", + "condition" : "INVALID_PARAMETER_VALUE.REGEX_GROUP_INDEX", "sqlState" : "22023", "messageParameters" : { "functionName" : "`regexp_extract`", @@ -50,7 +50,7 @@ struct<> -- !query output org.apache.spark.SparkRuntimeException { - "errorClass" : "INVALID_PARAMETER_VALUE.REGEX_GROUP_INDEX", + "condition" : "INVALID_PARAMETER_VALUE.REGEX_GROUP_INDEX", "sqlState" : "22023", "messageParameters" : { "functionName" : "`regexp_extract`", @@ -68,7 +68,7 @@ struct<> -- !query output org.apache.spark.SparkRuntimeException { - "errorClass" : "INVALID_PARAMETER_VALUE.REGEX_GROUP_INDEX", + "condition" : "INVALID_PARAMETER_VALUE.REGEX_GROUP_INDEX", "sqlState" : "22023", "messageParameters" : { "functionName" : "`regexp_extract`", @@ -134,7 +134,7 @@ struct<> -- !query output org.apache.spark.SparkRuntimeException { - "errorClass" : "INVALID_PARAMETER_VALUE.REGEX_GROUP_INDEX", + "condition" : "INVALID_PARAMETER_VALUE.REGEX_GROUP_INDEX", "sqlState" : "22023", "messageParameters" : { "functionName" : "`regexp_extract`", @@ -152,7 +152,7 @@ struct<> -- !query output org.apache.spark.SparkRuntimeException { - "errorClass" : "INVALID_PARAMETER_VALUE.REGEX_GROUP_INDEX", + "condition" : "INVALID_PARAMETER_VALUE.REGEX_GROUP_INDEX", "sqlState" : "22023", "messageParameters" : { "functionName" : "`regexp_extract`", @@ -186,7 +186,7 @@ struct<> -- !query output org.apache.spark.SparkRuntimeException { - "errorClass" : "INVALID_PARAMETER_VALUE.PATTERN", + "condition" : "INVALID_PARAMETER_VALUE.PATTERN", "sqlState" : "22023", "messageParameters" : { "functionName" : "`regexp_extract`", @@ -203,7 +203,7 @@ struct<> -- !query output org.apache.spark.SparkRuntimeException { - "errorClass" : "INVALID_PARAMETER_VALUE.REGEX_GROUP_INDEX", + "condition" : "INVALID_PARAMETER_VALUE.REGEX_GROUP_INDEX", "sqlState" : "22023", "messageParameters" : { "functionName" : "`regexp_extract_all`", @@ -229,7 +229,7 @@ struct<> -- !query output org.apache.spark.SparkRuntimeException { - "errorClass" : "INVALID_PARAMETER_VALUE.REGEX_GROUP_INDEX", + "condition" : "INVALID_PARAMETER_VALUE.REGEX_GROUP_INDEX", "sqlState" : "22023", "messageParameters" : { "functionName" : "`regexp_extract_all`", @@ -247,7 +247,7 @@ struct<> -- !query output org.apache.spark.SparkRuntimeException { - "errorClass" : "INVALID_PARAMETER_VALUE.REGEX_GROUP_INDEX", + "condition" : "INVALID_PARAMETER_VALUE.REGEX_GROUP_INDEX", "sqlState" : "22023", "messageParameters" : { "functionName" : "`regexp_extract_all`", @@ -265,7 +265,7 @@ struct<> -- !query output org.apache.spark.SparkRuntimeException { - "errorClass" : "INVALID_PARAMETER_VALUE.REGEX_GROUP_INDEX", + "condition" : "INVALID_PARAMETER_VALUE.REGEX_GROUP_INDEX", "sqlState" : "22023", "messageParameters" : { "functionName" : "`regexp_extract_all`", @@ -331,7 +331,7 @@ struct<> -- !query output org.apache.spark.SparkRuntimeException { - "errorClass" : "INVALID_PARAMETER_VALUE.REGEX_GROUP_INDEX", + "condition" : "INVALID_PARAMETER_VALUE.REGEX_GROUP_INDEX", "sqlState" : "22023", "messageParameters" : { "functionName" : "`regexp_extract_all`", @@ -349,7 +349,7 @@ struct<> -- !query output org.apache.spark.SparkRuntimeException { - "errorClass" : "INVALID_PARAMETER_VALUE.REGEX_GROUP_INDEX", + "condition" : "INVALID_PARAMETER_VALUE.REGEX_GROUP_INDEX", "sqlState" : "22023", "messageParameters" : { "functionName" : "`regexp_extract_all`", @@ -383,7 +383,7 @@ struct<> -- !query output org.apache.spark.SparkRuntimeException { - "errorClass" : "INVALID_PARAMETER_VALUE.PATTERN", + "condition" : "INVALID_PARAMETER_VALUE.PATTERN", "sqlState" : "22023", "messageParameters" : { "functionName" : "`regexp_extract_all`", @@ -408,7 +408,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.VALUE_OUT_OF_RANGE", + "condition" : "DATATYPE_MISMATCH.VALUE_OUT_OF_RANGE", "sqlState" : "42K09", "messageParameters" : { "currentValue" : "-2", @@ -433,7 +433,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.VALUE_OUT_OF_RANGE", + "condition" : "DATATYPE_MISMATCH.VALUE_OUT_OF_RANGE", "sqlState" : "42K09", "messageParameters" : { "currentValue" : "0", @@ -722,7 +722,7 @@ struct<> -- !query output org.apache.spark.SparkRuntimeException { - "errorClass" : "INVALID_PARAMETER_VALUE.PATTERN", + "condition" : "INVALID_PARAMETER_VALUE.PATTERN", "sqlState" : "22023", "messageParameters" : { "functionName" : "`regexp_instr`", diff --git a/sql/core/src/test/resources/sql-tests/results/selectExcept.sql.out b/sql/core/src/test/resources/sql-tests/results/selectExcept.sql.out index b01d1a04e59f..ee29248ec3a1 100644 --- a/sql/core/src/test/resources/sql-tests/results/selectExcept.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/selectExcept.sql.out @@ -245,7 +245,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'name'", @@ -261,7 +261,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "')'", @@ -277,7 +277,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`invalid_column`", @@ -300,7 +300,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`invalid_column`", @@ -323,7 +323,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "EXCEPT_OVERLAPPING_COLUMNS", + "condition" : "EXCEPT_OVERLAPPING_COLUMNS", "sqlState" : "42702", "messageParameters" : { "columns" : "id, id" @@ -345,7 +345,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "EXCEPT_OVERLAPPING_COLUMNS", + "condition" : "EXCEPT_OVERLAPPING_COLUMNS", "sqlState" : "42702", "messageParameters" : { "columns" : "data.s2, data.s2.f2" diff --git a/sql/core/src/test/resources/sql-tests/results/show-tables.sql.out b/sql/core/src/test/resources/sql-tests/results/show-tables.sql.out index 93c144477759..0911efe3e09c 100644 --- a/sql/core/src/test/resources/sql-tests/results/show-tables.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/show-tables.sql.out @@ -168,7 +168,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "end of input", @@ -195,7 +195,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'PARTITION'", @@ -211,7 +211,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "TABLE_OR_VIEW_NOT_FOUND", + "condition" : "TABLE_OR_VIEW_NOT_FOUND", "sqlState" : "42P01", "messageParameters" : { "relationName" : "`show_t*`" @@ -233,7 +233,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "_LEGACY_ERROR_TEMP_1232", + "condition" : "_LEGACY_ERROR_TEMP_1232", "messageParameters" : { "partitionColumnNames" : "c, d", "specKeys" : "c", @@ -249,7 +249,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "PARTITIONS_NOT_FOUND", + "condition" : "PARTITIONS_NOT_FOUND", "sqlState" : "428FT", "messageParameters" : { "partitionList" : "`a`", @@ -265,7 +265,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.analysis.NoSuchPartitionException { - "errorClass" : "PARTITIONS_NOT_FOUND", + "condition" : "PARTITIONS_NOT_FOUND", "sqlState" : "428FT", "messageParameters" : { "partitionList" : "PARTITION (`c` = Ch, `d` = 1)", diff --git a/sql/core/src/test/resources/sql-tests/results/show-views.sql.out b/sql/core/src/test/resources/sql-tests/results/show-views.sql.out index 249f5a1d9552..910aba40ee87 100644 --- a/sql/core/src/test/resources/sql-tests/results/show-views.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/show-views.sql.out @@ -140,7 +140,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.analysis.NoSuchNamespaceException { - "errorClass" : "SCHEMA_NOT_FOUND", + "condition" : "SCHEMA_NOT_FOUND", "sqlState" : "42704", "messageParameters" : { "schemaName" : "`spark_catalog`.`wrongdb`" diff --git a/sql/core/src/test/resources/sql-tests/results/show_columns.sql.out b/sql/core/src/test/resources/sql-tests/results/show_columns.sql.out index bb4e7e08c6f5..5ea827e517dd 100644 --- a/sql/core/src/test/resources/sql-tests/results/show_columns.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/show_columns.sql.out @@ -92,7 +92,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "TABLE_OR_VIEW_NOT_FOUND", + "condition" : "TABLE_OR_VIEW_NOT_FOUND", "sqlState" : "42P01", "messageParameters" : { "relationName" : "`showdb`.`badtable`" @@ -123,7 +123,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "SHOW_COLUMNS_WITH_CONFLICT_NAMESPACE", + "condition" : "SHOW_COLUMNS_WITH_CONFLICT_NAMESPACE", "sqlState" : "42K05", "messageParameters" : { "namespaceA" : "`baddb`", @@ -148,7 +148,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "TABLE_OR_VIEW_NOT_FOUND", + "condition" : "TABLE_OR_VIEW_NOT_FOUND", "sqlState" : "42P01", "messageParameters" : { "relationName" : "`showdb`.`showcolumn3`" @@ -170,7 +170,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "TABLE_OR_VIEW_NOT_FOUND", + "condition" : "TABLE_OR_VIEW_NOT_FOUND", "sqlState" : "42P01", "messageParameters" : { "relationName" : "`showdb`.`showcolumn3`" @@ -192,7 +192,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "TABLE_OR_VIEW_NOT_FOUND", + "condition" : "TABLE_OR_VIEW_NOT_FOUND", "sqlState" : "42P01", "messageParameters" : { "relationName" : "`showcolumn4`" diff --git a/sql/core/src/test/resources/sql-tests/results/sql-compatibility-functions.sql.out b/sql/core/src/test/resources/sql-tests/results/sql-compatibility-functions.sql.out index 0dd8c738d212..367f586a7e1b 100644 --- a/sql/core/src/test/resources/sql-tests/results/sql-compatibility-functions.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/sql-compatibility-functions.sql.out @@ -94,7 +94,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", + "condition" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", "sqlState" : "42605", "messageParameters" : { "actualNum" : "2", diff --git a/sql/core/src/test/resources/sql-tests/results/sql-on-files.sql.out b/sql/core/src/test/resources/sql-tests/results/sql-on-files.sql.out index 71d4216ea207..13f43167c4dd 100644 --- a/sql/core/src/test/resources/sql-tests/results/sql-on-files.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/sql-on-files.sql.out @@ -22,7 +22,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_EMPTY_LOCATION", + "condition" : "INVALID_EMPTY_LOCATION", "sqlState" : "42K05", "messageParameters" : { "location" : "" @@ -44,7 +44,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "PATH_NOT_FOUND", + "condition" : "PATH_NOT_FOUND", "sqlState" : "42K03", "messageParameters" : { "path" : "file:/file/not/found" @@ -83,7 +83,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_EMPTY_LOCATION", + "condition" : "INVALID_EMPTY_LOCATION", "sqlState" : "42K05", "messageParameters" : { "location" : "" @@ -105,7 +105,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "PATH_NOT_FOUND", + "condition" : "PATH_NOT_FOUND", "sqlState" : "42K03", "messageParameters" : { "path" : "file:/file/not/found" @@ -144,7 +144,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_EMPTY_LOCATION", + "condition" : "INVALID_EMPTY_LOCATION", "sqlState" : "42K05", "messageParameters" : { "location" : "" @@ -166,7 +166,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "PATH_NOT_FOUND", + "condition" : "PATH_NOT_FOUND", "sqlState" : "42K03", "messageParameters" : { "path" : "file:/file/not/found" @@ -205,7 +205,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_EMPTY_LOCATION", + "condition" : "INVALID_EMPTY_LOCATION", "sqlState" : "42K05", "messageParameters" : { "location" : "" @@ -227,7 +227,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "PATH_NOT_FOUND", + "condition" : "PATH_NOT_FOUND", "sqlState" : "42K03", "messageParameters" : { "path" : "file:/file/not/found" @@ -266,7 +266,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "FAILED_READ_FILE.UNSUPPORTED_FILE_SYSTEM", + "condition" : "FAILED_READ_FILE.UNSUPPORTED_FILE_SYSTEM", "sqlState" : "KD001", "messageParameters" : { "fileSystemClass" : "org.apache.hadoop.fs.http.HttpsFileSystem", diff --git a/sql/core/src/test/resources/sql-tests/results/sql-session-variables.sql.out b/sql/core/src/test/resources/sql-tests/results/sql-session-variables.sql.out index 249a03fdfbf8..5e26b0a125b8 100644 --- a/sql/core/src/test/resources/sql-tests/results/sql-session-variables.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/sql-session-variables.sql.out @@ -142,7 +142,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'EXISTS'", @@ -198,7 +198,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITHOUT_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITHOUT_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`var1`" @@ -220,7 +220,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "VARIABLE_NOT_FOUND", + "condition" : "VARIABLE_NOT_FOUND", "sqlState" : "42883", "messageParameters" : { "variableName" : "`system`.`session`.`var1`" @@ -251,7 +251,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'VARIABLE'", @@ -267,7 +267,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'VARIABLE'", @@ -443,7 +443,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "VARIABLE_NOT_FOUND", + "condition" : "VARIABLE_NOT_FOUND", "sqlState" : "42883", "messageParameters" : { "variableName" : "`system`.`session`.`var1`" @@ -474,7 +474,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "VARIABLE_NOT_FOUND", + "condition" : "VARIABLE_NOT_FOUND", "sqlState" : "42883", "messageParameters" : { "variableName" : "`system`.`session`.`var1`" @@ -505,7 +505,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "VARIABLE_NOT_FOUND", + "condition" : "VARIABLE_NOT_FOUND", "sqlState" : "42883", "messageParameters" : { "variableName" : "`system`.`session`.`var1`" @@ -536,7 +536,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "VARIABLE_NOT_FOUND", + "condition" : "VARIABLE_NOT_FOUND", "sqlState" : "42883", "messageParameters" : { "variableName" : "`system`.`session`.`var1`" @@ -576,7 +576,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "UNRESOLVED_VARIABLE", + "condition" : "UNRESOLVED_VARIABLE", "sqlState" : "42883", "messageParameters" : { "searchPath" : "`system`.`session`", @@ -592,7 +592,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "UNRESOLVED_VARIABLE", + "condition" : "UNRESOLVED_VARIABLE", "sqlState" : "42883", "messageParameters" : { "searchPath" : "`system`.`session`", @@ -608,7 +608,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "UNRESOLVED_VARIABLE", + "condition" : "UNRESOLVED_VARIABLE", "sqlState" : "42883", "messageParameters" : { "searchPath" : "`system`.`session`", @@ -632,7 +632,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITHOUT_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITHOUT_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`var`" @@ -654,7 +654,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITHOUT_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITHOUT_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`ses`.`var1`" @@ -676,7 +676,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITHOUT_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITHOUT_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`b`.`sesson`.`var1`" @@ -698,7 +698,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITHOUT_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITHOUT_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`builtn`.`session`.`var1`" @@ -720,7 +720,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "UNRESOLVED_VARIABLE", + "condition" : "UNRESOLVED_VARIABLE", "sqlState" : "42883", "messageParameters" : { "searchPath" : "`SYSTEM`.`SESSION`", @@ -736,7 +736,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "UNRESOLVED_VARIABLE", + "condition" : "UNRESOLVED_VARIABLE", "sqlState" : "42883", "messageParameters" : { "searchPath" : "`SYSTEM`.`SESSION`", @@ -896,7 +896,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_SQL_SYNTAX.VARIABLE_TYPE_OR_DEFAULT_REQUIRED", + "condition" : "INVALID_SQL_SYNTAX.VARIABLE_TYPE_OR_DEFAULT_REQUIRED", "sqlState" : "42000", "queryContext" : [ { "objectType" : "", @@ -923,7 +923,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INVALID_DEFAULT_VALUE.SUBQUERY_EXPRESSION", + "condition" : "INVALID_DEFAULT_VALUE.SUBQUERY_EXPRESSION", "sqlState" : "42623", "messageParameters" : { "colName" : "`system`.`session`.`var1`", @@ -940,7 +940,7 @@ struct<> -- !query output org.apache.spark.SparkNumberFormatException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'hello'", @@ -964,7 +964,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "DIVIDE_BY_ZERO", + "condition" : "DIVIDE_BY_ZERO", "sqlState" : "22012", "messageParameters" : { "config" : "\"spark.sql.ansi.enabled\"" @@ -986,7 +986,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "CAST_OVERFLOW", + "condition" : "CAST_OVERFLOW", "sqlState" : "22003", "messageParameters" : { "sourceType" : "\"INT\"", @@ -1099,7 +1099,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "CAST_OVERFLOW", + "condition" : "CAST_OVERFLOW", "sqlState" : "22003", "messageParameters" : { "sourceType" : "\"DOUBLE\"", @@ -1124,7 +1124,7 @@ struct<> -- !query output org.apache.spark.SparkException { - "errorClass" : "SCALAR_SUBQUERY_TOO_MANY_ROWS", + "condition" : "SCALAR_SUBQUERY_TOO_MANY_ROWS", "sqlState" : "21000", "queryContext" : [ { "objectType" : "", @@ -1143,7 +1143,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INVALID_SUBQUERY_EXPRESSION.SCALAR_SUBQUERY_RETURN_MORE_THAN_ONE_OUTPUT_COLUMN", + "condition" : "INVALID_SUBQUERY_EXPRESSION.SCALAR_SUBQUERY_RETURN_MORE_THAN_ONE_OUTPUT_COLUMN", "sqlState" : "42823", "messageParameters" : { "number" : "2" @@ -1165,7 +1165,7 @@ struct<> -- !query output org.apache.spark.SparkNumberFormatException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'hello'", @@ -1333,7 +1333,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "DUPLICATE_ASSIGNMENTS", + "condition" : "DUPLICATE_ASSIGNMENTS", "sqlState" : "42701", "messageParameters" : { "nameList" : "`var1`" @@ -1484,7 +1484,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_SET_SYNTAX", + "condition" : "INVALID_SET_SYNTAX", "sqlState" : "42000", "queryContext" : [ { "objectType" : "", @@ -1503,7 +1503,7 @@ struct<> -- !query output org.apache.spark.SparkException { - "errorClass" : "ROW_SUBQUERY_TOO_MANY_ROWS", + "condition" : "ROW_SUBQUERY_TOO_MANY_ROWS", "sqlState" : "21000" } @@ -1515,7 +1515,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "ASSIGNMENT_ARITY_MISMATCH", + "condition" : "ASSIGNMENT_ARITY_MISMATCH", "sqlState" : "42802", "messageParameters" : { "numExpr" : "2", @@ -1531,7 +1531,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "ASSIGNMENT_ARITY_MISMATCH", + "condition" : "ASSIGNMENT_ARITY_MISMATCH", "sqlState" : "42802", "messageParameters" : { "numExpr" : "4", @@ -1547,7 +1547,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "DUPLICATE_ASSIGNMENTS", + "condition" : "DUPLICATE_ASSIGNMENTS", "sqlState" : "42701", "messageParameters" : { "nameList" : "`var1`" @@ -1786,7 +1786,7 @@ struct<> -- !query output org.apache.spark.SparkException { - "errorClass" : "ROW_SUBQUERY_TOO_MANY_ROWS", + "condition" : "ROW_SUBQUERY_TOO_MANY_ROWS", "sqlState" : "21000" } @@ -1814,7 +1814,7 @@ struct<> -- !query output org.apache.spark.SparkException { - "errorClass" : "ROW_SUBQUERY_TOO_MANY_ROWS", + "condition" : "ROW_SUBQUERY_TOO_MANY_ROWS", "sqlState" : "21000" } @@ -1834,7 +1834,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "DEFAULT_PLACEMENT_INVALID", + "condition" : "DEFAULT_PLACEMENT_INVALID", "sqlState" : "42608" } @@ -1854,7 +1854,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "DEFAULT_PLACEMENT_INVALID", + "condition" : "DEFAULT_PLACEMENT_INVALID", "sqlState" : "42608" } @@ -1866,7 +1866,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITHOUT_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITHOUT_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`session`.`default`" @@ -1960,7 +1960,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "UNSUPPORTED_FEATURE.SET_VARIABLE_USING_SET", + "condition" : "UNSUPPORTED_FEATURE.SET_VARIABLE_USING_SET", "sqlState" : "0A000", "messageParameters" : { "variableName" : "`var1`" @@ -1975,7 +1975,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "UNSUPPORTED_FEATURE.SET_VARIABLE_USING_SET", + "condition" : "UNSUPPORTED_FEATURE.SET_VARIABLE_USING_SET", "sqlState" : "0A000", "messageParameters" : { "variableName" : "`session`.`var1`" @@ -1990,7 +1990,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "UNSUPPORTED_FEATURE.SET_VARIABLE_USING_SET", + "condition" : "UNSUPPORTED_FEATURE.SET_VARIABLE_USING_SET", "sqlState" : "0A000", "messageParameters" : { "variableName" : "`system`.`session`.`var1`" @@ -2005,7 +2005,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "UNSUPPORTED_FEATURE.SET_VARIABLE_USING_SET", + "condition" : "UNSUPPORTED_FEATURE.SET_VARIABLE_USING_SET", "sqlState" : "0A000", "messageParameters" : { "variableName" : "`vAr1`" @@ -2020,7 +2020,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "UNSUPPORTED_FEATURE.SET_VARIABLE_USING_SET", + "condition" : "UNSUPPORTED_FEATURE.SET_VARIABLE_USING_SET", "sqlState" : "0A000", "messageParameters" : { "variableName" : "`seSSion`.`var1`" @@ -2035,7 +2035,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "UNSUPPORTED_FEATURE.SET_VARIABLE_USING_SET", + "condition" : "UNSUPPORTED_FEATURE.SET_VARIABLE_USING_SET", "sqlState" : "0A000", "messageParameters" : { "variableName" : "`sYStem`.`session`.`var1`" @@ -2186,7 +2186,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'('", @@ -2274,7 +2274,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_TEMP_OBJ_REFERENCE", + "condition" : "INVALID_TEMP_OBJ_REFERENCE", "sqlState" : "42K0F", "messageParameters" : { "obj" : "VIEW", diff --git a/sql/core/src/test/resources/sql-tests/results/sql-udf.sql.out b/sql/core/src/test/resources/sql-tests/results/sql-udf.sql.out index 9f7af7c64487..08f2d75cce9d 100644 --- a/sql/core/src/test/resources/sql-tests/results/sql-udf.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/sql-udf.sql.out @@ -22,7 +22,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", + "condition" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", "sqlState" : "42605", "messageParameters" : { "actualNum" : "1", @@ -63,7 +63,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", + "condition" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", "sqlState" : "42605", "messageParameters" : { "actualNum" : "2", @@ -200,7 +200,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITHOUT_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITHOUT_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`a`" @@ -222,7 +222,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITHOUT_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITHOUT_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`a`" diff --git a/sql/core/src/test/resources/sql-tests/results/string-functions.sql.out b/sql/core/src/test/resources/sql-tests/results/string-functions.sql.out index 706673606625..5dfbe1c228c2 100644 --- a/sql/core/src/test/resources/sql-tests/results/string-functions.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/string-functions.sql.out @@ -6,7 +6,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", + "condition" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", "sqlState" : "42605", "messageParameters" : { "actualNum" : "0", @@ -24,7 +24,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", + "condition" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", "sqlState" : "42605", "messageParameters" : { "actualNum" : "0", @@ -98,7 +98,7 @@ struct<> -- !query output org.apache.spark.SparkNumberFormatException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'a'", @@ -138,7 +138,7 @@ struct<> -- !query output org.apache.spark.SparkNumberFormatException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'a'", @@ -250,7 +250,7 @@ struct<> -- !query output org.apache.spark.SparkRuntimeException { - "errorClass" : "INVALID_INDEX_OF_ZERO", + "condition" : "INVALID_INDEX_OF_ZERO", "sqlState" : "22003" } @@ -502,7 +502,7 @@ struct<> -- !query output org.apache.spark.SparkNumberFormatException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'invalid_length'", @@ -526,7 +526,7 @@ struct<> -- !query output org.apache.spark.SparkNumberFormatException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'invalid_length'", @@ -838,7 +838,7 @@ struct<> -- !query output org.apache.spark.SparkIllegalArgumentException { - "errorClass" : "INVALID_PARAMETER_VALUE.CHARSET", + "condition" : "INVALID_PARAMETER_VALUE.CHARSET", "sqlState" : "22023", "messageParameters" : { "charset" : "WINDOWS-1252", @@ -856,7 +856,7 @@ struct<> -- !query output org.apache.spark.SparkIllegalArgumentException { - "errorClass" : "INVALID_PARAMETER_VALUE.CHARSET", + "condition" : "INVALID_PARAMETER_VALUE.CHARSET", "sqlState" : "22023", "messageParameters" : { "charset" : "WINDOWS-1252", @@ -874,7 +874,7 @@ struct<> -- !query output org.apache.spark.SparkIllegalArgumentException { - "errorClass" : "INVALID_PARAMETER_VALUE.CHARSET", + "condition" : "INVALID_PARAMETER_VALUE.CHARSET", "sqlState" : "22023", "messageParameters" : { "charset" : "Windows-xxx", @@ -892,7 +892,7 @@ struct<> -- !query output org.apache.spark.SparkIllegalArgumentException { - "errorClass" : "INVALID_PARAMETER_VALUE.CHARSET", + "condition" : "INVALID_PARAMETER_VALUE.CHARSET", "sqlState" : "22023", "messageParameters" : { "charset" : "Windows-xxx", @@ -942,7 +942,7 @@ struct<> -- !query output org.apache.spark.SparkRuntimeException { - "errorClass" : "MALFORMED_CHARACTER_CODING", + "condition" : "MALFORMED_CHARACTER_CODING", "sqlState" : "22000", "messageParameters" : { "charset" : "US-ASCII", @@ -958,7 +958,7 @@ struct<> -- !query output org.apache.spark.SparkRuntimeException { - "errorClass" : "MALFORMED_CHARACTER_CODING", + "condition" : "MALFORMED_CHARACTER_CODING", "sqlState" : "22000", "messageParameters" : { "charset" : "US-ASCII", @@ -1014,7 +1014,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", + "condition" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", "sqlState" : "42605", "messageParameters" : { "actualNum" : "0", @@ -1039,7 +1039,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", + "condition" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", "sqlState" : "42605", "messageParameters" : { "actualNum" : "1", @@ -1136,7 +1136,7 @@ struct<> -- !query output org.apache.spark.SparkIllegalArgumentException { - "errorClass" : "INVALID_PARAMETER_VALUE.CHARSET", + "condition" : "INVALID_PARAMETER_VALUE.CHARSET", "sqlState" : "22023", "messageParameters" : { "charset" : "Windows-xxx", @@ -1154,7 +1154,7 @@ struct<> -- !query output org.apache.spark.SparkIllegalArgumentException { - "errorClass" : "INVALID_PARAMETER_VALUE.CHARSET", + "condition" : "INVALID_PARAMETER_VALUE.CHARSET", "sqlState" : "22023", "messageParameters" : { "charset" : "Windows-xxx", @@ -1204,7 +1204,7 @@ struct<> -- !query output org.apache.spark.SparkIllegalArgumentException { - "errorClass" : "INVALID_PARAMETER_VALUE.CHARSET", + "condition" : "INVALID_PARAMETER_VALUE.CHARSET", "sqlState" : "22023", "messageParameters" : { "charset" : "WINDOWS-1252", @@ -1222,7 +1222,7 @@ struct<> -- !query output org.apache.spark.SparkIllegalArgumentException { - "errorClass" : "INVALID_PARAMETER_VALUE.CHARSET", + "condition" : "INVALID_PARAMETER_VALUE.CHARSET", "sqlState" : "22023", "messageParameters" : { "charset" : "WINDOWS-1252", @@ -1272,7 +1272,7 @@ struct<> -- !query output org.apache.spark.SparkRuntimeException { - "errorClass" : "MALFORMED_CHARACTER_CODING", + "condition" : "MALFORMED_CHARACTER_CODING", "sqlState" : "22000", "messageParameters" : { "charset" : "US-ASCII", @@ -1288,7 +1288,7 @@ struct<> -- !query output org.apache.spark.SparkRuntimeException { - "errorClass" : "MALFORMED_CHARACTER_CODING", + "condition" : "MALFORMED_CHARACTER_CODING", "sqlState" : "22000", "messageParameters" : { "charset" : "US-ASCII", @@ -1664,7 +1664,7 @@ struct<> -- !query output org.apache.spark.SparkIllegalArgumentException { - "errorClass" : "CONVERSION_INVALID_INPUT", + "condition" : "CONVERSION_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "fmt" : "'BASE64'", @@ -1692,7 +1692,7 @@ struct<> -- !query output org.apache.spark.SparkIllegalArgumentException { - "errorClass" : "CONVERSION_INVALID_INPUT", + "condition" : "CONVERSION_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "fmt" : "'BASE64'", @@ -1710,7 +1710,7 @@ struct<> -- !query output org.apache.spark.SparkIllegalArgumentException { - "errorClass" : "CONVERSION_INVALID_INPUT", + "condition" : "CONVERSION_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "fmt" : "'BASE64'", @@ -1728,7 +1728,7 @@ struct<> -- !query output org.apache.spark.SparkIllegalArgumentException { - "errorClass" : "CONVERSION_INVALID_INPUT", + "condition" : "CONVERSION_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "fmt" : "'BASE64'", @@ -1746,7 +1746,7 @@ struct<> -- !query output org.apache.spark.SparkIllegalArgumentException { - "errorClass" : "CONVERSION_INVALID_INPUT", + "condition" : "CONVERSION_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "fmt" : "'BASE64'", @@ -1764,7 +1764,7 @@ struct<> -- !query output org.apache.spark.SparkIllegalArgumentException { - "errorClass" : "CONVERSION_INVALID_INPUT", + "condition" : "CONVERSION_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "fmt" : "'BASE64'", @@ -1782,7 +1782,7 @@ struct<> -- !query output org.apache.spark.SparkIllegalArgumentException { - "errorClass" : "CONVERSION_INVALID_INPUT", + "condition" : "CONVERSION_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "fmt" : "'BASE64'", @@ -1889,7 +1889,7 @@ struct<> -- !query output org.apache.spark.SparkIllegalArgumentException { - "errorClass" : "CONVERSION_INVALID_INPUT", + "condition" : "CONVERSION_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "fmt" : "'HEX'", @@ -1907,7 +1907,7 @@ struct<> -- !query output org.apache.spark.SparkIllegalArgumentException { - "errorClass" : "CONVERSION_INVALID_INPUT", + "condition" : "CONVERSION_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "fmt" : "'HEX'", @@ -1933,7 +1933,7 @@ struct<> -- !query output org.apache.spark.SparkIllegalArgumentException { - "errorClass" : "CONVERSION_INVALID_INPUT", + "condition" : "CONVERSION_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "fmt" : "'BASE64'", @@ -1951,7 +1951,7 @@ struct<> -- !query output org.apache.spark.SparkIllegalArgumentException { - "errorClass" : "CONVERSION_INVALID_INPUT", + "condition" : "CONVERSION_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "fmt" : "'HEX'", @@ -2010,7 +2010,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.INVALID_ARG_VALUE", + "condition" : "DATATYPE_MISMATCH.INVALID_ARG_VALUE", "sqlState" : "42K09", "messageParameters" : { "inputName" : "fmt", @@ -2036,7 +2036,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.INVALID_ARG_VALUE", + "condition" : "DATATYPE_MISMATCH.INVALID_ARG_VALUE", "sqlState" : "42K09", "messageParameters" : { "inputName" : "fmt", @@ -2070,7 +2070,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.NON_FOLDABLE_INPUT", + "condition" : "DATATYPE_MISMATCH.NON_FOLDABLE_INPUT", "sqlState" : "42K09", "messageParameters" : { "inputExpr" : "\"fmtField\"", @@ -2311,7 +2311,7 @@ struct<> -- !query output org.apache.spark.SparkIllegalArgumentException { - "errorClass" : "INVALID_UTF8_STRING", + "condition" : "INVALID_UTF8_STRING", "sqlState" : "22029", "messageParameters" : { "str" : "\\x80" diff --git a/sql/core/src/test/resources/sql-tests/results/subquery/exists-subquery/exists-in-join-condition.sql.out b/sql/core/src/test/resources/sql-tests/results/subquery/exists-subquery/exists-in-join-condition.sql.out index c9c68a5f0602..10825fdbbde1 100644 --- a/sql/core/src/test/resources/sql-tests/results/subquery/exists-subquery/exists-in-join-condition.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/subquery/exists-subquery/exists-in-join-condition.sql.out @@ -481,7 +481,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "UNSUPPORTED_SUBQUERY_EXPRESSION_CATEGORY.UNSUPPORTED_CORRELATED_EXPRESSION_IN_JOIN_CONDITION", + "condition" : "UNSUPPORTED_SUBQUERY_EXPRESSION_CATEGORY.UNSUPPORTED_CORRELATED_EXPRESSION_IN_JOIN_CONDITION", "sqlState" : "0A000", "messageParameters" : { "subqueryExpression" : "exists(x.x2, y.y2, (z.z2 = x.x2), (z.z2 = y.y2))" @@ -496,7 +496,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "UNSUPPORTED_SUBQUERY_EXPRESSION_CATEGORY.UNSUPPORTED_CORRELATED_EXPRESSION_IN_JOIN_CONDITION", + "condition" : "UNSUPPORTED_SUBQUERY_EXPRESSION_CATEGORY.UNSUPPORTED_CORRELATED_EXPRESSION_IN_JOIN_CONDITION", "sqlState" : "0A000", "messageParameters" : { "subqueryExpression" : "exists(x.x2, y.y2, (z.z2 = x.x2), (z.z2 = y.y2))" diff --git a/sql/core/src/test/resources/sql-tests/results/subquery/in-subquery/in-basic.sql.out b/sql/core/src/test/resources/sql-tests/results/subquery/in-subquery/in-basic.sql.out index 3a07dbdbf00e..02c403922217 100644 --- a/sql/core/src/test/resources/sql-tests/results/subquery/in-subquery/in-basic.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/subquery/in-subquery/in-basic.sql.out @@ -39,7 +39,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.IN_SUBQUERY_LENGTH_MISMATCH", + "condition" : "DATATYPE_MISMATCH.IN_SUBQUERY_LENGTH_MISMATCH", "sqlState" : "42K09", "messageParameters" : { "leftColumns" : "\"a1\", \"b1\"", diff --git a/sql/core/src/test/resources/sql-tests/results/subquery/in-subquery/in-subquery-in-join-condition.sql.out b/sql/core/src/test/resources/sql-tests/results/subquery/in-subquery/in-subquery-in-join-condition.sql.out index 13af4c81173a..b7970c00a3b7 100644 --- a/sql/core/src/test/resources/sql-tests/results/subquery/in-subquery/in-subquery-in-join-condition.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/subquery/in-subquery/in-subquery-in-join-condition.sql.out @@ -443,7 +443,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "UNSUPPORTED_SUBQUERY_EXPRESSION_CATEGORY.UNSUPPORTED_CORRELATED_EXPRESSION_IN_JOIN_CONDITION", + "condition" : "UNSUPPORTED_SUBQUERY_EXPRESSION_CATEGORY.UNSUPPORTED_CORRELATED_EXPRESSION_IN_JOIN_CONDITION", "sqlState" : "0A000", "messageParameters" : { "subqueryExpression" : "(x.x2 IN (listquery(x.x2, y.y2, (z.z2 = x.x2), (z.z2 = y.y2))))" @@ -458,7 +458,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "UNSUPPORTED_SUBQUERY_EXPRESSION_CATEGORY.UNSUPPORTED_CORRELATED_EXPRESSION_IN_JOIN_CONDITION", + "condition" : "UNSUPPORTED_SUBQUERY_EXPRESSION_CATEGORY.UNSUPPORTED_CORRELATED_EXPRESSION_IN_JOIN_CONDITION", "sqlState" : "0A000", "messageParameters" : { "subqueryExpression" : "(x.x2 IN (listquery(x.x2, y.y2, (z.z2 = x.x2), (z.z2 = y.y2))))" diff --git a/sql/core/src/test/resources/sql-tests/results/subquery/negative-cases/invalid-correlation.sql.out b/sql/core/src/test/resources/sql-tests/results/subquery/negative-cases/invalid-correlation.sql.out index 93c60fd49c58..61d5d02f2c87 100644 --- a/sql/core/src/test/resources/sql-tests/results/subquery/negative-cases/invalid-correlation.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/subquery/negative-cases/invalid-correlation.sql.out @@ -44,7 +44,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "MISSING_GROUP_BY", + "condition" : "MISSING_GROUP_BY", "sqlState" : "42803", "queryContext" : [ { "objectType" : "", @@ -71,7 +71,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "MISSING_ATTRIBUTES.RESOLVED_ATTRIBUTE_MISSING_FROM_INPUT", + "condition" : "MISSING_ATTRIBUTES.RESOLVED_ATTRIBUTE_MISSING_FROM_INPUT", "sqlState" : "XX000", "messageParameters" : { "input" : "\"min(t2a)\", \"t2c\"", @@ -101,7 +101,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "UNSUPPORTED_SUBQUERY_EXPRESSION_CATEGORY.AGGREGATE_FUNCTION_MIXED_OUTER_LOCAL_REFERENCES", + "condition" : "UNSUPPORTED_SUBQUERY_EXPRESSION_CATEGORY.AGGREGATE_FUNCTION_MIXED_OUTER_LOCAL_REFERENCES", "sqlState" : "0A000", "messageParameters" : { "function" : "min((outer(t1.t1a) + t2.t2a))" @@ -130,7 +130,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "UNSUPPORTED_SUBQUERY_EXPRESSION_CATEGORY.AGGREGATE_FUNCTION_MIXED_OUTER_LOCAL_REFERENCES", + "condition" : "UNSUPPORTED_SUBQUERY_EXPRESSION_CATEGORY.AGGREGATE_FUNCTION_MIXED_OUTER_LOCAL_REFERENCES", "sqlState" : "0A000", "messageParameters" : { "function" : "min((outer(t2.t2a) + t3.t3a))" @@ -157,7 +157,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNSUPPORTED_SUBQUERY_EXPRESSION_CATEGORY.CORRELATED_REFERENCE", + "condition" : "UNSUPPORTED_SUBQUERY_EXPRESSION_CATEGORY.CORRELATED_REFERENCE", "sqlState" : "0A000", "messageParameters" : { "sqlExprs" : "\"min(t2a) AS `min(outer(t2.t2a))`\"" @@ -192,7 +192,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`t1a`", diff --git a/sql/core/src/test/resources/sql-tests/results/subquery/negative-cases/subq-input-typecheck.sql.out b/sql/core/src/test/resources/sql-tests/results/subquery/negative-cases/subq-input-typecheck.sql.out index 52338368f25d..f488fb2a10a1 100644 --- a/sql/core/src/test/resources/sql-tests/results/subquery/negative-cases/subq-input-typecheck.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/subquery/negative-cases/subq-input-typecheck.sql.out @@ -62,7 +62,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INVALID_SUBQUERY_EXPRESSION.SCALAR_SUBQUERY_RETURN_MORE_THAN_ONE_OUTPUT_COLUMN", + "condition" : "INVALID_SUBQUERY_EXPRESSION.SCALAR_SUBQUERY_RETURN_MORE_THAN_ONE_OUTPUT_COLUMN", "sqlState" : "42823", "messageParameters" : { "number" : "2" @@ -90,7 +90,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INVALID_SUBQUERY_EXPRESSION.SCALAR_SUBQUERY_RETURN_MORE_THAN_ONE_OUTPUT_COLUMN", + "condition" : "INVALID_SUBQUERY_EXPRESSION.SCALAR_SUBQUERY_RETURN_MORE_THAN_ONE_OUTPUT_COLUMN", "sqlState" : "42823", "messageParameters" : { "number" : "2" @@ -116,7 +116,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.IN_SUBQUERY_LENGTH_MISMATCH", + "condition" : "DATATYPE_MISMATCH.IN_SUBQUERY_LENGTH_MISMATCH", "sqlState" : "42K09", "messageParameters" : { "leftColumns" : "\"t1a\"", @@ -146,7 +146,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.IN_SUBQUERY_LENGTH_MISMATCH", + "condition" : "DATATYPE_MISMATCH.IN_SUBQUERY_LENGTH_MISMATCH", "sqlState" : "42K09", "messageParameters" : { "leftColumns" : "\"t1a\", \"t1b\"", @@ -177,7 +177,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.IN_SUBQUERY_DATA_TYPE_MISMATCH", + "condition" : "DATATYPE_MISMATCH.IN_SUBQUERY_DATA_TYPE_MISMATCH", "sqlState" : "42K09", "messageParameters" : { "leftType" : "\"DOUBLE\", \"STRING\", \"STRING\"", diff --git a/sql/core/src/test/resources/sql-tests/results/subquery/scalar-subquery/scalar-subquery-group-by.sql.out b/sql/core/src/test/resources/sql-tests/results/subquery/scalar-subquery/scalar-subquery-group-by.sql.out index 56932edd4e54..bda13ce74dff 100644 --- a/sql/core/src/test/resources/sql-tests/results/subquery/scalar-subquery/scalar-subquery-group-by.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/subquery/scalar-subquery/scalar-subquery-group-by.sql.out @@ -38,7 +38,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNSUPPORTED_SUBQUERY_EXPRESSION_CATEGORY.CORRELATED_REFERENCE", + "condition" : "UNSUPPORTED_SUBQUERY_EXPRESSION_CATEGORY.CORRELATED_REFERENCE", "sqlState" : "0A000", "messageParameters" : { "sqlExprs" : "\"x1\"" @@ -60,7 +60,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNSUPPORTED_SUBQUERY_EXPRESSION_CATEGORY.CORRELATED_REFERENCE", + "condition" : "UNSUPPORTED_SUBQUERY_EXPRESSION_CATEGORY.CORRELATED_REFERENCE", "sqlState" : "0A000", "messageParameters" : { "sqlExprs" : "\"x1\"" @@ -127,7 +127,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNSUPPORTED_SUBQUERY_EXPRESSION_CATEGORY.NON_CORRELATED_COLUMNS_IN_GROUP_BY", + "condition" : "UNSUPPORTED_SUBQUERY_EXPRESSION_CATEGORY.NON_CORRELATED_COLUMNS_IN_GROUP_BY", "sqlState" : "0A000", "messageParameters" : { "value" : "y1" @@ -149,7 +149,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNSUPPORTED_SUBQUERY_EXPRESSION_CATEGORY.NON_CORRELATED_COLUMNS_IN_GROUP_BY", + "condition" : "UNSUPPORTED_SUBQUERY_EXPRESSION_CATEGORY.NON_CORRELATED_COLUMNS_IN_GROUP_BY", "sqlState" : "0A000", "messageParameters" : { "value" : "y1" @@ -171,7 +171,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNSUPPORTED_SUBQUERY_EXPRESSION_CATEGORY.NON_CORRELATED_COLUMNS_IN_GROUP_BY", + "condition" : "UNSUPPORTED_SUBQUERY_EXPRESSION_CATEGORY.NON_CORRELATED_COLUMNS_IN_GROUP_BY", "sqlState" : "0A000", "messageParameters" : { "value" : "y2" @@ -201,7 +201,7 @@ struct<> -- !query output org.apache.spark.SparkRuntimeException { - "errorClass" : "SCALAR_SUBQUERY_TOO_MANY_ROWS", + "condition" : "SCALAR_SUBQUERY_TOO_MANY_ROWS", "sqlState" : "21000" } @@ -213,7 +213,7 @@ struct<> -- !query output org.apache.spark.SparkRuntimeException { - "errorClass" : "SCALAR_SUBQUERY_TOO_MANY_ROWS", + "condition" : "SCALAR_SUBQUERY_TOO_MANY_ROWS", "sqlState" : "21000" } @@ -234,7 +234,7 @@ struct<> -- !query output org.apache.spark.SparkRuntimeException { - "errorClass" : "SCALAR_SUBQUERY_TOO_MANY_ROWS", + "condition" : "SCALAR_SUBQUERY_TOO_MANY_ROWS", "sqlState" : "21000" } @@ -246,7 +246,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNSUPPORTED_SUBQUERY_EXPRESSION_CATEGORY.ACCESSING_OUTER_QUERY_COLUMN_IS_NOT_ALLOWED", + "condition" : "UNSUPPORTED_SUBQUERY_EXPRESSION_CATEGORY.ACCESSING_OUTER_QUERY_COLUMN_IS_NOT_ALLOWED", "sqlState" : "0A000", "messageParameters" : { "treeNode" : "Filter (z1#x = outer(x1#x))\n+- SubqueryAlias z\n +- View (`z`, [z1#x, z2#x])\n +- Project [cast(col1#x as int) AS z1#x, cast(col2#x as int) AS z2#x]\n +- LocalRelation [col1#x, col2#x]\n" diff --git a/sql/core/src/test/resources/sql-tests/results/subquery/scalar-subquery/scalar-subquery-predicate.sql.out b/sql/core/src/test/resources/sql-tests/results/subquery/scalar-subquery/scalar-subquery-predicate.sql.out index 2460c2452ea5..6e78b0a63a58 100644 --- a/sql/core/src/test/resources/sql-tests/results/subquery/scalar-subquery/scalar-subquery-predicate.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/subquery/scalar-subquery/scalar-subquery-predicate.sql.out @@ -801,7 +801,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNSUPPORTED_SUBQUERY_EXPRESSION_CATEGORY.CORRELATED_REFERENCE", + "condition" : "UNSUPPORTED_SUBQUERY_EXPRESSION_CATEGORY.CORRELATED_REFERENCE", "sqlState" : "0A000", "messageParameters" : { "sqlExprs" : "\"sum(t0a) AS d\"" diff --git a/sql/core/src/test/resources/sql-tests/results/subquery/scalar-subquery/scalar-subquery-select.sql.out b/sql/core/src/test/resources/sql-tests/results/subquery/scalar-subquery/scalar-subquery-select.sql.out index 85bd9137602a..2642a4125488 100644 --- a/sql/core/src/test/resources/sql-tests/results/subquery/scalar-subquery/scalar-subquery-select.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/subquery/scalar-subquery/scalar-subquery-select.sql.out @@ -424,7 +424,7 @@ struct<> -- !query output org.apache.spark.SparkException { - "errorClass" : "SCALAR_SUBQUERY_TOO_MANY_ROWS", + "condition" : "SCALAR_SUBQUERY_TOO_MANY_ROWS", "sqlState" : "21000", "queryContext" : [ { "objectType" : "", diff --git a/sql/core/src/test/resources/sql-tests/results/subquery/scalar-subquery/scalar-subquery-set-op.sql.out b/sql/core/src/test/resources/sql-tests/results/subquery/scalar-subquery/scalar-subquery-set-op.sql.out index 33a57a73be08..9969fe758bec 100644 --- a/sql/core/src/test/resources/sql-tests/results/subquery/scalar-subquery/scalar-subquery-set-op.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/subquery/scalar-subquery/scalar-subquery-set-op.sql.out @@ -178,7 +178,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNSUPPORTED_SUBQUERY_EXPRESSION_CATEGORY.CORRELATED_REFERENCE", + "condition" : "UNSUPPORTED_SUBQUERY_EXPRESSION_CATEGORY.CORRELATED_REFERENCE", "sqlState" : "0A000", "messageParameters" : { "sqlExprs" : "\"sum(t0a) AS d\"" @@ -348,7 +348,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNSUPPORTED_SUBQUERY_EXPRESSION_CATEGORY.CORRELATED_REFERENCE", + "condition" : "UNSUPPORTED_SUBQUERY_EXPRESSION_CATEGORY.CORRELATED_REFERENCE", "sqlState" : "0A000", "messageParameters" : { "sqlExprs" : "\"sum(t0a) AS d\"" @@ -518,7 +518,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNSUPPORTED_SUBQUERY_EXPRESSION_CATEGORY.CORRELATED_REFERENCE", + "condition" : "UNSUPPORTED_SUBQUERY_EXPRESSION_CATEGORY.CORRELATED_REFERENCE", "sqlState" : "0A000", "messageParameters" : { "sqlExprs" : "\"sum(t0a) AS d\"" @@ -688,7 +688,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNSUPPORTED_SUBQUERY_EXPRESSION_CATEGORY.CORRELATED_REFERENCE", + "condition" : "UNSUPPORTED_SUBQUERY_EXPRESSION_CATEGORY.CORRELATED_REFERENCE", "sqlState" : "0A000", "messageParameters" : { "sqlExprs" : "\"sum(t0a) AS d\"" @@ -858,7 +858,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNSUPPORTED_SUBQUERY_EXPRESSION_CATEGORY.CORRELATED_REFERENCE", + "condition" : "UNSUPPORTED_SUBQUERY_EXPRESSION_CATEGORY.CORRELATED_REFERENCE", "sqlState" : "0A000", "messageParameters" : { "sqlExprs" : "\"sum(t0a) AS d\"" @@ -1028,7 +1028,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNSUPPORTED_SUBQUERY_EXPRESSION_CATEGORY.CORRELATED_REFERENCE", + "condition" : "UNSUPPORTED_SUBQUERY_EXPRESSION_CATEGORY.CORRELATED_REFERENCE", "sqlState" : "0A000", "messageParameters" : { "sqlExprs" : "\"sum(t0a) AS d\"" diff --git a/sql/core/src/test/resources/sql-tests/results/subquery/subquery-nested-data.sql.out b/sql/core/src/test/resources/sql-tests/results/subquery/subquery-nested-data.sql.out index e0fd040ca4bc..92c07626cf62 100644 --- a/sql/core/src/test/resources/sql-tests/results/subquery/subquery-nested-data.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/subquery/subquery-nested-data.sql.out @@ -197,7 +197,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNSUPPORTED_SUBQUERY_EXPRESSION_CATEGORY.CORRELATED_REFERENCE", + "condition" : "UNSUPPORTED_SUBQUERY_EXPRESSION_CATEGORY.CORRELATED_REFERENCE", "sqlState" : "0A000", "messageParameters" : { "sqlExprs" : "\"xm[1]\",\"xm[1] AS `outer(spark_catalog.default.x.xm)[1]`\",\"(xm[1] - sum(ym[1])) AS `(outer(spark_catalog.default.x.xm)[1] - sum(ym[1]))`\"" @@ -243,7 +243,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "UNSUPPORTED_SUBQUERY_EXPRESSION_CATEGORY.UNSUPPORTED_CORRELATED_REFERENCE_DATA_TYPE", + "condition" : "UNSUPPORTED_SUBQUERY_EXPRESSION_CATEGORY.UNSUPPORTED_CORRELATED_REFERENCE_DATA_TYPE", "sqlState" : "0A000", "messageParameters" : { "dataType" : "map", @@ -274,7 +274,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "UNSUPPORTED_SUBQUERY_EXPRESSION_CATEGORY.UNSUPPORTED_CORRELATED_REFERENCE_DATA_TYPE", + "condition" : "UNSUPPORTED_SUBQUERY_EXPRESSION_CATEGORY.UNSUPPORTED_CORRELATED_REFERENCE_DATA_TYPE", "sqlState" : "0A000", "messageParameters" : { "dataType" : "map", diff --git a/sql/core/src/test/resources/sql-tests/results/table-aliases.sql.out b/sql/core/src/test/resources/sql-tests/results/table-aliases.sql.out index 1878452d1c20..6fe77cd06225 100644 --- a/sql/core/src/test/resources/sql-tests/results/table-aliases.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/table-aliases.sql.out @@ -40,7 +40,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "ASSIGNMENT_ARITY_MISMATCH", + "condition" : "ASSIGNMENT_ARITY_MISMATCH", "sqlState" : "42802", "messageParameters" : { "numExpr" : "3", @@ -63,7 +63,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "ASSIGNMENT_ARITY_MISMATCH", + "condition" : "ASSIGNMENT_ARITY_MISMATCH", "sqlState" : "42802", "messageParameters" : { "numExpr" : "1", @@ -86,7 +86,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`a`", @@ -177,7 +177,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "CANNOT_RESOLVE_STAR_EXPAND", + "condition" : "CANNOT_RESOLVE_STAR_EXPAND", "sqlState" : "42704", "messageParameters" : { "columns" : "`id`, `v1`", @@ -200,7 +200,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`src1`.`id`", diff --git a/sql/core/src/test/resources/sql-tests/results/table-valued-functions.sql.out b/sql/core/src/test/resources/sql-tests/results/table-valued-functions.sql.out index 768f0e8c010e..282faa8c8e5a 100644 --- a/sql/core/src/test/resources/sql-tests/results/table-valued-functions.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/table-valued-functions.sql.out @@ -6,7 +6,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "UNRESOLVABLE_TABLE_VALUED_FUNCTION", + "condition" : "UNRESOLVABLE_TABLE_VALUED_FUNCTION", "sqlState" : "42883", "messageParameters" : { "name" : "`dummy`" @@ -81,7 +81,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", + "condition" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", "sqlState" : "42605", "messageParameters" : { "actualNum" : "5", @@ -106,7 +106,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "UNEXPECTED_INPUT_TYPE", + "condition" : "UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "functionName" : "`range`", @@ -132,7 +132,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "UNEXPECTED_INPUT_TYPE", + "condition" : "UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "functionName" : "`range`", @@ -158,7 +158,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "FAILED_FUNCTION_CALL", + "condition" : "FAILED_FUNCTION_CALL", "sqlState" : "38000", "messageParameters" : { "funcName" : "`range`" @@ -198,7 +198,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "NON_FOLDABLE_ARGUMENT", + "condition" : "NON_FOLDABLE_ARGUMENT", "sqlState" : "42K08", "messageParameters" : { "funcName" : "`range`", @@ -222,7 +222,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "NON_FOLDABLE_ARGUMENT", + "condition" : "NON_FOLDABLE_ARGUMENT", "sqlState" : "42K08", "messageParameters" : { "funcName" : "`range`", @@ -306,7 +306,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"NULL\"", @@ -332,7 +332,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"NULL\"", @@ -358,7 +358,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"1\"", @@ -384,7 +384,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", + "condition" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", "sqlState" : "42605", "messageParameters" : { "actualNum" : "2", @@ -409,7 +409,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "UNSUPPORTED_GENERATOR.NESTED_IN_EXPRESSIONS", + "condition" : "UNSUPPORTED_GENERATOR.NESTED_IN_EXPRESSIONS", "sqlState" : "42K0E", "messageParameters" : { "expression" : "\"explode(explode(array(1)))\"" @@ -431,7 +431,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "NUM_TABLE_VALUE_ALIASES_MISMATCH", + "condition" : "NUM_TABLE_VALUE_ALIASES_MISMATCH", "sqlState" : "42826", "messageParameters" : { "aliasesNum" : "2", @@ -535,7 +535,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"NULL\"", @@ -561,7 +561,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "NUM_TABLE_VALUE_ALIASES_MISMATCH", + "condition" : "NUM_TABLE_VALUE_ALIASES_MISMATCH", "sqlState" : "42826", "messageParameters" : { "aliasesNum" : "3", @@ -654,7 +654,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"1\"", @@ -680,7 +680,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", + "condition" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", "sqlState" : "42605", "messageParameters" : { "actualNum" : "2", @@ -705,7 +705,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "UNSUPPORTED_GENERATOR.NESTED_IN_EXPRESSIONS", + "condition" : "UNSUPPORTED_GENERATOR.NESTED_IN_EXPRESSIONS", "sqlState" : "42K0E", "messageParameters" : { "expression" : "\"posexplode(explode(array(1)))\"" @@ -727,7 +727,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "NUM_TABLE_VALUE_ALIASES_MISMATCH", + "condition" : "NUM_TABLE_VALUE_ALIASES_MISMATCH", "sqlState" : "42826", "messageParameters" : { "aliasesNum" : "1", @@ -825,7 +825,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", + "condition" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", "sqlState" : "42605", "messageParameters" : { "actualNum" : "0", @@ -843,7 +843,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", + "condition" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", "sqlState" : "42605", "messageParameters" : { "actualNum" : "1", @@ -861,7 +861,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.NON_STRING_TYPE", + "condition" : "DATATYPE_MISMATCH.NON_STRING_TYPE", "sqlState" : "42K09", "messageParameters" : { "funcName" : "`json_tuple`", @@ -884,7 +884,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.NON_STRING_TYPE", + "condition" : "DATATYPE_MISMATCH.NON_STRING_TYPE", "sqlState" : "42K09", "messageParameters" : { "funcName" : "`json_tuple`", @@ -907,7 +907,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "NUM_TABLE_VALUE_ALIASES_MISMATCH", + "condition" : "NUM_TABLE_VALUE_ALIASES_MISMATCH", "sqlState" : "42826", "messageParameters" : { "aliasesNum" : "1", @@ -987,7 +987,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", + "condition" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", "sqlState" : "42605", "messageParameters" : { "actualNum" : "0", @@ -1005,7 +1005,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "NUM_TABLE_VALUE_ALIASES_MISMATCH", + "condition" : "NUM_TABLE_VALUE_ALIASES_MISMATCH", "sqlState" : "42826", "messageParameters" : { "aliasesNum" : "3", @@ -1029,7 +1029,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.STACK_COLUMN_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.STACK_COLUMN_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "columnIndex" : "1", @@ -1056,7 +1056,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "UNSUPPORTED_GENERATOR.NESTED_IN_EXPRESSIONS", + "condition" : "UNSUPPORTED_GENERATOR.NESTED_IN_EXPRESSIONS", "sqlState" : "42K0E", "messageParameters" : { "expression" : "\"stack(2, explode(array(1, 2, 3)))\"" diff --git a/sql/core/src/test/resources/sql-tests/results/tablesample-negative.sql.out b/sql/core/src/test/resources/sql-tests/results/tablesample-negative.sql.out index fd852992397d..81050b43d4bf 100644 --- a/sql/core/src/test/resources/sql-tests/results/tablesample-negative.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/tablesample-negative.sql.out @@ -30,7 +30,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0064", + "condition" : "_LEGACY_ERROR_TEMP_0064", "messageParameters" : { "msg" : "Sampling fraction (-0.01) must be on interval [0, 1]" }, @@ -51,7 +51,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0064", + "condition" : "_LEGACY_ERROR_TEMP_0064", "messageParameters" : { "msg" : "Sampling fraction (1.01) must be on interval [0, 1]" }, diff --git a/sql/core/src/test/resources/sql-tests/results/timestamp-ltz.sql.out b/sql/core/src/test/resources/sql-tests/results/timestamp-ltz.sql.out index 963505615225..346da1791774 100644 --- a/sql/core/src/test/resources/sql-tests/results/timestamp-ltz.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/timestamp-ltz.sql.out @@ -54,7 +54,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "INVALID_FRACTION_OF_SECOND", + "condition" : "INVALID_FRACTION_OF_SECOND", "sqlState" : "22023", "messageParameters" : { "secAndMicros" : "60.007" diff --git a/sql/core/src/test/resources/sql-tests/results/timestamp-ntz.sql.out b/sql/core/src/test/resources/sql-tests/results/timestamp-ntz.sql.out index 3a473dad828a..0213ca73f609 100644 --- a/sql/core/src/test/resources/sql-tests/results/timestamp-ntz.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/timestamp-ntz.sql.out @@ -46,7 +46,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", + "condition" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", "sqlState" : "42605", "messageParameters" : { "actualNum" : "7", @@ -71,7 +71,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "INVALID_FRACTION_OF_SECOND", + "condition" : "INVALID_FRACTION_OF_SECOND", "sqlState" : "22023", "messageParameters" : { "secAndMicros" : "60.007" diff --git a/sql/core/src/test/resources/sql-tests/results/timestamp.sql.out b/sql/core/src/test/resources/sql-tests/results/timestamp.sql.out index 06a8fbad8a29..432f77c93f55 100644 --- a/sql/core/src/test/resources/sql-tests/results/timestamp.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/timestamp.sql.out @@ -14,7 +14,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_TYPED_LITERAL", + "condition" : "INVALID_TYPED_LITERAL", "sqlState" : "42604", "messageParameters" : { "value" : "'2019-01-01中文'", @@ -37,7 +37,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_TYPED_LITERAL", + "condition" : "INVALID_TYPED_LITERAL", "sqlState" : "42604", "messageParameters" : { "value" : "'4294967297'", @@ -60,7 +60,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_TYPED_LITERAL", + "condition" : "INVALID_TYPED_LITERAL", "sqlState" : "42604", "messageParameters" : { "value" : "'2021-01-01T12:30:4294967297.123456'", @@ -123,7 +123,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "INVALID_FRACTION_OF_SECOND", + "condition" : "INVALID_FRACTION_OF_SECOND", "sqlState" : "22023", "messageParameters" : { "secAndMicros" : "60.007" @@ -154,7 +154,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "DATETIME_FIELD_OUT_OF_BOUNDS", + "condition" : "DATETIME_FIELD_OUT_OF_BOUNDS", "sqlState" : "22023", "messageParameters" : { "ansiConfig" : "\"spark.sql.ansi.enabled\"", @@ -186,7 +186,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "DATETIME_FIELD_OUT_OF_BOUNDS", + "condition" : "DATETIME_FIELD_OUT_OF_BOUNDS", "sqlState" : "22023", "messageParameters" : { "ansiConfig" : "\"spark.sql.ansi.enabled\"", @@ -202,7 +202,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "DATETIME_FIELD_OUT_OF_BOUNDS", + "condition" : "DATETIME_FIELD_OUT_OF_BOUNDS", "sqlState" : "22023", "messageParameters" : { "ansiConfig" : "\"spark.sql.ansi.enabled\"", @@ -392,7 +392,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "CANNOT_PARSE_TIMESTAMP", + "condition" : "CANNOT_PARSE_TIMESTAMP", "sqlState" : "22007", "messageParameters" : { "func" : "`try_to_timestamp`", @@ -464,7 +464,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "CANNOT_PARSE_TIMESTAMP", + "condition" : "CANNOT_PARSE_TIMESTAMP", "sqlState" : "22007", "messageParameters" : { "func" : "`try_to_timestamp`", @@ -488,7 +488,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "CANNOT_PARSE_TIMESTAMP", + "condition" : "CANNOT_PARSE_TIMESTAMP", "sqlState" : "22007", "messageParameters" : { "func" : "`try_to_timestamp`", @@ -560,7 +560,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "CANNOT_PARSE_TIMESTAMP", + "condition" : "CANNOT_PARSE_TIMESTAMP", "sqlState" : "22007", "messageParameters" : { "func" : "`try_to_timestamp`", @@ -576,7 +576,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "CANNOT_PARSE_TIMESTAMP", + "condition" : "CANNOT_PARSE_TIMESTAMP", "sqlState" : "22007", "messageParameters" : { "func" : "`try_to_timestamp`", @@ -656,7 +656,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "CANNOT_PARSE_TIMESTAMP", + "condition" : "CANNOT_PARSE_TIMESTAMP", "sqlState" : "22007", "messageParameters" : { "func" : "`try_to_timestamp`", @@ -768,7 +768,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"TIMESTAMP\"", @@ -792,7 +792,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"TIMESTAMP\"", @@ -816,7 +816,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP\"", @@ -840,7 +840,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"VOID\"", @@ -887,7 +887,7 @@ struct<> -- !query output org.apache.spark.SparkUpgradeException { - "errorClass" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION.DATETIME_PATTERN_RECOGNITION", + "condition" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION.DATETIME_PATTERN_RECOGNITION", "sqlState" : "42K0B", "messageParameters" : { "config" : "\"spark.sql.legacy.timeParserPolicy\"", @@ -904,7 +904,7 @@ struct<> -- !query output org.apache.spark.SparkUpgradeException { - "errorClass" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION.DATETIME_PATTERN_RECOGNITION", + "condition" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION.DATETIME_PATTERN_RECOGNITION", "sqlState" : "42K0B", "messageParameters" : { "config" : "\"spark.sql.legacy.timeParserPolicy\"", @@ -921,7 +921,7 @@ struct<> -- !query output org.apache.spark.SparkUpgradeException { - "errorClass" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION.DATETIME_PATTERN_RECOGNITION", + "condition" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION.DATETIME_PATTERN_RECOGNITION", "sqlState" : "42K0B", "messageParameters" : { "config" : "\"spark.sql.legacy.timeParserPolicy\"", @@ -938,7 +938,7 @@ struct<> -- !query output org.apache.spark.SparkUpgradeException { - "errorClass" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION.DATETIME_PATTERN_RECOGNITION", + "condition" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION.DATETIME_PATTERN_RECOGNITION", "sqlState" : "42K0B", "messageParameters" : { "config" : "\"spark.sql.legacy.timeParserPolicy\"", @@ -955,7 +955,7 @@ struct<> -- !query output org.apache.spark.SparkUpgradeException { - "errorClass" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION.DATETIME_PATTERN_RECOGNITION", + "condition" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION.DATETIME_PATTERN_RECOGNITION", "sqlState" : "42K0B", "messageParameters" : { "config" : "\"spark.sql.legacy.timeParserPolicy\"", @@ -972,7 +972,7 @@ struct<> -- !query output org.apache.spark.SparkUpgradeException { - "errorClass" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION.DATETIME_PATTERN_RECOGNITION", + "condition" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION.DATETIME_PATTERN_RECOGNITION", "sqlState" : "42K0B", "messageParameters" : { "config" : "\"spark.sql.legacy.timeParserPolicy\"", @@ -1021,7 +1021,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", + "condition" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", "sqlState" : "22023", "messageParameters" : { "functionName" : "`timestampadd`", @@ -1045,7 +1045,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", + "condition" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", "sqlState" : "22023", "messageParameters" : { "functionName" : "`timestampadd`", @@ -1101,7 +1101,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", + "condition" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", "sqlState" : "22023", "messageParameters" : { "functionName" : "`timestampdiff`", @@ -1125,7 +1125,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", + "condition" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", "sqlState" : "22023", "messageParameters" : { "functionName" : "`timestampdiff`", @@ -1181,7 +1181,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", + "condition" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", "sqlState" : "22023", "messageParameters" : { "functionName" : "`timediff`", @@ -1205,7 +1205,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", + "condition" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", "sqlState" : "22023", "messageParameters" : { "functionName" : "`timediff`", diff --git a/sql/core/src/test/resources/sql-tests/results/timestampNTZ/timestamp-ansi.sql.out b/sql/core/src/test/resources/sql-tests/results/timestampNTZ/timestamp-ansi.sql.out index 933982f5cff6..b503287804bd 100644 --- a/sql/core/src/test/resources/sql-tests/results/timestampNTZ/timestamp-ansi.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/timestampNTZ/timestamp-ansi.sql.out @@ -14,7 +14,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_TYPED_LITERAL", + "condition" : "INVALID_TYPED_LITERAL", "sqlState" : "42604", "messageParameters" : { "value" : "'2019-01-01中文'", @@ -37,7 +37,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_TYPED_LITERAL", + "condition" : "INVALID_TYPED_LITERAL", "sqlState" : "42604", "messageParameters" : { "value" : "'4294967297'", @@ -60,7 +60,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_TYPED_LITERAL", + "condition" : "INVALID_TYPED_LITERAL", "sqlState" : "42604", "messageParameters" : { "value" : "'2021-01-01T12:30:4294967297.123456'", @@ -123,7 +123,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "INVALID_FRACTION_OF_SECOND", + "condition" : "INVALID_FRACTION_OF_SECOND", "sqlState" : "22023", "messageParameters" : { "secAndMicros" : "60.007" @@ -154,7 +154,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "DATETIME_FIELD_OUT_OF_BOUNDS", + "condition" : "DATETIME_FIELD_OUT_OF_BOUNDS", "sqlState" : "22023", "messageParameters" : { "ansiConfig" : "\"spark.sql.ansi.enabled\"", @@ -186,7 +186,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "DATETIME_FIELD_OUT_OF_BOUNDS", + "condition" : "DATETIME_FIELD_OUT_OF_BOUNDS", "sqlState" : "22023", "messageParameters" : { "ansiConfig" : "\"spark.sql.ansi.enabled\"", @@ -202,7 +202,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "DATETIME_FIELD_OUT_OF_BOUNDS", + "condition" : "DATETIME_FIELD_OUT_OF_BOUNDS", "sqlState" : "22023", "messageParameters" : { "ansiConfig" : "\"spark.sql.ansi.enabled\"", @@ -384,7 +384,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'1'", @@ -406,7 +406,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "CANNOT_PARSE_TIMESTAMP", + "condition" : "CANNOT_PARSE_TIMESTAMP", "sqlState" : "22007", "messageParameters" : { "func" : "`try_to_timestamp`", @@ -478,7 +478,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "CANNOT_PARSE_TIMESTAMP", + "condition" : "CANNOT_PARSE_TIMESTAMP", "sqlState" : "22007", "messageParameters" : { "func" : "`try_to_timestamp`", @@ -502,7 +502,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "CANNOT_PARSE_TIMESTAMP", + "condition" : "CANNOT_PARSE_TIMESTAMP", "sqlState" : "22007", "messageParameters" : { "func" : "`try_to_timestamp`", @@ -574,7 +574,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "CANNOT_PARSE_TIMESTAMP", + "condition" : "CANNOT_PARSE_TIMESTAMP", "sqlState" : "22007", "messageParameters" : { "func" : "`try_to_timestamp`", @@ -590,7 +590,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "CANNOT_PARSE_TIMESTAMP", + "condition" : "CANNOT_PARSE_TIMESTAMP", "sqlState" : "22007", "messageParameters" : { "func" : "`try_to_timestamp`", @@ -670,7 +670,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "CANNOT_PARSE_TIMESTAMP", + "condition" : "CANNOT_PARSE_TIMESTAMP", "sqlState" : "22007", "messageParameters" : { "func" : "`try_to_timestamp`", @@ -782,7 +782,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"TIMESTAMP_NTZ\"", @@ -806,7 +806,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"TIMESTAMP_NTZ\"", @@ -830,7 +830,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP_NTZ\"", @@ -854,7 +854,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"VOID\"", @@ -901,7 +901,7 @@ struct<> -- !query output org.apache.spark.SparkRuntimeException { - "errorClass" : "_LEGACY_ERROR_TEMP_2130", + "condition" : "_LEGACY_ERROR_TEMP_2130", "messageParameters" : { "docroot" : "https://spark.apache.org/docs/latest", "pattern" : "'yyyy-MM-dd GGGGG'" @@ -916,7 +916,7 @@ struct<> -- !query output org.apache.spark.SparkRuntimeException { - "errorClass" : "_LEGACY_ERROR_TEMP_2130", + "condition" : "_LEGACY_ERROR_TEMP_2130", "messageParameters" : { "docroot" : "https://spark.apache.org/docs/latest", "pattern" : "'dd MM yyyy EEEEEE'" @@ -931,7 +931,7 @@ struct<> -- !query output org.apache.spark.SparkRuntimeException { - "errorClass" : "_LEGACY_ERROR_TEMP_2130", + "condition" : "_LEGACY_ERROR_TEMP_2130", "messageParameters" : { "docroot" : "https://spark.apache.org/docs/latest", "pattern" : "'dd MM yyyy EEEEE'" @@ -946,7 +946,7 @@ struct<> -- !query output org.apache.spark.SparkUpgradeException { - "errorClass" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION.DATETIME_PATTERN_RECOGNITION", + "condition" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION.DATETIME_PATTERN_RECOGNITION", "sqlState" : "42K0B", "messageParameters" : { "config" : "\"spark.sql.legacy.timeParserPolicy\"", @@ -1011,7 +1011,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", + "condition" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", "sqlState" : "22023", "messageParameters" : { "functionName" : "`timestampadd`", @@ -1035,7 +1035,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", + "condition" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", "sqlState" : "22023", "messageParameters" : { "functionName" : "`timestampadd`", @@ -1091,7 +1091,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", + "condition" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", "sqlState" : "22023", "messageParameters" : { "functionName" : "`timestampdiff`", @@ -1115,7 +1115,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", + "condition" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", "sqlState" : "22023", "messageParameters" : { "functionName" : "`timestampdiff`", @@ -1171,7 +1171,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", + "condition" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", "sqlState" : "22023", "messageParameters" : { "functionName" : "`timediff`", @@ -1195,7 +1195,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", + "condition" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", "sqlState" : "22023", "messageParameters" : { "functionName" : "`timediff`", diff --git a/sql/core/src/test/resources/sql-tests/results/timestampNTZ/timestamp.sql.out b/sql/core/src/test/resources/sql-tests/results/timestampNTZ/timestamp.sql.out index 102935c90f9c..4766e760ced5 100644 --- a/sql/core/src/test/resources/sql-tests/results/timestampNTZ/timestamp.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/timestampNTZ/timestamp.sql.out @@ -14,7 +14,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_TYPED_LITERAL", + "condition" : "INVALID_TYPED_LITERAL", "sqlState" : "42604", "messageParameters" : { "value" : "'2019-01-01中文'", @@ -37,7 +37,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_TYPED_LITERAL", + "condition" : "INVALID_TYPED_LITERAL", "sqlState" : "42604", "messageParameters" : { "value" : "'4294967297'", @@ -60,7 +60,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_TYPED_LITERAL", + "condition" : "INVALID_TYPED_LITERAL", "sqlState" : "42604", "messageParameters" : { "value" : "'2021-01-01T12:30:4294967297.123456'", @@ -633,7 +633,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"2011-11-11 11:11:10\"", @@ -659,7 +659,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"2011-11-11 11:11:11\"", @@ -709,7 +709,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"str\"", @@ -735,7 +735,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"str\"", @@ -761,7 +761,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP_NTZ\"", @@ -785,7 +785,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DOUBLE\"", @@ -809,7 +809,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP_NTZ\"", @@ -833,7 +833,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"VOID\"", @@ -880,7 +880,7 @@ struct<> -- !query output org.apache.spark.SparkRuntimeException { - "errorClass" : "_LEGACY_ERROR_TEMP_2130", + "condition" : "_LEGACY_ERROR_TEMP_2130", "messageParameters" : { "docroot" : "https://spark.apache.org/docs/latest", "pattern" : "'yyyy-MM-dd GGGGG'" @@ -895,7 +895,7 @@ struct<> -- !query output org.apache.spark.SparkRuntimeException { - "errorClass" : "_LEGACY_ERROR_TEMP_2130", + "condition" : "_LEGACY_ERROR_TEMP_2130", "messageParameters" : { "docroot" : "https://spark.apache.org/docs/latest", "pattern" : "'dd MM yyyy EEEEEE'" @@ -910,7 +910,7 @@ struct<> -- !query output org.apache.spark.SparkRuntimeException { - "errorClass" : "_LEGACY_ERROR_TEMP_2130", + "condition" : "_LEGACY_ERROR_TEMP_2130", "messageParameters" : { "docroot" : "https://spark.apache.org/docs/latest", "pattern" : "'dd MM yyyy EEEEE'" @@ -925,7 +925,7 @@ struct<> -- !query output org.apache.spark.SparkUpgradeException { - "errorClass" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION.DATETIME_PATTERN_RECOGNITION", + "condition" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION.DATETIME_PATTERN_RECOGNITION", "sqlState" : "42K0B", "messageParameters" : { "config" : "\"spark.sql.legacy.timeParserPolicy\"", @@ -990,7 +990,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", + "condition" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", "sqlState" : "22023", "messageParameters" : { "functionName" : "`timestampadd`", @@ -1014,7 +1014,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", + "condition" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", "sqlState" : "22023", "messageParameters" : { "functionName" : "`timestampadd`", @@ -1070,7 +1070,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", + "condition" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", "sqlState" : "22023", "messageParameters" : { "functionName" : "`timestampdiff`", @@ -1094,7 +1094,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", + "condition" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", "sqlState" : "22023", "messageParameters" : { "functionName" : "`timestampdiff`", @@ -1150,7 +1150,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", + "condition" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", "sqlState" : "22023", "messageParameters" : { "functionName" : "`timediff`", @@ -1174,7 +1174,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", + "condition" : "INVALID_PARAMETER_VALUE.DATETIME_UNIT", "sqlState" : "22023", "messageParameters" : { "functionName" : "`timediff`", diff --git a/sql/core/src/test/resources/sql-tests/results/timezone.sql.out b/sql/core/src/test/resources/sql-tests/results/timezone.sql.out index 5f0fdef50e3d..4addda8eaf50 100644 --- a/sql/core/src/test/resources/sql-tests/results/timezone.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/timezone.sql.out @@ -46,7 +46,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0045", + "condition" : "_LEGACY_ERROR_TEMP_0045", "queryContext" : [ { "objectType" : "", "objectName" : "", @@ -64,7 +64,7 @@ struct<> -- !query output org.apache.spark.SparkIllegalArgumentException { - "errorClass" : "INVALID_CONF_VALUE.TIME_ZONE", + "condition" : "INVALID_CONF_VALUE.TIME_ZONE", "sqlState" : "22022", "messageParameters" : { "confName" : "spark.sql.session.timeZone", @@ -80,7 +80,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_INTERVAL_FORMAT.TIMEZONE_INTERVAL_OUT_OF_RANGE", + "condition" : "INVALID_INTERVAL_FORMAT.TIMEZONE_INTERVAL_OUT_OF_RANGE", "sqlState" : "22006", "messageParameters" : { "input" : "3" @@ -102,7 +102,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_INTERVAL_FORMAT.TIMEZONE_INTERVAL_OUT_OF_RANGE", + "condition" : "INVALID_INTERVAL_FORMAT.TIMEZONE_INTERVAL_OUT_OF_RANGE", "sqlState" : "22006", "messageParameters" : { "input" : "24" @@ -124,7 +124,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_INTERVAL_FORMAT.TIMEZONE_INTERVAL_OUT_OF_RANGE", + "condition" : "INVALID_INTERVAL_FORMAT.TIMEZONE_INTERVAL_OUT_OF_RANGE", "sqlState" : "22006", "messageParameters" : { "input" : "19" @@ -146,7 +146,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0045", + "condition" : "_LEGACY_ERROR_TEMP_0045", "queryContext" : [ { "objectType" : "", "objectName" : "", @@ -164,7 +164,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_INTERVAL_FORMAT.TIMEZONE_INTERVAL_OUT_OF_RANGE", + "condition" : "INVALID_INTERVAL_FORMAT.TIMEZONE_INTERVAL_OUT_OF_RANGE", "sqlState" : "22006", "messageParameters" : { "input" : "36000" diff --git a/sql/core/src/test/resources/sql-tests/results/to_from_avro.sql.out b/sql/core/src/test/resources/sql-tests/results/to_from_avro.sql.out index a94175b1df39..700fdcff1ed0 100644 --- a/sql/core/src/test/resources/sql-tests/results/to_from_avro.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/to_from_avro.sql.out @@ -33,7 +33,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.TYPE_CHECK_FAILURE_WITH_HINT", + "condition" : "DATATYPE_MISMATCH.TYPE_CHECK_FAILURE_WITH_HINT", "sqlState" : "42K09", "messageParameters" : { "hint" : "", @@ -57,7 +57,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.TYPE_CHECK_FAILURE_WITH_HINT", + "condition" : "DATATYPE_MISMATCH.TYPE_CHECK_FAILURE_WITH_HINT", "sqlState" : "42K09", "messageParameters" : { "hint" : "", @@ -81,7 +81,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.TYPE_CHECK_FAILURE_WITH_HINT", + "condition" : "DATATYPE_MISMATCH.TYPE_CHECK_FAILURE_WITH_HINT", "sqlState" : "42K09", "messageParameters" : { "hint" : "", @@ -105,7 +105,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "AVRO_NOT_LOADED_SQL_FUNCTIONS_UNUSABLE", + "condition" : "AVRO_NOT_LOADED_SQL_FUNCTIONS_UNUSABLE", "sqlState" : "22KD3", "messageParameters" : { "functionName" : "TO_AVRO" @@ -120,7 +120,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "AVRO_NOT_LOADED_SQL_FUNCTIONS_UNUSABLE", + "condition" : "AVRO_NOT_LOADED_SQL_FUNCTIONS_UNUSABLE", "sqlState" : "22KD3", "messageParameters" : { "functionName" : "FROM_AVRO" diff --git a/sql/core/src/test/resources/sql-tests/results/transform.sql.out b/sql/core/src/test/resources/sql-tests/results/transform.sql.out index 3c704dfcc461..7128d4c2d4df 100644 --- a/sql/core/src/test/resources/sql-tests/results/transform.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/transform.sql.out @@ -401,7 +401,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0064", + "condition" : "_LEGACY_ERROR_TEMP_0064", "messageParameters" : { "msg" : "LINES TERMINATED BY only supports newline '\\n' right now: @" }, @@ -699,7 +699,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "UNSUPPORTED_FEATURE.TRANSFORM_DISTINCT_ALL", + "condition" : "UNSUPPORTED_FEATURE.TRANSFORM_DISTINCT_ALL", "sqlState" : "0A000", "queryContext" : [ { "objectType" : "", @@ -721,7 +721,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "UNSUPPORTED_FEATURE.TRANSFORM_DISTINCT_ALL", + "condition" : "UNSUPPORTED_FEATURE.TRANSFORM_DISTINCT_ALL", "sqlState" : "0A000", "queryContext" : [ { "objectType" : "", @@ -744,7 +744,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'AS'", @@ -764,7 +764,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'b_1'", @@ -784,7 +784,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'AS'", diff --git a/sql/core/src/test/resources/sql-tests/results/try_aggregates.sql.out b/sql/core/src/test/resources/sql-tests/results/try_aggregates.sql.out index 94048ac8897b..f4208ae9d006 100644 --- a/sql/core/src/test/resources/sql-tests/results/try_aggregates.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/try_aggregates.sql.out @@ -86,7 +86,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "DIVIDE_BY_ZERO", + "condition" : "DIVIDE_BY_ZERO", "sqlState" : "22012", "messageParameters" : { "config" : "\"spark.sql.ansi.enabled\"" @@ -108,7 +108,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "DIVIDE_BY_ZERO", + "condition" : "DIVIDE_BY_ZERO", "sqlState" : "22012", "messageParameters" : { "config" : "\"spark.sql.ansi.enabled\"" @@ -130,7 +130,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "DIVIDE_BY_ZERO", + "condition" : "DIVIDE_BY_ZERO", "sqlState" : "22012", "messageParameters" : { "config" : "\"spark.sql.ansi.enabled\"" @@ -152,7 +152,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "ARITHMETIC_OVERFLOW", + "condition" : "ARITHMETIC_OVERFLOW", "sqlState" : "22003", "messageParameters" : { "alternative" : " Use 'try_add' to tolerate overflow and return NULL instead.", @@ -176,7 +176,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "INTERVAL_DIVIDED_BY_ZERO", + "condition" : "INTERVAL_DIVIDED_BY_ZERO", "sqlState" : "22012", "queryContext" : [ { "objectType" : "", @@ -195,7 +195,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "INTERVAL_DIVIDED_BY_ZERO", + "condition" : "INTERVAL_DIVIDED_BY_ZERO", "sqlState" : "22012", "queryContext" : [ { "objectType" : "", @@ -294,7 +294,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "DIVIDE_BY_ZERO", + "condition" : "DIVIDE_BY_ZERO", "sqlState" : "22012", "messageParameters" : { "config" : "\"spark.sql.ansi.enabled\"" @@ -316,7 +316,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "DIVIDE_BY_ZERO", + "condition" : "DIVIDE_BY_ZERO", "sqlState" : "22012", "messageParameters" : { "config" : "\"spark.sql.ansi.enabled\"" @@ -338,7 +338,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "DIVIDE_BY_ZERO", + "condition" : "DIVIDE_BY_ZERO", "sqlState" : "22012", "messageParameters" : { "config" : "\"spark.sql.ansi.enabled\"" @@ -360,7 +360,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "ARITHMETIC_OVERFLOW", + "condition" : "ARITHMETIC_OVERFLOW", "sqlState" : "22003", "messageParameters" : { "alternative" : " Use 'try_add' to tolerate overflow and return NULL instead.", @@ -384,7 +384,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "INTERVAL_DIVIDED_BY_ZERO", + "condition" : "INTERVAL_DIVIDED_BY_ZERO", "sqlState" : "22012", "queryContext" : [ { "objectType" : "", @@ -403,7 +403,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "INTERVAL_DIVIDED_BY_ZERO", + "condition" : "INTERVAL_DIVIDED_BY_ZERO", "sqlState" : "22012", "queryContext" : [ { "objectType" : "", diff --git a/sql/core/src/test/resources/sql-tests/results/try_aggregates.sql.out.java21 b/sql/core/src/test/resources/sql-tests/results/try_aggregates.sql.out.java21 index 9d3c97baecab..9c67eaebd75b 100644 --- a/sql/core/src/test/resources/sql-tests/results/try_aggregates.sql.out.java21 +++ b/sql/core/src/test/resources/sql-tests/results/try_aggregates.sql.out.java21 @@ -176,7 +176,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "INTERVAL_DIVIDED_BY_ZERO", + "condition" : "INTERVAL_DIVIDED_BY_ZERO", "sqlState" : "22012", "queryContext" : [ { "objectType" : "", @@ -195,7 +195,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "INTERVAL_DIVIDED_BY_ZERO", + "condition" : "INTERVAL_DIVIDED_BY_ZERO", "sqlState" : "22012", "queryContext" : [ { "objectType" : "", @@ -384,7 +384,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "INTERVAL_DIVIDED_BY_ZERO", + "condition" : "INTERVAL_DIVIDED_BY_ZERO", "sqlState" : "22012", "queryContext" : [ { "objectType" : "", @@ -403,7 +403,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "INTERVAL_DIVIDED_BY_ZERO", + "condition" : "INTERVAL_DIVIDED_BY_ZERO", "sqlState" : "22012", "queryContext" : [ { "objectType" : "", diff --git a/sql/core/src/test/resources/sql-tests/results/try_arithmetic.sql.out b/sql/core/src/test/resources/sql-tests/results/try_arithmetic.sql.out index acf6e70a50de..e8df5dd5af49 100644 --- a/sql/core/src/test/resources/sql-tests/results/try_arithmetic.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/try_arithmetic.sql.out @@ -62,7 +62,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "ARITHMETIC_OVERFLOW", + "condition" : "ARITHMETIC_OVERFLOW", "sqlState" : "22003", "messageParameters" : { "alternative" : " Use 'try_add' to tolerate overflow and return NULL instead.", @@ -86,7 +86,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "ARITHMETIC_OVERFLOW", + "condition" : "ARITHMETIC_OVERFLOW", "sqlState" : "22003", "messageParameters" : { "alternative" : " Use 'try_add' to tolerate overflow and return NULL instead.", @@ -110,7 +110,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "DIVIDE_BY_ZERO", + "condition" : "DIVIDE_BY_ZERO", "sqlState" : "22012", "messageParameters" : { "config" : "\"spark.sql.ansi.enabled\"" @@ -228,7 +228,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"INTERVAL '2' YEAR\"", @@ -294,7 +294,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "ARITHMETIC_OVERFLOW", + "condition" : "ARITHMETIC_OVERFLOW", "sqlState" : "22003", "messageParameters" : { "alternative" : " Use 'try_add' to tolerate overflow and return NULL instead.", @@ -318,7 +318,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "ARITHMETIC_OVERFLOW", + "condition" : "ARITHMETIC_OVERFLOW", "sqlState" : "22003", "messageParameters" : { "alternative" : " Use 'try_add' to tolerate overflow and return NULL instead.", @@ -342,7 +342,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "DIVIDE_BY_ZERO", + "condition" : "DIVIDE_BY_ZERO", "sqlState" : "22012", "messageParameters" : { "config" : "\"spark.sql.ansi.enabled\"" @@ -484,7 +484,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "ARITHMETIC_OVERFLOW", + "condition" : "ARITHMETIC_OVERFLOW", "sqlState" : "22003", "messageParameters" : { "alternative" : " Use 'try_add' to tolerate overflow and return NULL instead.", @@ -508,7 +508,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "ARITHMETIC_OVERFLOW", + "condition" : "ARITHMETIC_OVERFLOW", "sqlState" : "22003", "messageParameters" : { "alternative" : " Use 'try_add' to tolerate overflow and return NULL instead.", @@ -532,7 +532,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "DIVIDE_BY_ZERO", + "condition" : "DIVIDE_BY_ZERO", "sqlState" : "22012", "messageParameters" : { "config" : "\"spark.sql.ansi.enabled\"" @@ -642,7 +642,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "ARITHMETIC_OVERFLOW", + "condition" : "ARITHMETIC_OVERFLOW", "sqlState" : "22003", "messageParameters" : { "alternative" : " Use 'try_add' to tolerate overflow and return NULL instead.", @@ -666,7 +666,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "ARITHMETIC_OVERFLOW", + "condition" : "ARITHMETIC_OVERFLOW", "sqlState" : "22003", "messageParameters" : { "alternative" : " Use 'try_add' to tolerate overflow and return NULL instead.", @@ -690,7 +690,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "DIVIDE_BY_ZERO", + "condition" : "DIVIDE_BY_ZERO", "sqlState" : "22012", "messageParameters" : { "config" : "\"spark.sql.ansi.enabled\"" diff --git a/sql/core/src/test/resources/sql-tests/results/try_datetime_functions.sql.out b/sql/core/src/test/resources/sql-tests/results/try_datetime_functions.sql.out index 75a6f15bd363..a2383e7081d3 100644 --- a/sql/core/src/test/resources/sql-tests/results/try_datetime_functions.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/try_datetime_functions.sql.out @@ -46,7 +46,7 @@ struct<> -- !query output org.apache.spark.SparkUpgradeException { - "errorClass" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION.DATETIME_PATTERN_RECOGNITION", + "condition" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION.DATETIME_PATTERN_RECOGNITION", "sqlState" : "42K0B", "messageParameters" : { "config" : "\"spark.sql.legacy.timeParserPolicy\"", diff --git a/sql/core/src/test/resources/sql-tests/results/try_element_at.sql.out b/sql/core/src/test/resources/sql-tests/results/try_element_at.sql.out index 0437f9d6dd9e..ea0c68680c18 100644 --- a/sql/core/src/test/resources/sql-tests/results/try_element_at.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/try_element_at.sql.out @@ -6,7 +6,7 @@ struct<> -- !query output org.apache.spark.SparkRuntimeException { - "errorClass" : "INVALID_INDEX_OF_ZERO", + "condition" : "INVALID_INDEX_OF_ZERO", "sqlState" : "22003" } diff --git a/sql/core/src/test/resources/sql-tests/results/try_reflect.sql.out b/sql/core/src/test/resources/sql-tests/results/try_reflect.sql.out index 8b2c4f0ed737..30ab8e0a3b5e 100644 --- a/sql/core/src/test/resources/sql-tests/results/try_reflect.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/try_reflect.sql.out @@ -62,7 +62,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_CLASS_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_CLASS_TYPE", "sqlState" : "42K09", "messageParameters" : { "className" : "java.wrongclass.Math", @@ -85,7 +85,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_STATIC_METHOD", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_STATIC_METHOD", "sqlState" : "42K09", "messageParameters" : { "className" : "java.lang.Math", @@ -109,7 +109,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", + "condition" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", "sqlState" : "42605", "messageParameters" : { "actualNum" : "1", @@ -127,7 +127,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"2.5\"", @@ -153,7 +153,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_STATIC_METHOD", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_STATIC_METHOD", "sqlState" : "42K09", "messageParameters" : { "className" : "java.lang.Object", diff --git a/sql/core/src/test/resources/sql-tests/results/typeCoercion/native/binaryComparison.sql.out b/sql/core/src/test/resources/sql-tests/results/typeCoercion/native/binaryComparison.sql.out index f3263241a556..4f946ea9c7f4 100644 --- a/sql/core/src/test/resources/sql-tests/results/typeCoercion/native/binaryComparison.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/typeCoercion/native/binaryComparison.sql.out @@ -14,7 +14,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", + "condition" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", "sqlState" : "42K09", "messageParameters" : { "config" : "\"spark.sql.ansi.enabled\"", @@ -40,7 +40,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", + "condition" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", "sqlState" : "42K09", "messageParameters" : { "config" : "\"spark.sql.ansi.enabled\"", @@ -66,7 +66,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", + "condition" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", "sqlState" : "42K09", "messageParameters" : { "config" : "\"spark.sql.ansi.enabled\"", @@ -92,7 +92,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", + "condition" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", "sqlState" : "42K09", "messageParameters" : { "config" : "\"spark.sql.ansi.enabled\"", @@ -118,7 +118,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", + "condition" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", "sqlState" : "42K09", "messageParameters" : { "config" : "\"spark.sql.ansi.enabled\"", @@ -144,7 +144,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", + "condition" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", "sqlState" : "42K09", "messageParameters" : { "config" : "\"spark.sql.ansi.enabled\"", @@ -170,7 +170,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", + "condition" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", "sqlState" : "42K09", "messageParameters" : { "config" : "\"spark.sql.ansi.enabled\"", @@ -196,7 +196,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", + "condition" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", "sqlState" : "42K09", "messageParameters" : { "config" : "\"spark.sql.ansi.enabled\"", @@ -222,7 +222,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", + "condition" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", "sqlState" : "42K09", "messageParameters" : { "config" : "\"spark.sql.ansi.enabled\"", @@ -248,7 +248,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", + "condition" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", "sqlState" : "42K09", "messageParameters" : { "config" : "\"spark.sql.ansi.enabled\"", @@ -274,7 +274,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", + "condition" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", "sqlState" : "42K09", "messageParameters" : { "config" : "\"spark.sql.ansi.enabled\"", @@ -300,7 +300,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", + "condition" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", "sqlState" : "42K09", "messageParameters" : { "config" : "\"spark.sql.ansi.enabled\"", @@ -326,7 +326,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", + "condition" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", "sqlState" : "42K09", "messageParameters" : { "config" : "\"spark.sql.ansi.enabled\"", @@ -352,7 +352,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", + "condition" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", "sqlState" : "42K09", "messageParameters" : { "config" : "\"spark.sql.ansi.enabled\"", @@ -378,7 +378,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", + "condition" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", "sqlState" : "42K09", "messageParameters" : { "config" : "\"spark.sql.ansi.enabled\"", @@ -404,7 +404,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", + "condition" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", "sqlState" : "42K09", "messageParameters" : { "config" : "\"spark.sql.ansi.enabled\"", @@ -430,7 +430,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", + "condition" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", "sqlState" : "42K09", "messageParameters" : { "config" : "\"spark.sql.ansi.enabled\"", @@ -456,7 +456,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", + "condition" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", "sqlState" : "42K09", "messageParameters" : { "config" : "\"spark.sql.ansi.enabled\"", @@ -482,7 +482,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", + "condition" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", "sqlState" : "42K09", "messageParameters" : { "config" : "\"spark.sql.ansi.enabled\"", @@ -508,7 +508,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", + "condition" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", "sqlState" : "42K09", "messageParameters" : { "config" : "\"spark.sql.ansi.enabled\"", @@ -534,7 +534,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", + "condition" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", "sqlState" : "42K09", "messageParameters" : { "config" : "\"spark.sql.ansi.enabled\"", @@ -560,7 +560,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", + "condition" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", "sqlState" : "42K09", "messageParameters" : { "config" : "\"spark.sql.ansi.enabled\"", @@ -586,7 +586,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", + "condition" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", "sqlState" : "42K09", "messageParameters" : { "config" : "\"spark.sql.ansi.enabled\"", @@ -612,7 +612,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", + "condition" : "DATATYPE_MISMATCH.CAST_WITH_CONF_SUGGESTION", "sqlState" : "42K09", "messageParameters" : { "config" : "\"spark.sql.ansi.enabled\"", diff --git a/sql/core/src/test/resources/sql-tests/results/typeCoercion/native/booleanEquality.sql.out b/sql/core/src/test/resources/sql-tests/results/typeCoercion/native/booleanEquality.sql.out index f83284e9bf6f..432c6c06cabc 100644 --- a/sql/core/src/test/resources/sql-tests/results/typeCoercion/native/booleanEquality.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/typeCoercion/native/booleanEquality.sql.out @@ -14,7 +14,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BOOLEAN\"", @@ -38,7 +38,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BOOLEAN\"", @@ -62,7 +62,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BOOLEAN\"", @@ -86,7 +86,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BOOLEAN\"", @@ -110,7 +110,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BOOLEAN\"", @@ -134,7 +134,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BOOLEAN\"", @@ -158,7 +158,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BOOLEAN\"", @@ -190,7 +190,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BOOLEAN\"", @@ -222,7 +222,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BOOLEAN\"", @@ -246,7 +246,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BOOLEAN\"", @@ -270,7 +270,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BOOLEAN\"", @@ -294,7 +294,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BOOLEAN\"", @@ -318,7 +318,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BOOLEAN\"", @@ -342,7 +342,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BOOLEAN\"", @@ -366,7 +366,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BOOLEAN\"", @@ -390,7 +390,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BOOLEAN\"", @@ -414,7 +414,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BOOLEAN\"", @@ -446,7 +446,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BOOLEAN\"", @@ -478,7 +478,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BOOLEAN\"", @@ -502,7 +502,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BOOLEAN\"", @@ -526,7 +526,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TINYINT\"", @@ -550,7 +550,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"SMALLINT\"", @@ -574,7 +574,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"INT\"", @@ -598,7 +598,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BIGINT\"", @@ -622,7 +622,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"FLOAT\"", @@ -646,7 +646,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DOUBLE\"", @@ -670,7 +670,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(10,0)\"", @@ -702,7 +702,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BINARY\"", @@ -734,7 +734,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP\"", @@ -758,7 +758,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DATE\"", @@ -782,7 +782,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TINYINT\"", @@ -806,7 +806,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"SMALLINT\"", @@ -830,7 +830,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"INT\"", @@ -854,7 +854,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BIGINT\"", @@ -878,7 +878,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"FLOAT\"", @@ -902,7 +902,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DOUBLE\"", @@ -926,7 +926,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(10,0)\"", @@ -958,7 +958,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BINARY\"", @@ -990,7 +990,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP\"", @@ -1014,7 +1014,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DATE\"", @@ -1038,7 +1038,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BOOLEAN\"", @@ -1062,7 +1062,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BOOLEAN\"", @@ -1086,7 +1086,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BOOLEAN\"", @@ -1110,7 +1110,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BOOLEAN\"", @@ -1134,7 +1134,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BOOLEAN\"", @@ -1158,7 +1158,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BOOLEAN\"", @@ -1182,7 +1182,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BOOLEAN\"", @@ -1214,7 +1214,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BOOLEAN\"", @@ -1246,7 +1246,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BOOLEAN\"", @@ -1270,7 +1270,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BOOLEAN\"", @@ -1294,7 +1294,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BOOLEAN\"", @@ -1318,7 +1318,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BOOLEAN\"", @@ -1342,7 +1342,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BOOLEAN\"", @@ -1366,7 +1366,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BOOLEAN\"", @@ -1390,7 +1390,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BOOLEAN\"", @@ -1414,7 +1414,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BOOLEAN\"", @@ -1438,7 +1438,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BOOLEAN\"", @@ -1470,7 +1470,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BOOLEAN\"", @@ -1502,7 +1502,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BOOLEAN\"", @@ -1526,7 +1526,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BOOLEAN\"", @@ -1550,7 +1550,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TINYINT\"", @@ -1574,7 +1574,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"SMALLINT\"", @@ -1598,7 +1598,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"INT\"", @@ -1622,7 +1622,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BIGINT\"", @@ -1646,7 +1646,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"FLOAT\"", @@ -1670,7 +1670,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DOUBLE\"", @@ -1694,7 +1694,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(10,0)\"", @@ -1726,7 +1726,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BINARY\"", @@ -1758,7 +1758,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP\"", @@ -1782,7 +1782,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DATE\"", @@ -1806,7 +1806,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TINYINT\"", @@ -1830,7 +1830,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"SMALLINT\"", @@ -1854,7 +1854,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"INT\"", @@ -1878,7 +1878,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BIGINT\"", @@ -1902,7 +1902,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"FLOAT\"", @@ -1926,7 +1926,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DOUBLE\"", @@ -1950,7 +1950,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(10,0)\"", @@ -1982,7 +1982,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BINARY\"", @@ -2014,7 +2014,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP\"", @@ -2038,7 +2038,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DATE\"", diff --git a/sql/core/src/test/resources/sql-tests/results/typeCoercion/native/caseWhenCoercion.sql.out b/sql/core/src/test/resources/sql-tests/results/typeCoercion/native/caseWhenCoercion.sql.out index 35ff9e79d980..1a3c4292caf8 100644 --- a/sql/core/src/test/resources/sql-tests/results/typeCoercion/native/caseWhenCoercion.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/typeCoercion/native/caseWhenCoercion.sql.out @@ -78,7 +78,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"TINYINT\", \"BINARY\"]", @@ -102,7 +102,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"TINYINT\", \"BOOLEAN\"]", @@ -126,7 +126,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"TINYINT\", \"TIMESTAMP\"]", @@ -150,7 +150,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"TINYINT\", \"DATE\"]", @@ -238,7 +238,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"SMALLINT\", \"BINARY\"]", @@ -262,7 +262,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"SMALLINT\", \"BOOLEAN\"]", @@ -286,7 +286,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"SMALLINT\", \"TIMESTAMP\"]", @@ -310,7 +310,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"SMALLINT\", \"DATE\"]", @@ -398,7 +398,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"INT\", \"BINARY\"]", @@ -422,7 +422,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"INT\", \"BOOLEAN\"]", @@ -446,7 +446,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"INT\", \"TIMESTAMP\"]", @@ -470,7 +470,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"INT\", \"DATE\"]", @@ -558,7 +558,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BIGINT\", \"BINARY\"]", @@ -582,7 +582,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BIGINT\", \"BOOLEAN\"]", @@ -606,7 +606,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BIGINT\", \"TIMESTAMP\"]", @@ -630,7 +630,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BIGINT\", \"DATE\"]", @@ -718,7 +718,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"FLOAT\", \"BINARY\"]", @@ -742,7 +742,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"FLOAT\", \"BOOLEAN\"]", @@ -766,7 +766,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"FLOAT\", \"TIMESTAMP\"]", @@ -790,7 +790,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"FLOAT\", \"DATE\"]", @@ -878,7 +878,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DOUBLE\", \"BINARY\"]", @@ -902,7 +902,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DOUBLE\", \"BOOLEAN\"]", @@ -926,7 +926,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DOUBLE\", \"TIMESTAMP\"]", @@ -950,7 +950,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DOUBLE\", \"DATE\"]", @@ -1038,7 +1038,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DECIMAL(10,0)\", \"BINARY\"]", @@ -1062,7 +1062,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DECIMAL(10,0)\", \"BOOLEAN\"]", @@ -1086,7 +1086,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DECIMAL(10,0)\", \"TIMESTAMP\"]", @@ -1110,7 +1110,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DECIMAL(10,0)\", \"DATE\"]", @@ -1214,7 +1214,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'1'", @@ -1238,7 +1238,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'1'", @@ -1262,7 +1262,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BINARY\", \"TINYINT\"]", @@ -1286,7 +1286,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BINARY\", \"SMALLINT\"]", @@ -1310,7 +1310,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BINARY\", \"INT\"]", @@ -1334,7 +1334,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BINARY\", \"BIGINT\"]", @@ -1358,7 +1358,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BINARY\", \"FLOAT\"]", @@ -1382,7 +1382,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BINARY\", \"DOUBLE\"]", @@ -1406,7 +1406,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BINARY\", \"DECIMAL(10,0)\"]", @@ -1446,7 +1446,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BINARY\", \"BOOLEAN\"]", @@ -1470,7 +1470,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BINARY\", \"TIMESTAMP\"]", @@ -1494,7 +1494,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BINARY\", \"DATE\"]", @@ -1518,7 +1518,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BOOLEAN\", \"TINYINT\"]", @@ -1542,7 +1542,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BOOLEAN\", \"SMALLINT\"]", @@ -1566,7 +1566,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BOOLEAN\", \"INT\"]", @@ -1590,7 +1590,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BOOLEAN\", \"BIGINT\"]", @@ -1614,7 +1614,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BOOLEAN\", \"FLOAT\"]", @@ -1638,7 +1638,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BOOLEAN\", \"DOUBLE\"]", @@ -1662,7 +1662,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BOOLEAN\", \"DECIMAL(10,0)\"]", @@ -1694,7 +1694,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BOOLEAN\", \"BINARY\"]", @@ -1726,7 +1726,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BOOLEAN\", \"TIMESTAMP\"]", @@ -1750,7 +1750,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BOOLEAN\", \"DATE\"]", @@ -1774,7 +1774,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"TIMESTAMP\", \"TINYINT\"]", @@ -1798,7 +1798,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"TIMESTAMP\", \"SMALLINT\"]", @@ -1822,7 +1822,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"TIMESTAMP\", \"INT\"]", @@ -1846,7 +1846,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"TIMESTAMP\", \"BIGINT\"]", @@ -1870,7 +1870,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"TIMESTAMP\", \"FLOAT\"]", @@ -1894,7 +1894,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"TIMESTAMP\", \"DOUBLE\"]", @@ -1918,7 +1918,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"TIMESTAMP\", \"DECIMAL(10,0)\"]", @@ -1950,7 +1950,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"TIMESTAMP\", \"BINARY\"]", @@ -1974,7 +1974,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"TIMESTAMP\", \"BOOLEAN\"]", @@ -2014,7 +2014,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DATE\", \"TINYINT\"]", @@ -2038,7 +2038,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DATE\", \"SMALLINT\"]", @@ -2062,7 +2062,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DATE\", \"INT\"]", @@ -2086,7 +2086,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DATE\", \"BIGINT\"]", @@ -2110,7 +2110,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DATE\", \"FLOAT\"]", @@ -2134,7 +2134,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DATE\", \"DOUBLE\"]", @@ -2158,7 +2158,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DATE\", \"DECIMAL(10,0)\"]", @@ -2190,7 +2190,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DATE\", \"BINARY\"]", @@ -2214,7 +2214,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DATE\", \"BOOLEAN\"]", diff --git a/sql/core/src/test/resources/sql-tests/results/typeCoercion/native/concat.sql.out b/sql/core/src/test/resources/sql-tests/results/typeCoercion/native/concat.sql.out index 0f42834d2824..057cf8735137 100644 --- a/sql/core/src/test/resources/sql-tests/results/typeCoercion/native/concat.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/typeCoercion/native/concat.sql.out @@ -315,7 +315,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'a'", diff --git a/sql/core/src/test/resources/sql-tests/results/typeCoercion/native/dateTimeOperations.sql.out b/sql/core/src/test/resources/sql-tests/results/typeCoercion/native/dateTimeOperations.sql.out index 10cceabca052..7b3b763ff6f7 100644 --- a/sql/core/src/test/resources/sql-tests/results/typeCoercion/native/dateTimeOperations.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/typeCoercion/native/dateTimeOperations.sql.out @@ -14,7 +14,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"CAST(1 AS TINYINT)\"", @@ -40,7 +40,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"CAST(1 AS SMALLINT)\"", @@ -66,7 +66,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"CAST(1 AS INT)\"", @@ -92,7 +92,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"CAST(1 AS BIGINT)\"", @@ -118,7 +118,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"CAST(1 AS FLOAT)\"", @@ -144,7 +144,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"CAST(1 AS DOUBLE)\"", @@ -170,7 +170,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"CAST(1 AS DECIMAL(10,0))\"", @@ -212,7 +212,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"CAST(1 AS BINARY)\"", @@ -238,7 +238,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"CAST(1 AS BOOLEAN)\"", @@ -280,7 +280,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"CAST(1 AS TINYINT)\"", @@ -306,7 +306,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"CAST(1 AS SMALLINT)\"", @@ -332,7 +332,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"CAST(1 AS INT)\"", @@ -358,7 +358,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"CAST(1 AS BIGINT)\"", @@ -384,7 +384,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"CAST(1 AS FLOAT)\"", @@ -410,7 +410,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"CAST(1 AS DOUBLE)\"", @@ -436,7 +436,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"CAST(1 AS DECIMAL(10,0))\"", @@ -478,7 +478,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"CAST(1 AS BINARY)\"", @@ -504,7 +504,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"CAST(1 AS BOOLEAN)\"", @@ -546,7 +546,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"CAST(1 AS TINYINT)\"", @@ -572,7 +572,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"CAST(1 AS SMALLINT)\"", @@ -598,7 +598,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"CAST(1 AS INT)\"", @@ -624,7 +624,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"CAST(1 AS BIGINT)\"", @@ -650,7 +650,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"CAST(1 AS FLOAT)\"", @@ -676,7 +676,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"CAST(1 AS DOUBLE)\"", @@ -702,7 +702,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"CAST(1 AS DECIMAL(10,0))\"", @@ -744,7 +744,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"CAST(1 AS BINARY)\"", @@ -770,7 +770,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"CAST(1 AS BOOLEAN)\"", diff --git a/sql/core/src/test/resources/sql-tests/results/typeCoercion/native/decimalPrecision.sql.out b/sql/core/src/test/resources/sql-tests/results/typeCoercion/native/decimalPrecision.sql.out index 54e26851ba57..bea318c72da3 100644 --- a/sql/core/src/test/resources/sql-tests/results/typeCoercion/native/decimalPrecision.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/typeCoercion/native/decimalPrecision.sql.out @@ -238,7 +238,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BINARY\"", @@ -262,7 +262,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BINARY\"", @@ -286,7 +286,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BINARY\"", @@ -310,7 +310,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BINARY\"", @@ -334,7 +334,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP\"", @@ -358,7 +358,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP\"", @@ -382,7 +382,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP\"", @@ -406,7 +406,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP\"", @@ -430,7 +430,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"CAST(1 AS DECIMAL(3,0))\"", @@ -456,7 +456,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"CAST(1 AS DECIMAL(5,0))\"", @@ -482,7 +482,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"CAST(1 AS DECIMAL(10,0))\"", @@ -508,7 +508,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"CAST(1 AS DECIMAL(20,0))\"", @@ -790,7 +790,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(3,0)\"", @@ -814,7 +814,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(5,0)\"", @@ -838,7 +838,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(10,0)\"", @@ -862,7 +862,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(20,0)\"", @@ -886,7 +886,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(3,0)\"", @@ -910,7 +910,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(5,0)\"", @@ -934,7 +934,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(10,0)\"", @@ -958,7 +958,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(20,0)\"", @@ -982,7 +982,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(3,0)\"", @@ -1006,7 +1006,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(5,0)\"", @@ -1030,7 +1030,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(10,0)\"", @@ -1054,7 +1054,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(20,0)\"", @@ -1078,7 +1078,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"CAST(1 AS DECIMAL(3,0))\"", @@ -1104,7 +1104,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"CAST(1 AS DECIMAL(5,0))\"", @@ -1130,7 +1130,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"CAST(1 AS DECIMAL(10,0))\"", @@ -1156,7 +1156,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"CAST(1 AS DECIMAL(20,0))\"", @@ -1406,7 +1406,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BINARY\"", @@ -1430,7 +1430,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BINARY\"", @@ -1454,7 +1454,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BINARY\"", @@ -1478,7 +1478,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BINARY\"", @@ -1502,7 +1502,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"CAST(1 AS DECIMAL(3,0))\"", @@ -1528,7 +1528,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"CAST(1 AS DECIMAL(5,0))\"", @@ -1554,7 +1554,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"CAST(1 AS DECIMAL(10,0))\"", @@ -1580,7 +1580,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"CAST(1 AS DECIMAL(20,0))\"", @@ -1606,7 +1606,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"CAST(1 AS DECIMAL(3,0))\"", @@ -1632,7 +1632,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"CAST(1 AS DECIMAL(5,0))\"", @@ -1658,7 +1658,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"CAST(1 AS DECIMAL(10,0))\"", @@ -1684,7 +1684,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"CAST(1 AS DECIMAL(20,0))\"", @@ -1966,7 +1966,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(3,0)\"", @@ -1990,7 +1990,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(5,0)\"", @@ -2014,7 +2014,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(10,0)\"", @@ -2038,7 +2038,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(20,0)\"", @@ -2062,7 +2062,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(3,0)\"", @@ -2086,7 +2086,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(5,0)\"", @@ -2110,7 +2110,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(10,0)\"", @@ -2134,7 +2134,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(20,0)\"", @@ -2158,7 +2158,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"CAST(1 AS DECIMAL(3,0))\"", @@ -2184,7 +2184,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"CAST(1 AS DECIMAL(5,0))\"", @@ -2210,7 +2210,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"CAST(1 AS DECIMAL(10,0))\"", @@ -2236,7 +2236,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"CAST(1 AS DECIMAL(20,0))\"", @@ -2262,7 +2262,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"CAST(1 AS DECIMAL(3,0))\"", @@ -2288,7 +2288,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"CAST(1 AS DECIMAL(5,0))\"", @@ -2314,7 +2314,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"CAST(1 AS DECIMAL(10,0))\"", @@ -2340,7 +2340,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"CAST(1 AS DECIMAL(20,0))\"", @@ -2590,7 +2590,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BINARY\"", @@ -2614,7 +2614,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BINARY\"", @@ -2638,7 +2638,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BINARY\"", @@ -2662,7 +2662,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BINARY\"", @@ -2686,7 +2686,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP\"", @@ -2710,7 +2710,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP\"", @@ -2734,7 +2734,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP\"", @@ -2758,7 +2758,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP\"", @@ -2782,7 +2782,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DATE\"", @@ -2806,7 +2806,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DATE\"", @@ -2830,7 +2830,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DATE\"", @@ -2854,7 +2854,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DATE\"", @@ -3134,7 +3134,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(3,0)\"", @@ -3158,7 +3158,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(5,0)\"", @@ -3182,7 +3182,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(10,0)\"", @@ -3206,7 +3206,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(20,0)\"", @@ -3230,7 +3230,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(3,0)\"", @@ -3254,7 +3254,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(5,0)\"", @@ -3278,7 +3278,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(10,0)\"", @@ -3302,7 +3302,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(20,0)\"", @@ -3326,7 +3326,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(3,0)\"", @@ -3350,7 +3350,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(5,0)\"", @@ -3374,7 +3374,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(10,0)\"", @@ -3398,7 +3398,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(20,0)\"", @@ -3422,7 +3422,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(3,0)\"", @@ -3446,7 +3446,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(5,0)\"", @@ -3470,7 +3470,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(10,0)\"", @@ -3494,7 +3494,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(20,0)\"", @@ -3742,7 +3742,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BINARY\"", @@ -3766,7 +3766,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BINARY\"", @@ -3790,7 +3790,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BINARY\"", @@ -3814,7 +3814,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BINARY\"", @@ -3838,7 +3838,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP\"", @@ -3862,7 +3862,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP\"", @@ -3886,7 +3886,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP\"", @@ -3910,7 +3910,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP\"", @@ -3934,7 +3934,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DATE\"", @@ -3958,7 +3958,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DATE\"", @@ -3982,7 +3982,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DATE\"", @@ -4006,7 +4006,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DATE\"", @@ -4286,7 +4286,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(3,0)\"", @@ -4310,7 +4310,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(5,0)\"", @@ -4334,7 +4334,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(10,0)\"", @@ -4358,7 +4358,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(20,0)\"", @@ -4382,7 +4382,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(3,0)\"", @@ -4406,7 +4406,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(5,0)\"", @@ -4430,7 +4430,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(10,0)\"", @@ -4454,7 +4454,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(20,0)\"", @@ -4478,7 +4478,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(3,0)\"", @@ -4502,7 +4502,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(5,0)\"", @@ -4526,7 +4526,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(10,0)\"", @@ -4550,7 +4550,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(20,0)\"", @@ -4574,7 +4574,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(3,0)\"", @@ -4598,7 +4598,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(5,0)\"", @@ -4622,7 +4622,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(10,0)\"", @@ -4646,7 +4646,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(20,0)\"", @@ -4894,7 +4894,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BINARY\"", @@ -4918,7 +4918,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BINARY\"", @@ -4942,7 +4942,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BINARY\"", @@ -4966,7 +4966,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BINARY\"", @@ -4990,7 +4990,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP\"", @@ -5014,7 +5014,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP\"", @@ -5038,7 +5038,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP\"", @@ -5062,7 +5062,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP\"", @@ -5086,7 +5086,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DATE\"", @@ -5110,7 +5110,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DATE\"", @@ -5134,7 +5134,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DATE\"", @@ -5158,7 +5158,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DATE\"", @@ -5438,7 +5438,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(3,0)\"", @@ -5462,7 +5462,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(5,0)\"", @@ -5486,7 +5486,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(10,0)\"", @@ -5510,7 +5510,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(20,0)\"", @@ -5534,7 +5534,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(3,0)\"", @@ -5558,7 +5558,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(5,0)\"", @@ -5582,7 +5582,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(10,0)\"", @@ -5606,7 +5606,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(20,0)\"", @@ -5630,7 +5630,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(3,0)\"", @@ -5654,7 +5654,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(5,0)\"", @@ -5678,7 +5678,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(10,0)\"", @@ -5702,7 +5702,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(20,0)\"", @@ -5726,7 +5726,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(3,0)\"", @@ -5750,7 +5750,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(5,0)\"", @@ -5774,7 +5774,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(10,0)\"", @@ -5798,7 +5798,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(20,0)\"", @@ -6046,7 +6046,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BINARY\"", @@ -6070,7 +6070,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BINARY\"", @@ -6094,7 +6094,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BINARY\"", @@ -6118,7 +6118,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BINARY\"", @@ -6142,7 +6142,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP\"", @@ -6166,7 +6166,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP\"", @@ -6190,7 +6190,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP\"", @@ -6214,7 +6214,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP\"", @@ -6238,7 +6238,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DATE\"", @@ -6262,7 +6262,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DATE\"", @@ -6286,7 +6286,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DATE\"", @@ -6310,7 +6310,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DATE\"", @@ -6590,7 +6590,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(3,0)\"", @@ -6614,7 +6614,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(5,0)\"", @@ -6638,7 +6638,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(10,0)\"", @@ -6662,7 +6662,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(20,0)\"", @@ -6686,7 +6686,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(3,0)\"", @@ -6710,7 +6710,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(5,0)\"", @@ -6734,7 +6734,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(10,0)\"", @@ -6758,7 +6758,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(20,0)\"", @@ -6782,7 +6782,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(3,0)\"", @@ -6806,7 +6806,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(5,0)\"", @@ -6830,7 +6830,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(10,0)\"", @@ -6854,7 +6854,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(20,0)\"", @@ -6878,7 +6878,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(3,0)\"", @@ -6902,7 +6902,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(5,0)\"", @@ -6926,7 +6926,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(10,0)\"", @@ -6950,7 +6950,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(20,0)\"", @@ -7198,7 +7198,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BINARY\"", @@ -7222,7 +7222,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BINARY\"", @@ -7246,7 +7246,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BINARY\"", @@ -7270,7 +7270,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BINARY\"", @@ -7294,7 +7294,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP\"", @@ -7318,7 +7318,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP\"", @@ -7342,7 +7342,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP\"", @@ -7366,7 +7366,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP\"", @@ -7390,7 +7390,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DATE\"", @@ -7414,7 +7414,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DATE\"", @@ -7438,7 +7438,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DATE\"", @@ -7462,7 +7462,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DATE\"", @@ -7742,7 +7742,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(3,0)\"", @@ -7766,7 +7766,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(5,0)\"", @@ -7790,7 +7790,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(10,0)\"", @@ -7814,7 +7814,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(20,0)\"", @@ -7838,7 +7838,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(3,0)\"", @@ -7862,7 +7862,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(5,0)\"", @@ -7886,7 +7886,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(10,0)\"", @@ -7910,7 +7910,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(20,0)\"", @@ -7934,7 +7934,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(3,0)\"", @@ -7958,7 +7958,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(5,0)\"", @@ -7982,7 +7982,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(10,0)\"", @@ -8006,7 +8006,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(20,0)\"", @@ -8030,7 +8030,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(3,0)\"", @@ -8054,7 +8054,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(5,0)\"", @@ -8078,7 +8078,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(10,0)\"", @@ -8102,7 +8102,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(20,0)\"", @@ -8350,7 +8350,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BINARY\"", @@ -8374,7 +8374,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BINARY\"", @@ -8398,7 +8398,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BINARY\"", @@ -8422,7 +8422,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BINARY\"", @@ -8446,7 +8446,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP\"", @@ -8470,7 +8470,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP\"", @@ -8494,7 +8494,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP\"", @@ -8518,7 +8518,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP\"", @@ -8542,7 +8542,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DATE\"", @@ -8566,7 +8566,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DATE\"", @@ -8590,7 +8590,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DATE\"", @@ -8614,7 +8614,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DATE\"", @@ -8894,7 +8894,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(3,0)\"", @@ -8918,7 +8918,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(5,0)\"", @@ -8942,7 +8942,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(10,0)\"", @@ -8966,7 +8966,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(20,0)\"", @@ -8990,7 +8990,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(3,0)\"", @@ -9014,7 +9014,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(5,0)\"", @@ -9038,7 +9038,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(10,0)\"", @@ -9062,7 +9062,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(20,0)\"", @@ -9086,7 +9086,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(3,0)\"", @@ -9110,7 +9110,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(5,0)\"", @@ -9134,7 +9134,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(10,0)\"", @@ -9158,7 +9158,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(20,0)\"", @@ -9182,7 +9182,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(3,0)\"", @@ -9206,7 +9206,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(5,0)\"", @@ -9230,7 +9230,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(10,0)\"", @@ -9254,7 +9254,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(20,0)\"", @@ -9502,7 +9502,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BINARY\"", @@ -9526,7 +9526,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BINARY\"", @@ -9550,7 +9550,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BINARY\"", @@ -9574,7 +9574,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BINARY\"", @@ -9598,7 +9598,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP\"", @@ -9622,7 +9622,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP\"", @@ -9646,7 +9646,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP\"", @@ -9670,7 +9670,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP\"", @@ -9694,7 +9694,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DATE\"", @@ -9718,7 +9718,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DATE\"", @@ -9742,7 +9742,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DATE\"", @@ -9766,7 +9766,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DATE\"", @@ -10046,7 +10046,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(3,0)\"", @@ -10070,7 +10070,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(5,0)\"", @@ -10094,7 +10094,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(10,0)\"", @@ -10118,7 +10118,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(20,0)\"", @@ -10142,7 +10142,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(3,0)\"", @@ -10166,7 +10166,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(5,0)\"", @@ -10190,7 +10190,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(10,0)\"", @@ -10214,7 +10214,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(20,0)\"", @@ -10238,7 +10238,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(3,0)\"", @@ -10262,7 +10262,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(5,0)\"", @@ -10286,7 +10286,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(10,0)\"", @@ -10310,7 +10310,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(20,0)\"", @@ -10334,7 +10334,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(3,0)\"", @@ -10358,7 +10358,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(5,0)\"", @@ -10382,7 +10382,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(10,0)\"", @@ -10406,7 +10406,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(20,0)\"", @@ -10654,7 +10654,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BINARY\"", @@ -10678,7 +10678,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BINARY\"", @@ -10702,7 +10702,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BINARY\"", @@ -10726,7 +10726,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BINARY\"", @@ -10750,7 +10750,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP\"", @@ -10774,7 +10774,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP\"", @@ -10798,7 +10798,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP\"", @@ -10822,7 +10822,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP\"", @@ -10846,7 +10846,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DATE\"", @@ -10870,7 +10870,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DATE\"", @@ -10894,7 +10894,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DATE\"", @@ -10918,7 +10918,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DATE\"", @@ -11198,7 +11198,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(3,0)\"", @@ -11222,7 +11222,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(5,0)\"", @@ -11246,7 +11246,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(10,0)\"", @@ -11270,7 +11270,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(20,0)\"", @@ -11294,7 +11294,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(3,0)\"", @@ -11318,7 +11318,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(5,0)\"", @@ -11342,7 +11342,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(10,0)\"", @@ -11366,7 +11366,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(20,0)\"", @@ -11390,7 +11390,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(3,0)\"", @@ -11414,7 +11414,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(5,0)\"", @@ -11438,7 +11438,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(10,0)\"", @@ -11462,7 +11462,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(20,0)\"", @@ -11486,7 +11486,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(3,0)\"", @@ -11510,7 +11510,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(5,0)\"", @@ -11534,7 +11534,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(10,0)\"", @@ -11558,7 +11558,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(20,0)\"", @@ -11806,7 +11806,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BINARY\"", @@ -11830,7 +11830,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BINARY\"", @@ -11854,7 +11854,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BINARY\"", @@ -11878,7 +11878,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BINARY\"", @@ -11902,7 +11902,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP\"", @@ -11926,7 +11926,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP\"", @@ -11950,7 +11950,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP\"", @@ -11974,7 +11974,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP\"", @@ -11998,7 +11998,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DATE\"", @@ -12022,7 +12022,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DATE\"", @@ -12046,7 +12046,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DATE\"", @@ -12070,7 +12070,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DATE\"", @@ -12350,7 +12350,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(3,0)\"", @@ -12374,7 +12374,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(5,0)\"", @@ -12398,7 +12398,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(10,0)\"", @@ -12422,7 +12422,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(20,0)\"", @@ -12446,7 +12446,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(3,0)\"", @@ -12470,7 +12470,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(5,0)\"", @@ -12494,7 +12494,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(10,0)\"", @@ -12518,7 +12518,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(20,0)\"", @@ -12542,7 +12542,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(3,0)\"", @@ -12566,7 +12566,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(5,0)\"", @@ -12590,7 +12590,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(10,0)\"", @@ -12614,7 +12614,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(20,0)\"", @@ -12638,7 +12638,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(3,0)\"", @@ -12662,7 +12662,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(5,0)\"", @@ -12686,7 +12686,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(10,0)\"", @@ -12710,7 +12710,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(20,0)\"", @@ -12958,7 +12958,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BINARY\"", @@ -12982,7 +12982,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BINARY\"", @@ -13006,7 +13006,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BINARY\"", @@ -13030,7 +13030,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BINARY\"", @@ -13054,7 +13054,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP\"", @@ -13078,7 +13078,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP\"", @@ -13102,7 +13102,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP\"", @@ -13126,7 +13126,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP\"", @@ -13150,7 +13150,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DATE\"", @@ -13174,7 +13174,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DATE\"", @@ -13198,7 +13198,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DATE\"", @@ -13222,7 +13222,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DATE\"", @@ -13502,7 +13502,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(3,0)\"", @@ -13526,7 +13526,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(5,0)\"", @@ -13550,7 +13550,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(10,0)\"", @@ -13574,7 +13574,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(20,0)\"", @@ -13598,7 +13598,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(3,0)\"", @@ -13622,7 +13622,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(5,0)\"", @@ -13646,7 +13646,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(10,0)\"", @@ -13670,7 +13670,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(20,0)\"", @@ -13694,7 +13694,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(3,0)\"", @@ -13718,7 +13718,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(5,0)\"", @@ -13742,7 +13742,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(10,0)\"", @@ -13766,7 +13766,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(20,0)\"", @@ -13790,7 +13790,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(3,0)\"", @@ -13814,7 +13814,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(5,0)\"", @@ -13838,7 +13838,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(10,0)\"", @@ -13862,7 +13862,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(20,0)\"", @@ -14110,7 +14110,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BINARY\"", @@ -14134,7 +14134,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BINARY\"", @@ -14158,7 +14158,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BINARY\"", @@ -14182,7 +14182,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BINARY\"", @@ -14206,7 +14206,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP\"", @@ -14230,7 +14230,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP\"", @@ -14254,7 +14254,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP\"", @@ -14278,7 +14278,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP\"", @@ -14302,7 +14302,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DATE\"", @@ -14326,7 +14326,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DATE\"", @@ -14350,7 +14350,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DATE\"", @@ -14374,7 +14374,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DATE\"", @@ -14654,7 +14654,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(3,0)\"", @@ -14678,7 +14678,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(5,0)\"", @@ -14702,7 +14702,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(10,0)\"", @@ -14726,7 +14726,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(20,0)\"", @@ -14750,7 +14750,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(3,0)\"", @@ -14774,7 +14774,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(5,0)\"", @@ -14798,7 +14798,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(10,0)\"", @@ -14822,7 +14822,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(20,0)\"", @@ -14846,7 +14846,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(3,0)\"", @@ -14870,7 +14870,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(5,0)\"", @@ -14894,7 +14894,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(10,0)\"", @@ -14918,7 +14918,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(20,0)\"", @@ -14942,7 +14942,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(3,0)\"", @@ -14966,7 +14966,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(5,0)\"", @@ -14990,7 +14990,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(10,0)\"", @@ -15014,7 +15014,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(20,0)\"", diff --git a/sql/core/src/test/resources/sql-tests/results/typeCoercion/native/division.sql.out b/sql/core/src/test/resources/sql-tests/results/typeCoercion/native/division.sql.out index cf6931a4ffda..2c03abf26e9a 100644 --- a/sql/core/src/test/resources/sql-tests/results/typeCoercion/native/division.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/typeCoercion/native/division.sql.out @@ -78,7 +78,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TINYINT\"", @@ -102,7 +102,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TINYINT\"", @@ -126,7 +126,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TINYINT\"", @@ -150,7 +150,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TINYINT\"", @@ -238,7 +238,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"SMALLINT\"", @@ -262,7 +262,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"SMALLINT\"", @@ -286,7 +286,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"SMALLINT\"", @@ -310,7 +310,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"SMALLINT\"", @@ -398,7 +398,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"INT\"", @@ -422,7 +422,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"INT\"", @@ -446,7 +446,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"INT\"", @@ -470,7 +470,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"INT\"", @@ -558,7 +558,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BIGINT\"", @@ -582,7 +582,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BIGINT\"", @@ -606,7 +606,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BIGINT\"", @@ -630,7 +630,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BIGINT\"", @@ -718,7 +718,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"FLOAT\"", @@ -742,7 +742,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"FLOAT\"", @@ -766,7 +766,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"FLOAT\"", @@ -790,7 +790,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"FLOAT\"", @@ -878,7 +878,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DOUBLE\"", @@ -902,7 +902,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DOUBLE\"", @@ -926,7 +926,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DOUBLE\"", @@ -950,7 +950,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DOUBLE\"", @@ -1038,7 +1038,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(10,0)\"", @@ -1062,7 +1062,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(10,0)\"", @@ -1086,7 +1086,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(10,0)\"", @@ -1110,7 +1110,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DECIMAL(10,0)\"", @@ -1190,7 +1190,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"STRING\"", @@ -1214,7 +1214,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"BINARY\"", @@ -1238,7 +1238,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"BOOLEAN\"", @@ -1262,7 +1262,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"TIMESTAMP\"", @@ -1286,7 +1286,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"DATE\"", @@ -1310,7 +1310,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BINARY\"", @@ -1334,7 +1334,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BINARY\"", @@ -1358,7 +1358,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BINARY\"", @@ -1382,7 +1382,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BINARY\"", @@ -1406,7 +1406,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BINARY\"", @@ -1430,7 +1430,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BINARY\"", @@ -1454,7 +1454,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BINARY\"", @@ -1478,7 +1478,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"BINARY\"", @@ -1502,7 +1502,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"BINARY\"", @@ -1526,7 +1526,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BINARY\"", @@ -1550,7 +1550,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BINARY\"", @@ -1574,7 +1574,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BINARY\"", @@ -1598,7 +1598,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BOOLEAN\"", @@ -1622,7 +1622,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BOOLEAN\"", @@ -1646,7 +1646,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BOOLEAN\"", @@ -1670,7 +1670,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BOOLEAN\"", @@ -1694,7 +1694,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BOOLEAN\"", @@ -1718,7 +1718,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BOOLEAN\"", @@ -1742,7 +1742,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BOOLEAN\"", @@ -1766,7 +1766,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"BOOLEAN\"", @@ -1790,7 +1790,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BOOLEAN\"", @@ -1814,7 +1814,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"BOOLEAN\"", @@ -1838,7 +1838,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BOOLEAN\"", @@ -1862,7 +1862,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"BOOLEAN\"", @@ -1886,7 +1886,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP\"", @@ -1910,7 +1910,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP\"", @@ -1934,7 +1934,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP\"", @@ -1958,7 +1958,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP\"", @@ -1982,7 +1982,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP\"", @@ -2006,7 +2006,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP\"", @@ -2030,7 +2030,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP\"", @@ -2054,7 +2054,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"TIMESTAMP\"", @@ -2078,7 +2078,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP\"", @@ -2102,7 +2102,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP\"", @@ -2126,7 +2126,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"TIMESTAMP\"", @@ -2150,7 +2150,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"TIMESTAMP\"", @@ -2174,7 +2174,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DATE\"", @@ -2198,7 +2198,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DATE\"", @@ -2222,7 +2222,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DATE\"", @@ -2246,7 +2246,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DATE\"", @@ -2270,7 +2270,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DATE\"", @@ -2294,7 +2294,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DATE\"", @@ -2318,7 +2318,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DATE\"", @@ -2342,7 +2342,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"DATE\"", @@ -2366,7 +2366,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DATE\"", @@ -2390,7 +2390,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DATE\"", @@ -2414,7 +2414,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"DATE\"", @@ -2438,7 +2438,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"DATE\"", diff --git a/sql/core/src/test/resources/sql-tests/results/typeCoercion/native/ifCoercion.sql.out b/sql/core/src/test/resources/sql-tests/results/typeCoercion/native/ifCoercion.sql.out index b2e0f50028a4..2afdca6d00d7 100644 --- a/sql/core/src/test/resources/sql-tests/results/typeCoercion/native/ifCoercion.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/typeCoercion/native/ifCoercion.sql.out @@ -78,7 +78,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"TINYINT\", \"BINARY\"]", @@ -102,7 +102,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"TINYINT\", \"BOOLEAN\"]", @@ -126,7 +126,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"TINYINT\", \"TIMESTAMP\"]", @@ -150,7 +150,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"TINYINT\", \"DATE\"]", @@ -238,7 +238,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"SMALLINT\", \"BINARY\"]", @@ -262,7 +262,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"SMALLINT\", \"BOOLEAN\"]", @@ -286,7 +286,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"SMALLINT\", \"TIMESTAMP\"]", @@ -310,7 +310,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"SMALLINT\", \"DATE\"]", @@ -398,7 +398,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"INT\", \"BINARY\"]", @@ -422,7 +422,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"INT\", \"BOOLEAN\"]", @@ -446,7 +446,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"INT\", \"TIMESTAMP\"]", @@ -470,7 +470,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"INT\", \"DATE\"]", @@ -558,7 +558,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BIGINT\", \"BINARY\"]", @@ -582,7 +582,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BIGINT\", \"BOOLEAN\"]", @@ -606,7 +606,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BIGINT\", \"TIMESTAMP\"]", @@ -630,7 +630,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BIGINT\", \"DATE\"]", @@ -718,7 +718,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"FLOAT\", \"BINARY\"]", @@ -742,7 +742,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"FLOAT\", \"BOOLEAN\"]", @@ -766,7 +766,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"FLOAT\", \"TIMESTAMP\"]", @@ -790,7 +790,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"FLOAT\", \"DATE\"]", @@ -878,7 +878,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DOUBLE\", \"BINARY\"]", @@ -902,7 +902,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DOUBLE\", \"BOOLEAN\"]", @@ -926,7 +926,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DOUBLE\", \"TIMESTAMP\"]", @@ -950,7 +950,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DOUBLE\", \"DATE\"]", @@ -1038,7 +1038,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DECIMAL(10,0)\", \"BINARY\"]", @@ -1062,7 +1062,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DECIMAL(10,0)\", \"BOOLEAN\"]", @@ -1086,7 +1086,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DECIMAL(10,0)\", \"TIMESTAMP\"]", @@ -1110,7 +1110,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DECIMAL(10,0)\", \"DATE\"]", @@ -1214,7 +1214,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'1'", @@ -1238,7 +1238,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'1'", @@ -1262,7 +1262,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BINARY\", \"TINYINT\"]", @@ -1286,7 +1286,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BINARY\", \"SMALLINT\"]", @@ -1310,7 +1310,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BINARY\", \"INT\"]", @@ -1334,7 +1334,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BINARY\", \"BIGINT\"]", @@ -1358,7 +1358,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BINARY\", \"FLOAT\"]", @@ -1382,7 +1382,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BINARY\", \"DOUBLE\"]", @@ -1406,7 +1406,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BINARY\", \"DECIMAL(10,0)\"]", @@ -1446,7 +1446,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BINARY\", \"BOOLEAN\"]", @@ -1470,7 +1470,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BINARY\", \"TIMESTAMP\"]", @@ -1494,7 +1494,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BINARY\", \"DATE\"]", @@ -1518,7 +1518,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BOOLEAN\", \"TINYINT\"]", @@ -1542,7 +1542,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BOOLEAN\", \"SMALLINT\"]", @@ -1566,7 +1566,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BOOLEAN\", \"INT\"]", @@ -1590,7 +1590,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BOOLEAN\", \"BIGINT\"]", @@ -1614,7 +1614,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BOOLEAN\", \"FLOAT\"]", @@ -1638,7 +1638,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BOOLEAN\", \"DOUBLE\"]", @@ -1662,7 +1662,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BOOLEAN\", \"DECIMAL(10,0)\"]", @@ -1694,7 +1694,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BOOLEAN\", \"BINARY\"]", @@ -1726,7 +1726,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BOOLEAN\", \"TIMESTAMP\"]", @@ -1750,7 +1750,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BOOLEAN\", \"DATE\"]", @@ -1774,7 +1774,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"TIMESTAMP\", \"TINYINT\"]", @@ -1798,7 +1798,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"TIMESTAMP\", \"SMALLINT\"]", @@ -1822,7 +1822,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"TIMESTAMP\", \"INT\"]", @@ -1846,7 +1846,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"TIMESTAMP\", \"BIGINT\"]", @@ -1870,7 +1870,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"TIMESTAMP\", \"FLOAT\"]", @@ -1894,7 +1894,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"TIMESTAMP\", \"DOUBLE\"]", @@ -1918,7 +1918,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"TIMESTAMP\", \"DECIMAL(10,0)\"]", @@ -1950,7 +1950,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"TIMESTAMP\", \"BINARY\"]", @@ -1974,7 +1974,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"TIMESTAMP\", \"BOOLEAN\"]", @@ -2014,7 +2014,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DATE\", \"TINYINT\"]", @@ -2038,7 +2038,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DATE\", \"SMALLINT\"]", @@ -2062,7 +2062,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DATE\", \"INT\"]", @@ -2086,7 +2086,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DATE\", \"BIGINT\"]", @@ -2110,7 +2110,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DATE\", \"FLOAT\"]", @@ -2134,7 +2134,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DATE\", \"DOUBLE\"]", @@ -2158,7 +2158,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DATE\", \"DECIMAL(10,0)\"]", @@ -2190,7 +2190,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DATE\", \"BINARY\"]", @@ -2214,7 +2214,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DATE\", \"BOOLEAN\"]", diff --git a/sql/core/src/test/resources/sql-tests/results/typeCoercion/native/inConversion.sql.out b/sql/core/src/test/resources/sql-tests/results/typeCoercion/native/inConversion.sql.out index 7c9152a66a9c..a71184490b75 100644 --- a/sql/core/src/test/resources/sql-tests/results/typeCoercion/native/inConversion.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/typeCoercion/native/inConversion.sql.out @@ -78,7 +78,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"TINYINT\", \"BINARY\"]", @@ -102,7 +102,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"TINYINT\", \"BOOLEAN\"]", @@ -126,7 +126,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"TINYINT\", \"TIMESTAMP\"]", @@ -150,7 +150,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"TINYINT\", \"DATE\"]", @@ -238,7 +238,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"SMALLINT\", \"BINARY\"]", @@ -262,7 +262,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"SMALLINT\", \"BOOLEAN\"]", @@ -286,7 +286,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"SMALLINT\", \"TIMESTAMP\"]", @@ -310,7 +310,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"SMALLINT\", \"DATE\"]", @@ -398,7 +398,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"INT\", \"BINARY\"]", @@ -422,7 +422,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"INT\", \"BOOLEAN\"]", @@ -446,7 +446,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"INT\", \"TIMESTAMP\"]", @@ -470,7 +470,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"INT\", \"DATE\"]", @@ -558,7 +558,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BIGINT\", \"BINARY\"]", @@ -582,7 +582,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BIGINT\", \"BOOLEAN\"]", @@ -606,7 +606,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BIGINT\", \"TIMESTAMP\"]", @@ -630,7 +630,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BIGINT\", \"DATE\"]", @@ -718,7 +718,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"FLOAT\", \"BINARY\"]", @@ -742,7 +742,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"FLOAT\", \"BOOLEAN\"]", @@ -766,7 +766,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"FLOAT\", \"TIMESTAMP\"]", @@ -790,7 +790,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"FLOAT\", \"DATE\"]", @@ -878,7 +878,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DOUBLE\", \"BINARY\"]", @@ -902,7 +902,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DOUBLE\", \"BOOLEAN\"]", @@ -926,7 +926,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DOUBLE\", \"TIMESTAMP\"]", @@ -950,7 +950,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DOUBLE\", \"DATE\"]", @@ -1038,7 +1038,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DECIMAL(10,0)\", \"BINARY\"]", @@ -1062,7 +1062,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DECIMAL(10,0)\", \"BOOLEAN\"]", @@ -1086,7 +1086,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DECIMAL(10,0)\", \"TIMESTAMP\"]", @@ -1110,7 +1110,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DECIMAL(10,0)\", \"DATE\"]", @@ -1214,7 +1214,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'1'", @@ -1238,7 +1238,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'1'", @@ -1262,7 +1262,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BINARY\", \"TINYINT\"]", @@ -1286,7 +1286,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BINARY\", \"SMALLINT\"]", @@ -1310,7 +1310,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BINARY\", \"INT\"]", @@ -1334,7 +1334,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BINARY\", \"BIGINT\"]", @@ -1358,7 +1358,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BINARY\", \"FLOAT\"]", @@ -1382,7 +1382,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BINARY\", \"DOUBLE\"]", @@ -1406,7 +1406,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BINARY\", \"DECIMAL(10,0)\"]", @@ -1446,7 +1446,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BINARY\", \"BOOLEAN\"]", @@ -1470,7 +1470,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BINARY\", \"TIMESTAMP\"]", @@ -1494,7 +1494,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BINARY\", \"DATE\"]", @@ -1518,7 +1518,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BOOLEAN\", \"TINYINT\"]", @@ -1542,7 +1542,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BOOLEAN\", \"SMALLINT\"]", @@ -1566,7 +1566,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BOOLEAN\", \"INT\"]", @@ -1590,7 +1590,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BOOLEAN\", \"BIGINT\"]", @@ -1614,7 +1614,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BOOLEAN\", \"FLOAT\"]", @@ -1638,7 +1638,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BOOLEAN\", \"DOUBLE\"]", @@ -1662,7 +1662,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BOOLEAN\", \"DECIMAL(10,0)\"]", @@ -1694,7 +1694,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BOOLEAN\", \"BINARY\"]", @@ -1726,7 +1726,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BOOLEAN\", \"TIMESTAMP\"]", @@ -1750,7 +1750,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BOOLEAN\", \"DATE\"]", @@ -1774,7 +1774,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"TIMESTAMP\", \"TINYINT\"]", @@ -1798,7 +1798,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"TIMESTAMP\", \"SMALLINT\"]", @@ -1822,7 +1822,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"TIMESTAMP\", \"INT\"]", @@ -1846,7 +1846,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"TIMESTAMP\", \"BIGINT\"]", @@ -1870,7 +1870,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"TIMESTAMP\", \"FLOAT\"]", @@ -1894,7 +1894,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"TIMESTAMP\", \"DOUBLE\"]", @@ -1918,7 +1918,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"TIMESTAMP\", \"DECIMAL(10,0)\"]", @@ -1942,7 +1942,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'2'", @@ -1966,7 +1966,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"TIMESTAMP\", \"BINARY\"]", @@ -1990,7 +1990,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"TIMESTAMP\", \"BOOLEAN\"]", @@ -2030,7 +2030,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DATE\", \"TINYINT\"]", @@ -2054,7 +2054,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DATE\", \"SMALLINT\"]", @@ -2078,7 +2078,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DATE\", \"INT\"]", @@ -2102,7 +2102,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DATE\", \"BIGINT\"]", @@ -2126,7 +2126,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DATE\", \"FLOAT\"]", @@ -2150,7 +2150,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DATE\", \"DOUBLE\"]", @@ -2174,7 +2174,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DATE\", \"DECIMAL(10,0)\"]", @@ -2198,7 +2198,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'2'", @@ -2222,7 +2222,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DATE\", \"BINARY\"]", @@ -2246,7 +2246,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DATE\", \"BOOLEAN\"]", @@ -2350,7 +2350,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"TINYINT\", \"TINYINT\", \"BINARY\"]", @@ -2374,7 +2374,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"TINYINT\", \"TINYINT\", \"BOOLEAN\"]", @@ -2398,7 +2398,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"TINYINT\", \"TINYINT\", \"TIMESTAMP\"]", @@ -2422,7 +2422,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"TINYINT\", \"TINYINT\", \"DATE\"]", @@ -2510,7 +2510,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"SMALLINT\", \"SMALLINT\", \"BINARY\"]", @@ -2534,7 +2534,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"SMALLINT\", \"SMALLINT\", \"BOOLEAN\"]", @@ -2558,7 +2558,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"SMALLINT\", \"SMALLINT\", \"TIMESTAMP\"]", @@ -2582,7 +2582,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"SMALLINT\", \"SMALLINT\", \"DATE\"]", @@ -2670,7 +2670,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"INT\", \"INT\", \"BINARY\"]", @@ -2694,7 +2694,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"INT\", \"INT\", \"BOOLEAN\"]", @@ -2718,7 +2718,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"INT\", \"INT\", \"TIMESTAMP\"]", @@ -2742,7 +2742,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"INT\", \"INT\", \"DATE\"]", @@ -2830,7 +2830,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BIGINT\", \"BIGINT\", \"BINARY\"]", @@ -2854,7 +2854,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BIGINT\", \"BIGINT\", \"BOOLEAN\"]", @@ -2878,7 +2878,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BIGINT\", \"BIGINT\", \"TIMESTAMP\"]", @@ -2902,7 +2902,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BIGINT\", \"BIGINT\", \"DATE\"]", @@ -2990,7 +2990,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"FLOAT\", \"FLOAT\", \"BINARY\"]", @@ -3014,7 +3014,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"FLOAT\", \"FLOAT\", \"BOOLEAN\"]", @@ -3038,7 +3038,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"FLOAT\", \"FLOAT\", \"TIMESTAMP\"]", @@ -3062,7 +3062,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"FLOAT\", \"FLOAT\", \"DATE\"]", @@ -3150,7 +3150,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DOUBLE\", \"DOUBLE\", \"BINARY\"]", @@ -3174,7 +3174,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DOUBLE\", \"DOUBLE\", \"BOOLEAN\"]", @@ -3198,7 +3198,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DOUBLE\", \"DOUBLE\", \"TIMESTAMP\"]", @@ -3222,7 +3222,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DOUBLE\", \"DOUBLE\", \"DATE\"]", @@ -3310,7 +3310,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DECIMAL(10,0)\", \"DECIMAL(10,0)\", \"BINARY\"]", @@ -3334,7 +3334,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DECIMAL(10,0)\", \"DECIMAL(10,0)\", \"BOOLEAN\"]", @@ -3358,7 +3358,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DECIMAL(10,0)\", \"DECIMAL(10,0)\", \"TIMESTAMP\"]", @@ -3382,7 +3382,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DECIMAL(10,0)\", \"DECIMAL(10,0)\", \"DATE\"]", @@ -3486,7 +3486,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'1'", @@ -3510,7 +3510,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'1'", @@ -3534,7 +3534,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BINARY\", \"BINARY\", \"TINYINT\"]", @@ -3558,7 +3558,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BINARY\", \"BINARY\", \"SMALLINT\"]", @@ -3582,7 +3582,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BINARY\", \"BINARY\", \"INT\"]", @@ -3606,7 +3606,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BINARY\", \"BINARY\", \"BIGINT\"]", @@ -3630,7 +3630,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BINARY\", \"BINARY\", \"FLOAT\"]", @@ -3654,7 +3654,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BINARY\", \"BINARY\", \"DOUBLE\"]", @@ -3678,7 +3678,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BINARY\", \"BINARY\", \"DECIMAL(10,0)\"]", @@ -3718,7 +3718,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BINARY\", \"BINARY\", \"BOOLEAN\"]", @@ -3742,7 +3742,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BINARY\", \"BINARY\", \"TIMESTAMP\"]", @@ -3766,7 +3766,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BINARY\", \"BINARY\", \"DATE\"]", @@ -3790,7 +3790,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BOOLEAN\", \"BOOLEAN\", \"TINYINT\"]", @@ -3814,7 +3814,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BOOLEAN\", \"BOOLEAN\", \"SMALLINT\"]", @@ -3838,7 +3838,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BOOLEAN\", \"BOOLEAN\", \"INT\"]", @@ -3862,7 +3862,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BOOLEAN\", \"BOOLEAN\", \"BIGINT\"]", @@ -3886,7 +3886,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BOOLEAN\", \"BOOLEAN\", \"FLOAT\"]", @@ -3910,7 +3910,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BOOLEAN\", \"BOOLEAN\", \"DOUBLE\"]", @@ -3934,7 +3934,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BOOLEAN\", \"BOOLEAN\", \"DECIMAL(10,0)\"]", @@ -3966,7 +3966,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BOOLEAN\", \"BOOLEAN\", \"BINARY\"]", @@ -3998,7 +3998,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BOOLEAN\", \"BOOLEAN\", \"TIMESTAMP\"]", @@ -4022,7 +4022,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"BOOLEAN\", \"BOOLEAN\", \"DATE\"]", @@ -4046,7 +4046,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"TIMESTAMP\", \"TIMESTAMP\", \"TINYINT\"]", @@ -4070,7 +4070,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"TIMESTAMP\", \"TIMESTAMP\", \"SMALLINT\"]", @@ -4094,7 +4094,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"TIMESTAMP\", \"TIMESTAMP\", \"INT\"]", @@ -4118,7 +4118,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"TIMESTAMP\", \"TIMESTAMP\", \"BIGINT\"]", @@ -4142,7 +4142,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"TIMESTAMP\", \"TIMESTAMP\", \"FLOAT\"]", @@ -4166,7 +4166,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"TIMESTAMP\", \"TIMESTAMP\", \"DOUBLE\"]", @@ -4190,7 +4190,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"TIMESTAMP\", \"TIMESTAMP\", \"DECIMAL(10,0)\"]", @@ -4222,7 +4222,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"TIMESTAMP\", \"TIMESTAMP\", \"BINARY\"]", @@ -4246,7 +4246,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"TIMESTAMP\", \"TIMESTAMP\", \"BOOLEAN\"]", @@ -4286,7 +4286,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DATE\", \"DATE\", \"TINYINT\"]", @@ -4310,7 +4310,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DATE\", \"DATE\", \"SMALLINT\"]", @@ -4334,7 +4334,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DATE\", \"DATE\", \"INT\"]", @@ -4358,7 +4358,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DATE\", \"DATE\", \"BIGINT\"]", @@ -4382,7 +4382,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DATE\", \"DATE\", \"FLOAT\"]", @@ -4406,7 +4406,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DATE\", \"DATE\", \"DOUBLE\"]", @@ -4430,7 +4430,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DATE\", \"DATE\", \"DECIMAL(10,0)\"]", @@ -4462,7 +4462,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DATE\", \"DATE\", \"BINARY\"]", @@ -4486,7 +4486,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "[\"DATE\", \"DATE\", \"BOOLEAN\"]", diff --git a/sql/core/src/test/resources/sql-tests/results/typeCoercion/native/mapZipWith.sql.out b/sql/core/src/test/resources/sql-tests/results/typeCoercion/native/mapZipWith.sql.out index a0bd111f6ba6..59feebe6f10f 100644 --- a/sql/core/src/test/resources/sql-tests/results/typeCoercion/native/mapZipWith.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/typeCoercion/native/mapZipWith.sql.out @@ -128,7 +128,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.MAP_ZIP_WITH_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.MAP_ZIP_WITH_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "functionName" : "`map_zip_with`", @@ -154,7 +154,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.MAP_ZIP_WITH_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.MAP_ZIP_WITH_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "functionName" : "`map_zip_with`", @@ -180,7 +180,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.MAP_ZIP_WITH_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.MAP_ZIP_WITH_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "functionName" : "`map_zip_with`", @@ -206,7 +206,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.MAP_ZIP_WITH_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.MAP_ZIP_WITH_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "functionName" : "`map_zip_with`", diff --git a/sql/core/src/test/resources/sql-tests/results/typeCoercion/native/mapconcat.sql.out b/sql/core/src/test/resources/sql-tests/results/typeCoercion/native/mapconcat.sql.out index 893e9b511986..3ce8a4674867 100644 --- a/sql/core/src/test/resources/sql-tests/results/typeCoercion/native/mapconcat.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/typeCoercion/native/mapconcat.sql.out @@ -82,7 +82,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "(\"MAP\" or \"MAP\")", @@ -108,7 +108,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "(\"MAP\" or \"MAP, ARRAY>\")", @@ -134,7 +134,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "(\"MAP\" or \"MAP\")", @@ -160,7 +160,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "(\"MAP\" or \"MAP, STRUCT>\")", @@ -186,7 +186,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "(\"MAP, STRUCT>\" or \"MAP, ARRAY>\")", @@ -212,7 +212,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.DATA_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "dataType" : "(\"MAP\" or \"MAP, ARRAY>\")", diff --git a/sql/core/src/test/resources/sql-tests/results/typeCoercion/native/promoteStrings.sql.out b/sql/core/src/test/resources/sql-tests/results/typeCoercion/native/promoteStrings.sql.out index a97abd5dd181..2a3b6d8e6cd9 100644 --- a/sql/core/src/test/resources/sql-tests/results/typeCoercion/native/promoteStrings.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/typeCoercion/native/promoteStrings.sql.out @@ -70,7 +70,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"STRING\"", @@ -94,7 +94,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"BINARY\"", @@ -118,7 +118,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"BOOLEAN\"", @@ -142,7 +142,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"TIMESTAMP\"", @@ -166,7 +166,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"CAST(2017-12-11 09:30:00 AS DATE)\"", @@ -248,7 +248,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"STRING\"", @@ -272,7 +272,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"BINARY\"", @@ -296,7 +296,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"BOOLEAN\"", @@ -320,7 +320,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'1'", @@ -344,7 +344,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'1'", @@ -424,7 +424,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"STRING\"", @@ -448,7 +448,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"BINARY\"", @@ -472,7 +472,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"BOOLEAN\"", @@ -496,7 +496,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"TIMESTAMP\"", @@ -520,7 +520,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"DATE\"", @@ -600,7 +600,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"STRING\"", @@ -624,7 +624,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"BINARY\"", @@ -648,7 +648,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"BOOLEAN\"", @@ -672,7 +672,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"TIMESTAMP\"", @@ -696,7 +696,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"DATE\"", @@ -776,7 +776,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"STRING\"", @@ -800,7 +800,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"BINARY\"", @@ -824,7 +824,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"BOOLEAN\"", @@ -848,7 +848,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"TIMESTAMP\"", @@ -872,7 +872,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"DATE\"", @@ -952,7 +952,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"STRING\"", @@ -976,7 +976,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"BINARY\"", @@ -1000,7 +1000,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"BOOLEAN\"", @@ -1024,7 +1024,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"TIMESTAMP\"", @@ -1048,7 +1048,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"DATE\"", @@ -1128,7 +1128,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"BINARY\"", @@ -1152,7 +1152,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"BOOLEAN\"", @@ -1176,7 +1176,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"TIMESTAMP\"", @@ -1200,7 +1200,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"1\"", @@ -1282,7 +1282,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"BINARY\"", @@ -1306,7 +1306,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"BOOLEAN\"", @@ -1330,7 +1330,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'1'", @@ -1354,7 +1354,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'1'", @@ -1434,7 +1434,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"BINARY\"", @@ -1458,7 +1458,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"BOOLEAN\"", @@ -1482,7 +1482,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"TIMESTAMP\"", @@ -1506,7 +1506,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"DATE\"", @@ -1586,7 +1586,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"BINARY\"", @@ -1610,7 +1610,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"BOOLEAN\"", @@ -1634,7 +1634,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"TIMESTAMP\"", @@ -1658,7 +1658,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"DATE\"", @@ -1738,7 +1738,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"BINARY\"", @@ -1762,7 +1762,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"BOOLEAN\"", @@ -1786,7 +1786,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"TIMESTAMP\"", @@ -1810,7 +1810,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"DATE\"", @@ -1890,7 +1890,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"BINARY\"", @@ -1914,7 +1914,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"BOOLEAN\"", @@ -1938,7 +1938,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"TIMESTAMP\"", @@ -1962,7 +1962,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_WRONG_TYPE", "sqlState" : "42K09", "messageParameters" : { "actualDataType" : "\"DATE\"", @@ -2066,7 +2066,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'1'", @@ -2090,7 +2090,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'1'", @@ -2186,7 +2186,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'1'", @@ -2210,7 +2210,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'1'", @@ -2314,7 +2314,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'1'", @@ -2338,7 +2338,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'1'", @@ -2434,7 +2434,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'1'", @@ -2458,7 +2458,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'1'", @@ -2562,7 +2562,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'1'", @@ -2586,7 +2586,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'1'", @@ -2690,7 +2690,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'1'", @@ -2714,7 +2714,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'1'", @@ -2818,7 +2818,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'1'", @@ -2842,7 +2842,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'1'", @@ -2946,7 +2946,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'1'", @@ -2970,7 +2970,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'1'", @@ -3074,7 +3074,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'1'", @@ -3098,7 +3098,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'1'", @@ -3202,7 +3202,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'1'", @@ -3226,7 +3226,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'1'", @@ -3330,7 +3330,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'1'", @@ -3354,7 +3354,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'1'", @@ -3458,7 +3458,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'1'", @@ -3482,7 +3482,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'1'", @@ -3586,7 +3586,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'1'", @@ -3610,7 +3610,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'1'", @@ -3714,7 +3714,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'1'", @@ -3738,7 +3738,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'1'", diff --git a/sql/core/src/test/resources/sql-tests/results/typeCoercion/native/stringCastAndExpressions.sql.out b/sql/core/src/test/resources/sql-tests/results/typeCoercion/native/stringCastAndExpressions.sql.out index ea17c7f5289f..01ba71a6a678 100644 --- a/sql/core/src/test/resources/sql-tests/results/typeCoercion/native/stringCastAndExpressions.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/typeCoercion/native/stringCastAndExpressions.sql.out @@ -14,7 +14,7 @@ struct<> -- !query output org.apache.spark.SparkNumberFormatException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'aa'", @@ -38,7 +38,7 @@ struct<> -- !query output org.apache.spark.SparkNumberFormatException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'aa'", @@ -62,7 +62,7 @@ struct<> -- !query output org.apache.spark.SparkNumberFormatException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'aa'", @@ -86,7 +86,7 @@ struct<> -- !query output org.apache.spark.SparkNumberFormatException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'aa'", @@ -110,7 +110,7 @@ struct<> -- !query output org.apache.spark.SparkNumberFormatException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'aa'", @@ -134,7 +134,7 @@ struct<> -- !query output org.apache.spark.SparkNumberFormatException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'aa'", @@ -158,7 +158,7 @@ struct<> -- !query output org.apache.spark.SparkNumberFormatException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'aa'", @@ -182,7 +182,7 @@ struct<> -- !query output org.apache.spark.SparkRuntimeException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'aa'", @@ -206,7 +206,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'aa'", @@ -230,7 +230,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'aa'", @@ -262,7 +262,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.CAST_WITHOUT_SUGGESTION", + "condition" : "DATATYPE_MISMATCH.CAST_WITHOUT_SUGGESTION", "sqlState" : "42K09", "messageParameters" : { "sqlExpr" : "\"a\"", @@ -286,7 +286,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.CAST_WITHOUT_SUGGESTION", + "condition" : "DATATYPE_MISMATCH.CAST_WITHOUT_SUGGESTION", "sqlState" : "42K09", "messageParameters" : { "sqlExpr" : "\"a\"", @@ -310,7 +310,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.CAST_WITHOUT_SUGGESTION", + "condition" : "DATATYPE_MISMATCH.CAST_WITHOUT_SUGGESTION", "sqlState" : "42K09", "messageParameters" : { "sqlExpr" : "\"a\"", @@ -334,7 +334,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'aa'", @@ -356,7 +356,7 @@ struct<> -- !query output org.apache.spark.SparkUpgradeException { - "errorClass" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION.DATETIME_PATTERN_RECOGNITION", + "condition" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION.DATETIME_PATTERN_RECOGNITION", "sqlState" : "42K0B", "messageParameters" : { "config" : "\"spark.sql.legacy.timeParserPolicy\"", @@ -373,7 +373,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "CANNOT_PARSE_TIMESTAMP", + "condition" : "CANNOT_PARSE_TIMESTAMP", "sqlState" : "22007", "messageParameters" : { "func" : "`try_to_timestamp`", @@ -389,7 +389,7 @@ struct<> -- !query output org.apache.spark.SparkUpgradeException { - "errorClass" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION.DATETIME_PATTERN_RECOGNITION", + "condition" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION.DATETIME_PATTERN_RECOGNITION", "sqlState" : "42K0B", "messageParameters" : { "config" : "\"spark.sql.legacy.timeParserPolicy\"", @@ -406,7 +406,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "CANNOT_PARSE_TIMESTAMP", + "condition" : "CANNOT_PARSE_TIMESTAMP", "sqlState" : "22007", "messageParameters" : { "func" : "`try_to_timestamp`", @@ -422,7 +422,7 @@ struct<> -- !query output org.apache.spark.SparkUpgradeException { - "errorClass" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION.DATETIME_PATTERN_RECOGNITION", + "condition" : "INCONSISTENT_BEHAVIOR_CROSS_VERSION.DATETIME_PATTERN_RECOGNITION", "sqlState" : "42K0B", "messageParameters" : { "config" : "\"spark.sql.legacy.timeParserPolicy\"", @@ -439,7 +439,7 @@ struct<> -- !query output org.apache.spark.SparkNumberFormatException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'aa'", @@ -463,7 +463,7 @@ struct<> -- !query output org.apache.spark.SparkNumberFormatException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'2018-01-01'", @@ -487,7 +487,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'aa'", @@ -511,7 +511,7 @@ struct<> -- !query output org.apache.spark.SparkIllegalArgumentException { - "errorClass" : "ILLEGAL_DAY_OF_WEEK", + "condition" : "ILLEGAL_DAY_OF_WEEK", "sqlState" : "22009", "messageParameters" : { "string" : "aa" @@ -526,7 +526,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'aa'", @@ -566,7 +566,7 @@ struct<> -- !query output org.apache.spark.SparkNumberFormatException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'aa'", diff --git a/sql/core/src/test/resources/sql-tests/results/typeCoercion/native/widenSetOperationTypes.sql.out b/sql/core/src/test/resources/sql-tests/results/typeCoercion/native/widenSetOperationTypes.sql.out index 678d00a39b2d..1c9edf4d9597 100644 --- a/sql/core/src/test/resources/sql-tests/results/typeCoercion/native/widenSetOperationTypes.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/typeCoercion/native/widenSetOperationTypes.sql.out @@ -86,7 +86,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -113,7 +113,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -140,7 +140,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -167,7 +167,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -266,7 +266,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -293,7 +293,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -320,7 +320,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -347,7 +347,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -446,7 +446,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -473,7 +473,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -500,7 +500,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -527,7 +527,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -626,7 +626,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -653,7 +653,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -680,7 +680,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -707,7 +707,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -806,7 +806,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -833,7 +833,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -860,7 +860,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -887,7 +887,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -986,7 +986,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -1013,7 +1013,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -1040,7 +1040,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -1067,7 +1067,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -1166,7 +1166,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -1193,7 +1193,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -1220,7 +1220,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -1247,7 +1247,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -1363,7 +1363,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'1'", @@ -1387,7 +1387,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'1'", @@ -1411,7 +1411,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -1438,7 +1438,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -1465,7 +1465,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -1492,7 +1492,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -1519,7 +1519,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -1546,7 +1546,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -1573,7 +1573,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -1618,7 +1618,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -1645,7 +1645,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -1672,7 +1672,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -1699,7 +1699,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -1726,7 +1726,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -1753,7 +1753,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -1780,7 +1780,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -1807,7 +1807,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -1834,7 +1834,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -1861,7 +1861,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -1888,7 +1888,7 @@ struct<> -- !query output org.apache.spark.SparkRuntimeException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'2'", @@ -1912,7 +1912,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -1947,7 +1947,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -1974,7 +1974,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -2001,7 +2001,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -2028,7 +2028,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -2055,7 +2055,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -2082,7 +2082,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -2109,7 +2109,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -2136,7 +2136,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -2163,7 +2163,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -2190,7 +2190,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'2'", @@ -2214,7 +2214,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -2241,7 +2241,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -2286,7 +2286,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -2313,7 +2313,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -2340,7 +2340,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -2367,7 +2367,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -2394,7 +2394,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -2421,7 +2421,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -2448,7 +2448,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -2475,7 +2475,7 @@ struct<> -- !query output org.apache.spark.SparkDateTimeException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'2'", @@ -2499,7 +2499,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -2526,7 +2526,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", diff --git a/sql/core/src/test/resources/sql-tests/results/typeCoercion/native/windowFrameCoercion.sql.out b/sql/core/src/test/resources/sql-tests/results/typeCoercion/native/windowFrameCoercion.sql.out index 67645cfb732f..4d2ff9c098da 100644 --- a/sql/core/src/test/resources/sql-tests/results/typeCoercion/native/windowFrameCoercion.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/typeCoercion/native/windowFrameCoercion.sql.out @@ -166,7 +166,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.SPECIFIED_WINDOW_FRAME_UNACCEPTED_TYPE", + "condition" : "DATATYPE_MISMATCH.SPECIFIED_WINDOW_FRAME_UNACCEPTED_TYPE", "sqlState" : "42K09", "messageParameters" : { "expectedType" : "(\"NUMERIC\" or \"INTERVAL DAY TO SECOND\" or \"INTERVAL YEAR TO MONTH\" or \"INTERVAL\")", @@ -191,7 +191,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.RANGE_FRAME_INVALID_TYPE", + "condition" : "DATATYPE_MISMATCH.RANGE_FRAME_INVALID_TYPE", "sqlState" : "42K09", "messageParameters" : { "orderSpecType" : "\"BINARY\"", @@ -215,7 +215,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.SPECIFIED_WINDOW_FRAME_UNACCEPTED_TYPE", + "condition" : "DATATYPE_MISMATCH.SPECIFIED_WINDOW_FRAME_UNACCEPTED_TYPE", "sqlState" : "42K09", "messageParameters" : { "expectedType" : "(\"NUMERIC\" or \"INTERVAL DAY TO SECOND\" or \"INTERVAL YEAR TO MONTH\" or \"INTERVAL\")", @@ -240,7 +240,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.RANGE_FRAME_INVALID_TYPE", + "condition" : "DATATYPE_MISMATCH.RANGE_FRAME_INVALID_TYPE", "sqlState" : "42K09", "messageParameters" : { "orderSpecType" : "\"TIMESTAMP\"", diff --git a/sql/core/src/test/resources/sql-tests/results/udaf/udaf-group-analytics.sql.out b/sql/core/src/test/resources/sql-tests/results/udaf/udaf-group-analytics.sql.out index f0be6f436423..0b40a82e8d23 100644 --- a/sql/core/src/test/resources/sql-tests/results/udaf/udaf-group-analytics.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/udaf/udaf-group-analytics.sql.out @@ -132,7 +132,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0051", + "condition" : "_LEGACY_ERROR_TEMP_0051", "messageParameters" : { "element" : "ROLLUP" }, @@ -201,7 +201,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0051", + "condition" : "_LEGACY_ERROR_TEMP_0051", "messageParameters" : { "element" : "CUBE" }, diff --git a/sql/core/src/test/resources/sql-tests/results/udaf/udaf-group-by-ordinal.sql.out b/sql/core/src/test/resources/sql-tests/results/udaf/udaf-group-by-ordinal.sql.out index 45a19ba2c3f1..b99f4a17ba5c 100644 --- a/sql/core/src/test/resources/sql-tests/results/udaf/udaf-group-by-ordinal.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/udaf/udaf-group-by-ordinal.sql.out @@ -93,7 +93,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "GROUP_BY_POS_AGGREGATE", + "condition" : "GROUP_BY_POS_AGGREGATE", "sqlState" : "42903", "messageParameters" : { "aggExpr" : "udaf(b#x) AS `udaf(b)`", @@ -116,7 +116,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "GROUP_BY_POS_AGGREGATE", + "condition" : "GROUP_BY_POS_AGGREGATE", "sqlState" : "42903", "messageParameters" : { "aggExpr" : "(udaf(b#x) + 2) AS `(udaf(b) + 2)`", @@ -342,7 +342,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "GROUP_BY_POS_OUT_OF_RANGE", + "condition" : "GROUP_BY_POS_OUT_OF_RANGE", "sqlState" : "42805", "messageParameters" : { "index" : "-1", @@ -365,7 +365,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "GROUP_BY_POS_AGGREGATE", + "condition" : "GROUP_BY_POS_AGGREGATE", "sqlState" : "42903", "messageParameters" : { "aggExpr" : "udaf(1) AS `udaf(1)`", @@ -388,7 +388,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "GROUP_BY_POS_OUT_OF_RANGE", + "condition" : "GROUP_BY_POS_OUT_OF_RANGE", "sqlState" : "42805", "messageParameters" : { "index" : "-1", @@ -411,7 +411,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "GROUP_BY_POS_AGGREGATE", + "condition" : "GROUP_BY_POS_AGGREGATE", "sqlState" : "42903", "messageParameters" : { "aggExpr" : "udaf(1) AS `udaf(1)`", diff --git a/sql/core/src/test/resources/sql-tests/results/udaf/udaf-group-by.sql.out b/sql/core/src/test/resources/sql-tests/results/udaf/udaf-group-by.sql.out index ad12f8bd03fd..de513f91f643 100644 --- a/sql/core/src/test/resources/sql-tests/results/udaf/udaf-group-by.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/udaf/udaf-group-by.sql.out @@ -16,7 +16,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "MISSING_GROUP_BY", + "condition" : "MISSING_GROUP_BY", "sqlState" : "42803", "queryContext" : [ { "objectType" : "", @@ -54,7 +54,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "MISSING_AGGREGATION", + "condition" : "MISSING_AGGREGATION", "sqlState" : "42803", "messageParameters" : { "expression" : "\"a\"", @@ -117,7 +117,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "MISSING_AGGREGATION", + "condition" : "MISSING_AGGREGATION", "sqlState" : "42803", "messageParameters" : { "expression" : "\"a\"", @@ -145,7 +145,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_PANDAS_UDF_PLACEMENT", + "condition" : "INVALID_PANDAS_UDF_PLACEMENT", "sqlState" : "0A000", "messageParameters" : { "functionList" : "`udaf`" @@ -160,7 +160,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_SQL_SYNTAX.FUNCTION_WITH_UNSUPPORTED_SYNTAX", + "condition" : "INVALID_SQL_SYNTAX.FUNCTION_WITH_UNSUPPORTED_SYNTAX", "sqlState" : "42000", "messageParameters" : { "prettyName" : "`pythonudaf`", @@ -203,7 +203,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`non_existing`", @@ -235,7 +235,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "MISSING_AGGREGATION", + "condition" : "MISSING_AGGREGATION", "sqlState" : "42803", "messageParameters" : { "expression" : "\"k\"", @@ -259,7 +259,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`k`", @@ -341,7 +341,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_PANDAS_UDF_PLACEMENT", + "condition" : "INVALID_PANDAS_UDF_PLACEMENT", "sqlState" : "0A000", "messageParameters" : { "functionList" : "`udaf`" @@ -356,7 +356,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_PANDAS_UDF_PLACEMENT", + "condition" : "INVALID_PANDAS_UDF_PLACEMENT", "sqlState" : "0A000", "messageParameters" : { "functionList" : "`udaf`" @@ -371,7 +371,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_PANDAS_UDF_PLACEMENT", + "condition" : "INVALID_PANDAS_UDF_PLACEMENT", "sqlState" : "0A000", "messageParameters" : { "functionList" : "`udaf`" @@ -386,7 +386,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_PANDAS_UDF_PLACEMENT", + "condition" : "INVALID_PANDAS_UDF_PLACEMENT", "sqlState" : "0A000", "messageParameters" : { "functionList" : "`udaf`" @@ -401,7 +401,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", + "condition" : "DATATYPE_MISMATCH.BINARY_OP_DIFF_TYPES", "sqlState" : "42K09", "messageParameters" : { "left" : "\"INT\"", diff --git a/sql/core/src/test/resources/sql-tests/results/udaf/udaf-grouping-set.sql.out b/sql/core/src/test/resources/sql-tests/results/udaf/udaf-grouping-set.sql.out index db24bc7ec572..4a683070bb04 100644 --- a/sql/core/src/test/resources/sql-tests/results/udaf/udaf-grouping-set.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/udaf/udaf-grouping-set.sql.out @@ -100,7 +100,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'ROLLUP'", @@ -116,7 +116,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'CUBE'", diff --git a/sql/core/src/test/resources/sql-tests/results/udaf/udaf.sql.out b/sql/core/src/test/resources/sql-tests/results/udaf/udaf.sql.out index 3ea85b1369fb..eac5ae09e36b 100644 --- a/sql/core/src/test/resources/sql-tests/results/udaf/udaf.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/udaf/udaf.sql.out @@ -32,7 +32,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", + "condition" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", "sqlState" : "42605", "messageParameters" : { "actualNum" : "2", @@ -65,7 +65,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "CANNOT_LOAD_FUNCTION_CLASS", + "condition" : "CANNOT_LOAD_FUNCTION_CLASS", "sqlState" : "46103", "messageParameters" : { "className" : "test.non.existent.udaf", diff --git a/sql/core/src/test/resources/sql-tests/results/udf/postgreSQL/udf-aggregates_part1.sql.out b/sql/core/src/test/resources/sql-tests/results/udf/postgreSQL/udf-aggregates_part1.sql.out index 6e955c1e4f1e..8d4218750273 100644 --- a/sql/core/src/test/resources/sql-tests/results/udf/postgreSQL/udf-aggregates_part1.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/udf/postgreSQL/udf-aggregates_part1.sql.out @@ -476,7 +476,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "UNSUPPORTED_SUBQUERY_EXPRESSION_CATEGORY.AGGREGATE_FUNCTION_MIXED_OUTER_LOCAL_REFERENCES", + "condition" : "UNSUPPORTED_SUBQUERY_EXPRESSION_CATEGORY.AGGREGATE_FUNCTION_MIXED_OUTER_LOCAL_REFERENCES", "sqlState" : "0A000", "messageParameters" : { "function" : "sum(DISTINCT (outer(a.four) + b.four))" @@ -500,7 +500,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`o`.`unique1`", diff --git a/sql/core/src/test/resources/sql-tests/results/udf/postgreSQL/udf-aggregates_part3.sql.out b/sql/core/src/test/resources/sql-tests/results/udf/postgreSQL/udf-aggregates_part3.sql.out index e31538c16642..00aeffb69003 100644 --- a/sql/core/src/test/resources/sql-tests/results/udf/postgreSQL/udf-aggregates_part3.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/udf/postgreSQL/udf-aggregates_part3.sql.out @@ -6,7 +6,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "NESTED_AGGREGATE_FUNCTION", + "condition" : "NESTED_AGGREGATE_FUNCTION", "sqlState" : "42607", "queryContext" : [ { "objectType" : "", diff --git a/sql/core/src/test/resources/sql-tests/results/udf/postgreSQL/udf-join.sql.out b/sql/core/src/test/resources/sql-tests/results/udf/postgreSQL/udf-join.sql.out index 0eb291c07c74..3a83214d9d36 100644 --- a/sql/core/src/test/resources/sql-tests/results/udf/postgreSQL/udf-join.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/udf/postgreSQL/udf-join.sql.out @@ -547,7 +547,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "AMBIGUOUS_REFERENCE", + "condition" : "AMBIGUOUS_REFERENCE", "sqlState" : "42704", "messageParameters" : { "name" : "`i`", @@ -3278,7 +3278,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "AMBIGUOUS_REFERENCE", + "condition" : "AMBIGUOUS_REFERENCE", "sqlState" : "42704", "messageParameters" : { "name" : "`f1`", @@ -3302,7 +3302,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`y`.`f1`", @@ -3335,7 +3335,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`t1`.`uunique1`", @@ -3359,7 +3359,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`t2`.`uunique1`", @@ -3383,7 +3383,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`uunique1`", @@ -3597,7 +3597,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITHOUT_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITHOUT_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`f1`" @@ -3619,7 +3619,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITHOUT_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITHOUT_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`a`.`f1`" @@ -3641,7 +3641,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITHOUT_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITHOUT_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`f1`" @@ -3663,7 +3663,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITHOUT_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITHOUT_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`a`.`f1`" diff --git a/sql/core/src/test/resources/sql-tests/results/udf/postgreSQL/udf-select_having.sql.out b/sql/core/src/test/resources/sql-tests/results/udf/postgreSQL/udf-select_having.sql.out index a220e6487f81..389555cd04d2 100644 --- a/sql/core/src/test/resources/sql-tests/results/udf/postgreSQL/udf-select_having.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/udf/postgreSQL/udf-select_having.sql.out @@ -141,7 +141,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "MISSING_GROUP_BY", + "condition" : "MISSING_GROUP_BY", "sqlState" : "42803", "queryContext" : [ { "objectType" : "", @@ -160,7 +160,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`a`", @@ -199,7 +199,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "DIVIDE_BY_ZERO", + "condition" : "DIVIDE_BY_ZERO", "sqlState" : "22012", "messageParameters" : { "config" : "\"spark.sql.ansi.enabled\"" diff --git a/sql/core/src/test/resources/sql-tests/results/udf/postgreSQL/udf-select_implicit.sql.out b/sql/core/src/test/resources/sql-tests/results/udf/postgreSQL/udf-select_implicit.sql.out index a3a7cee4eaa7..16f12a11a8ce 100755 --- a/sql/core/src/test/resources/sql-tests/results/udf/postgreSQL/udf-select_implicit.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/udf/postgreSQL/udf-select_implicit.sql.out @@ -123,7 +123,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`b`", @@ -218,7 +218,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "GROUP_BY_POS_OUT_OF_RANGE", + "condition" : "GROUP_BY_POS_OUT_OF_RANGE", "sqlState" : "42805", "messageParameters" : { "index" : "3", @@ -243,7 +243,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "AMBIGUOUS_REFERENCE", + "condition" : "AMBIGUOUS_REFERENCE", "sqlState" : "42704", "messageParameters" : { "name" : "`b`", @@ -370,7 +370,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`b`", @@ -447,7 +447,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "AMBIGUOUS_REFERENCE", + "condition" : "AMBIGUOUS_REFERENCE", "sqlState" : "42704", "messageParameters" : { "name" : "`b`", @@ -486,7 +486,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "AMBIGUOUS_REFERENCE", + "condition" : "AMBIGUOUS_REFERENCE", "sqlState" : "42704", "messageParameters" : { "name" : "`b`", diff --git a/sql/core/src/test/resources/sql-tests/results/udf/udf-except-all.sql.out b/sql/core/src/test/resources/sql-tests/results/udf/udf-except-all.sql.out index ad1b0aabcdec..accf2741af26 100644 --- a/sql/core/src/test/resources/sql-tests/results/udf/udf-except-all.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/udf/udf-except-all.sql.out @@ -139,7 +139,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -229,7 +229,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "NUM_COLUMNS_MISMATCH", + "condition" : "NUM_COLUMNS_MISMATCH", "sqlState" : "42826", "messageParameters" : { "firstNumColumns" : "1", diff --git a/sql/core/src/test/resources/sql-tests/results/udf/udf-group-analytics.sql.out b/sql/core/src/test/resources/sql-tests/results/udf/udf-group-analytics.sql.out index 9beee9972ab7..5f0ae6dc21f2 100644 --- a/sql/core/src/test/resources/sql-tests/results/udf/udf-group-analytics.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/udf/udf-group-analytics.sql.out @@ -208,7 +208,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNSUPPORTED_GROUPING_EXPRESSION", + "condition" : "UNSUPPORTED_GROUPING_EXPRESSION", "sqlState" : "42K0E", "queryContext" : [ { "objectType" : "", @@ -227,7 +227,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNSUPPORTED_GROUPING_EXPRESSION", + "condition" : "UNSUPPORTED_GROUPING_EXPRESSION", "sqlState" : "42K0E", "queryContext" : [ { "objectType" : "", @@ -273,7 +273,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "UNSUPPORTED_GROUPING_EXPRESSION", + "condition" : "UNSUPPORTED_GROUPING_EXPRESSION", "sqlState" : "42K0E" } @@ -285,7 +285,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "UNSUPPORTED_GROUPING_EXPRESSION", + "condition" : "UNSUPPORTED_GROUPING_EXPRESSION", "sqlState" : "42K0E" } @@ -343,7 +343,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "UNSUPPORTED_GROUPING_EXPRESSION", + "condition" : "UNSUPPORTED_GROUPING_EXPRESSION", "sqlState" : "42K0E" } @@ -355,7 +355,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "UNSUPPORTED_GROUPING_EXPRESSION", + "condition" : "UNSUPPORTED_GROUPING_EXPRESSION", "sqlState" : "42K0E" } diff --git a/sql/core/src/test/resources/sql-tests/results/udf/udf-group-by.sql.out b/sql/core/src/test/resources/sql-tests/results/udf/udf-group-by.sql.out index 6a70c8b96841..1ed9e6b00fc9 100644 --- a/sql/core/src/test/resources/sql-tests/results/udf/udf-group-by.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/udf/udf-group-by.sql.out @@ -16,7 +16,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "MISSING_GROUP_BY", + "condition" : "MISSING_GROUP_BY", "sqlState" : "42803", "queryContext" : [ { "objectType" : "", @@ -54,7 +54,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "MISSING_AGGREGATION", + "condition" : "MISSING_AGGREGATION", "sqlState" : "42803", "messageParameters" : { "expression" : "\"a\"", @@ -125,7 +125,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "MISSING_AGGREGATION", + "condition" : "MISSING_AGGREGATION", "sqlState" : "42803", "messageParameters" : { "expression" : "\"a\"", @@ -189,7 +189,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "GROUP_BY_AGGREGATE", + "condition" : "GROUP_BY_AGGREGATE", "sqlState" : "42903", "messageParameters" : { "sqlExpr" : "CAST(udf(cast(count(b) as string)) AS BIGINT)" @@ -220,7 +220,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "MISSING_AGGREGATION", + "condition" : "MISSING_AGGREGATION", "sqlState" : "42803", "messageParameters" : { "expression" : "\"k\"", @@ -244,7 +244,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`k`", @@ -330,7 +330,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "MISSING_GROUP_BY", + "condition" : "MISSING_GROUP_BY", "sqlState" : "42803", "queryContext" : [ { "objectType" : "", @@ -446,7 +446,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"udf(1)\"", @@ -472,7 +472,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"udf(1)\"", @@ -498,7 +498,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"udf(1)\"", @@ -609,7 +609,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INVALID_WHERE_CONDITION", + "condition" : "INVALID_WHERE_CONDITION", "sqlState" : "42903", "messageParameters" : { "condition" : "\"(count(1) > 1)\"", @@ -632,7 +632,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INVALID_WHERE_CONDITION", + "condition" : "INVALID_WHERE_CONDITION", "sqlState" : "42903", "messageParameters" : { "condition" : "\"((count(1) + 1) > 1)\"", @@ -655,7 +655,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INVALID_WHERE_CONDITION", + "condition" : "INVALID_WHERE_CONDITION", "sqlState" : "42903", "messageParameters" : { "condition" : "\"(((k = 1) OR (k = 2)) OR (((count(1) + 1) > 1) OR (max(k) > 1)))\"", diff --git a/sql/core/src/test/resources/sql-tests/results/udf/udf-inline-table.sql.out b/sql/core/src/test/resources/sql-tests/results/udf/udf-inline-table.sql.out index 3e84ec09c215..6ef6a7a413ea 100644 --- a/sql/core/src/test/resources/sql-tests/results/udf/udf-inline-table.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/udf/udf-inline-table.sql.out @@ -95,7 +95,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_INLINE_TABLE.CANNOT_EVALUATE_EXPRESSION_IN_INLINE_TABLE", + "condition" : "INVALID_INLINE_TABLE.CANNOT_EVALUATE_EXPRESSION_IN_INLINE_TABLE", "sqlState" : "42000", "messageParameters" : { "expr" : "\"rand(5)\"" @@ -117,7 +117,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_INLINE_TABLE.NUM_COLUMNS_MISMATCH", + "condition" : "INVALID_INLINE_TABLE.NUM_COLUMNS_MISMATCH", "sqlState" : "42000", "messageParameters" : { "actualNumCols" : "1", @@ -141,7 +141,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_INLINE_TABLE.INCOMPATIBLE_TYPES_IN_INLINE_TABLE", + "condition" : "INVALID_INLINE_TABLE.INCOMPATIBLE_TYPES_IN_INLINE_TABLE", "sqlState" : "42000", "messageParameters" : { "colName" : "`b`" @@ -163,7 +163,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_INLINE_TABLE.NUM_COLUMNS_MISMATCH", + "condition" : "INVALID_INLINE_TABLE.NUM_COLUMNS_MISMATCH", "sqlState" : "42000", "messageParameters" : { "actualNumCols" : "1", @@ -187,7 +187,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "UNRESOLVED_ROUTINE", + "condition" : "UNRESOLVED_ROUTINE", "sqlState" : "42883", "messageParameters" : { "routineName" : "`random_not_exist_func`", @@ -210,7 +210,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_INLINE_TABLE.CANNOT_EVALUATE_EXPRESSION_IN_INLINE_TABLE", + "condition" : "INVALID_INLINE_TABLE.CANNOT_EVALUATE_EXPRESSION_IN_INLINE_TABLE", "sqlState" : "42000", "messageParameters" : { "expr" : "\"count(1)\"" diff --git a/sql/core/src/test/resources/sql-tests/results/udf/udf-intersect-all.sql.out b/sql/core/src/test/resources/sql-tests/results/udf/udf-intersect-all.sql.out index 240469b0cab6..3b85f230b31f 100644 --- a/sql/core/src/test/resources/sql-tests/results/udf/udf-intersect-all.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/udf/udf-intersect-all.sql.out @@ -96,7 +96,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "INCOMPATIBLE_COLUMN_TYPE", + "condition" : "INCOMPATIBLE_COLUMN_TYPE", "sqlState" : "42825", "messageParameters" : { "columnOrdinalNumber" : "first", @@ -125,7 +125,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "NUM_COLUMNS_MISMATCH", + "condition" : "NUM_COLUMNS_MISMATCH", "sqlState" : "42826", "messageParameters" : { "firstNumColumns" : "1", diff --git a/sql/core/src/test/resources/sql-tests/results/udf/udf-pivot.sql.out b/sql/core/src/test/resources/sql-tests/results/udf/udf-pivot.sql.out index 50376fb86a50..5e2fcd9717ec 100644 --- a/sql/core/src/test/resources/sql-tests/results/udf/udf-pivot.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/udf/udf-pivot.sql.out @@ -200,7 +200,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "_LEGACY_ERROR_TEMP_1006", + "condition" : "_LEGACY_ERROR_TEMP_1006", "messageParameters" : { "sql" : "coursesales.earnings" } @@ -220,7 +220,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "_LEGACY_ERROR_TEMP_1006", + "condition" : "_LEGACY_ERROR_TEMP_1006", "messageParameters" : { "sql" : "__auto_generated_subquery_name.year" } @@ -240,7 +240,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`year`", @@ -284,7 +284,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "NESTED_AGGREGATE_FUNCTION", + "condition" : "NESTED_AGGREGATE_FUNCTION", "sqlState" : "42607", "queryContext" : [ { "objectType" : "", @@ -345,7 +345,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "PIVOT_VALUE_DATA_TYPE_MISMATCH", + "condition" : "PIVOT_VALUE_DATA_TYPE_MISMATCH", "sqlState" : "42K09", "messageParameters" : { "pivotType" : "struct", @@ -366,7 +366,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`s`", @@ -393,7 +393,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "NON_LITERAL_PIVOT_VALUES", + "condition" : "NON_LITERAL_PIVOT_VALUES", "sqlState" : "42K08", "messageParameters" : { "expression" : "\"course\"" @@ -484,7 +484,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "INCOMPARABLE_PIVOT_COLUMN", + "condition" : "INCOMPARABLE_PIVOT_COLUMN", "sqlState" : "42818", "messageParameters" : { "columnName" : "`m`" @@ -507,7 +507,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "INCOMPARABLE_PIVOT_COLUMN", + "condition" : "INCOMPARABLE_PIVOT_COLUMN", "sqlState" : "42818", "messageParameters" : { "columnName" : "`named_struct('course', __auto_generated_subquery_name`.`course, 'm', __auto_generated_subquery_name`.`m)`" diff --git a/sql/core/src/test/resources/sql-tests/results/udf/udf-udaf.sql.out b/sql/core/src/test/resources/sql-tests/results/udf/udf-udaf.sql.out index ad8b42d9a5db..f8e04b34d175 100644 --- a/sql/core/src/test/resources/sql-tests/results/udf/udf-udaf.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/udf/udf-udaf.sql.out @@ -32,7 +32,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", + "condition" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", "sqlState" : "42605", "messageParameters" : { "actualNum" : "2", @@ -65,7 +65,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "CANNOT_LOAD_FUNCTION_CLASS", + "condition" : "CANNOT_LOAD_FUNCTION_CLASS", "sqlState" : "46103", "messageParameters" : { "className" : "test.non.existent.udaf", diff --git a/sql/core/src/test/resources/sql-tests/results/udf/udf-union.sql.out b/sql/core/src/test/resources/sql-tests/results/udf/udf-union.sql.out index 7f2931c3ade0..2429f82b6dc2 100644 --- a/sql/core/src/test/resources/sql-tests/results/udf/udf-union.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/udf/udf-union.sql.out @@ -41,7 +41,7 @@ struct<> -- !query output org.apache.spark.SparkNumberFormatException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'a'", @@ -121,7 +121,7 @@ struct<> -- !query output org.apache.spark.SparkNumberFormatException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'str'", @@ -147,7 +147,7 @@ struct<> -- !query output org.apache.spark.SparkNumberFormatException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'str'", diff --git a/sql/core/src/test/resources/sql-tests/results/udf/udf-window.sql.out b/sql/core/src/test/resources/sql-tests/results/udf/udf-window.sql.out index 40e24e7b4e87..18d38da39b8a 100644 --- a/sql/core/src/test/resources/sql-tests/results/udf/udf-window.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/udf/udf-window.sql.out @@ -59,7 +59,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.SPECIFIED_WINDOW_FRAME_UNACCEPTED_TYPE", + "condition" : "DATATYPE_MISMATCH.SPECIFIED_WINDOW_FRAME_UNACCEPTED_TYPE", "sqlState" : "42K09", "messageParameters" : { "expectedType" : "\"INT\"", @@ -205,7 +205,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.SPECIFIED_WINDOW_FRAME_INVALID_BOUND", + "condition" : "DATATYPE_MISMATCH.SPECIFIED_WINDOW_FRAME_INVALID_BOUND", "sqlState" : "42K09", "messageParameters" : { "lower" : "\"UNBOUNDED FOLLOWING\"", @@ -230,7 +230,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.RANGE_FRAME_WITHOUT_ORDER", + "condition" : "DATATYPE_MISMATCH.RANGE_FRAME_WITHOUT_ORDER", "sqlState" : "42K09", "messageParameters" : { "sqlExpr" : "\"(PARTITION BY udf(cate) RANGE BETWEEN CURRENT ROW AND 1 FOLLOWING)\"" @@ -253,7 +253,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.RANGE_FRAME_MULTI_ORDER", + "condition" : "DATATYPE_MISMATCH.RANGE_FRAME_MULTI_ORDER", "sqlState" : "42K09", "messageParameters" : { "orderSpec" : "cast(udf(cast(val#x as string)) as int) ASC NULLS FIRST,cate#x ASC NULLS FIRST", @@ -277,7 +277,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.RANGE_FRAME_INVALID_TYPE", + "condition" : "DATATYPE_MISMATCH.RANGE_FRAME_INVALID_TYPE", "sqlState" : "42K09", "messageParameters" : { "orderSpecType" : "\"TIMESTAMP\"", @@ -302,7 +302,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.SPECIFIED_WINDOW_FRAME_WRONG_COMPARISON", + "condition" : "DATATYPE_MISMATCH.SPECIFIED_WINDOW_FRAME_WRONG_COMPARISON", "sqlState" : "42K09", "messageParameters" : { "comparison" : "less than or equal", @@ -326,7 +326,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0064", + "condition" : "_LEGACY_ERROR_TEMP_0064", "messageParameters" : { "msg" : "Frame bound value must be a literal." }, @@ -383,7 +383,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "DIVIDE_BY_ZERO", + "condition" : "DIVIDE_BY_ZERO", "sqlState" : "22012", "messageParameters" : { "config" : "\"spark.sql.ansi.enabled\"" @@ -421,7 +421,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "_LEGACY_ERROR_TEMP_1037", + "condition" : "_LEGACY_ERROR_TEMP_1037", "messageParameters" : { "wf" : "row_number()" } diff --git a/sql/core/src/test/resources/sql-tests/results/udtf/udtf.sql.out b/sql/core/src/test/resources/sql-tests/results/udtf/udtf.sql.out index b20516151146..deb634c47a48 100644 --- a/sql/core/src/test/resources/sql-tests/results/udtf/udtf.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/udtf/udtf.sql.out @@ -145,7 +145,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNSUPPORTED_SUBQUERY_EXPRESSION_CATEGORY.NON_DETERMINISTIC_LATERAL_SUBQUERIES", + "condition" : "UNSUPPORTED_SUBQUERY_EXPRESSION_CATEGORY.NON_DETERMINISTIC_LATERAL_SUBQUERIES", "sqlState" : "0A000", "messageParameters" : { "treeNode" : "LateralJoin lateral-subquery#x [], Inner\n: +- Project [count#x, total#x, last#x]\n: +- LateralJoin lateral-subquery#x [c#x], Inner\n: : +- SubqueryAlias __auto_generated_subquery_name_1\n: : +- Generate UDTFCountSumLast(outer(c#x))#x, false, [count#x, total#x, last#x]\n: : +- OneRowRelation\n: +- SubqueryAlias __auto_generated_subquery_name_0\n: +- Project [named_struct(partition_col, partition_col#x, input, input#x, partition_by_0, partition_by_0#x) AS c#x]\n: +- Sort [partition_by_0#x ASC NULLS FIRST, input#x DESC NULLS LAST], false\n: +- RepartitionByExpression [partition_by_0#x]\n: +- Project [partition_col#x, input#x, partition_col#x AS partition_by_0#x]\n: +- SubqueryAlias t2\n: +- View (`t2`, [partition_col#x, input#x])\n: +- Project [cast(partition_col#x as int) AS partition_col#x, cast(input#x as int) AS input#x]\n: +- SubqueryAlias t\n: +- LocalRelation [partition_col#x, input#x]\n+- SubqueryAlias t\n +- LocalRelation [col#x]\n" @@ -183,7 +183,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "TABLE_VALUED_FUNCTION_REQUIRED_METADATA_INCOMPATIBLE_WITH_CALL", + "condition" : "TABLE_VALUED_FUNCTION_REQUIRED_METADATA_INCOMPATIBLE_WITH_CALL", "sqlState" : "22023", "messageParameters" : { "functionName" : "UDTFWithSinglePartition", @@ -207,7 +207,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "TABLE_VALUED_FUNCTION_REQUIRED_METADATA_INCOMPATIBLE_WITH_CALL", + "condition" : "TABLE_VALUED_FUNCTION_REQUIRED_METADATA_INCOMPATIBLE_WITH_CALL", "sqlState" : "22023", "messageParameters" : { "functionName" : "UDTFWithSinglePartition", @@ -234,7 +234,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "TABLE_VALUED_FUNCTION_REQUIRED_METADATA_INCOMPATIBLE_WITH_CALL", + "condition" : "TABLE_VALUED_FUNCTION_REQUIRED_METADATA_INCOMPATIBLE_WITH_CALL", "sqlState" : "22023", "messageParameters" : { "functionName" : "UDTFWithSinglePartition", @@ -267,7 +267,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "TABLE_VALUED_FUNCTION_REQUIRED_METADATA_INCOMPATIBLE_WITH_CALL", + "condition" : "TABLE_VALUED_FUNCTION_REQUIRED_METADATA_INCOMPATIBLE_WITH_CALL", "sqlState" : "22023", "messageParameters" : { "functionName" : "UDTFPartitionByOrderBy", @@ -291,7 +291,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "TABLE_VALUED_FUNCTION_REQUIRED_METADATA_INCOMPATIBLE_WITH_CALL", + "condition" : "TABLE_VALUED_FUNCTION_REQUIRED_METADATA_INCOMPATIBLE_WITH_CALL", "sqlState" : "22023", "messageParameters" : { "functionName" : "UDTFPartitionByOrderBy", @@ -318,7 +318,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "TABLE_VALUED_FUNCTION_REQUIRED_METADATA_INCOMPATIBLE_WITH_CALL", + "condition" : "TABLE_VALUED_FUNCTION_REQUIRED_METADATA_INCOMPATIBLE_WITH_CALL", "sqlState" : "22023", "messageParameters" : { "functionName" : "UDTFPartitionByOrderBy", @@ -382,7 +382,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`unparsable`", @@ -405,7 +405,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "TABLE_VALUED_FUNCTION_FAILED_TO_ANALYZE_IN_PYTHON", + "condition" : "TABLE_VALUED_FUNCTION_FAILED_TO_ANALYZE_IN_PYTHON", "sqlState" : "38000", "messageParameters" : { "msg" : "Failed to evaluate the user-defined table function 'UDTFInvalidSelectExprStringValue' because the static 'analyze' method returned an 'AnalyzeResult' object with the 'select' field set to a value besides a list or tuple of 'SelectedColumn' objects. Please update the table function and then try the query again." @@ -427,7 +427,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "UDTF_INVALID_REQUESTED_SELECTED_EXPRESSION_FROM_ANALYZE_METHOD_REQUIRES_ALIAS", + "condition" : "UDTF_INVALID_REQUESTED_SELECTED_EXPRESSION_FROM_ANALYZE_METHOD_REQUIRES_ALIAS", "sqlState" : "42802", "messageParameters" : { "expression" : "(input + 1)" @@ -449,7 +449,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "UDTF_INVALID_ALIAS_IN_REQUESTED_ORDERING_STRING_FROM_ANALYZE_METHOD", + "condition" : "UDTF_INVALID_ALIAS_IN_REQUESTED_ORDERING_STRING_FROM_ANALYZE_METHOD", "sqlState" : "42802", "messageParameters" : { "aliasName" : "ASC" @@ -471,7 +471,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "TABLE_VALUED_FUNCTION_FAILED_TO_ANALYZE_IN_PYTHON", + "condition" : "TABLE_VALUED_FUNCTION_FAILED_TO_ANALYZE_IN_PYTHON", "sqlState" : "38000", "messageParameters" : { "msg" : "Failed to evaluate the user-defined table function 'UDTFInvalidOrderByStringList' because the static 'analyze' method returned an 'AnalyzeResult' object with the 'orderBy' field set to a value besides a list or tuple of 'OrderingColumn' objects. Please update the table function and then try the query again." @@ -493,7 +493,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "TABLE_VALUED_FUNCTION_REQUIRED_METADATA_INVALID", + "condition" : "TABLE_VALUED_FUNCTION_REQUIRED_METADATA_INVALID", "sqlState" : "22023", "messageParameters" : { "functionName" : "UDTFInvalidPartitionByAndWithSinglePartition", @@ -516,7 +516,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "TABLE_VALUED_FUNCTION_REQUIRED_METADATA_INVALID", + "condition" : "TABLE_VALUED_FUNCTION_REQUIRED_METADATA_INVALID", "sqlState" : "22023", "messageParameters" : { "functionName" : "UDTFInvalidPartitionByAndWithSinglePartition", @@ -539,7 +539,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "TABLE_VALUED_FUNCTION_REQUIRED_METADATA_INVALID", + "condition" : "TABLE_VALUED_FUNCTION_REQUIRED_METADATA_INVALID", "sqlState" : "22023", "messageParameters" : { "functionName" : "UDTFInvalidPartitionByAndWithSinglePartition", @@ -565,7 +565,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "TABLE_VALUED_FUNCTION_REQUIRED_METADATA_INVALID", + "condition" : "TABLE_VALUED_FUNCTION_REQUIRED_METADATA_INVALID", "sqlState" : "22023", "messageParameters" : { "functionName" : "UDTFInvalidPartitionByAndWithSinglePartition", @@ -588,7 +588,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "TABLE_VALUED_FUNCTION_REQUIRED_METADATA_INVALID", + "condition" : "TABLE_VALUED_FUNCTION_REQUIRED_METADATA_INVALID", "sqlState" : "22023", "messageParameters" : { "functionName" : "UDTFInvalidOrderByWithoutPartitionBy", @@ -611,7 +611,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "TABLE_VALUED_FUNCTION_REQUIRED_METADATA_INVALID", + "condition" : "TABLE_VALUED_FUNCTION_REQUIRED_METADATA_INVALID", "sqlState" : "22023", "messageParameters" : { "functionName" : "UDTFInvalidOrderByWithoutPartitionBy", @@ -634,7 +634,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "TABLE_VALUED_FUNCTION_REQUIRED_METADATA_INVALID", + "condition" : "TABLE_VALUED_FUNCTION_REQUIRED_METADATA_INVALID", "sqlState" : "22023", "messageParameters" : { "functionName" : "UDTFInvalidOrderByWithoutPartitionBy", @@ -660,7 +660,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "TABLE_VALUED_FUNCTION_REQUIRED_METADATA_INVALID", + "condition" : "TABLE_VALUED_FUNCTION_REQUIRED_METADATA_INVALID", "sqlState" : "22023", "messageParameters" : { "functionName" : "UDTFInvalidOrderByWithoutPartitionBy", @@ -782,7 +782,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "TABLE_VALUED_FUNCTION_FAILED_TO_ANALYZE_IN_PYTHON", + "condition" : "TABLE_VALUED_FUNCTION_FAILED_TO_ANALYZE_IN_PYTHON", "sqlState" : "38000", "messageParameters" : { "msg" : "Failed to evaluate the user-defined table function 'UDTFForwardStateFromAnalyzeWithKwargs' because the function arguments did not match the expected signature of the static 'analyze' method (too many positional arguments). Please update the query so that this table function call provides arguments matching the expected signature, or else update the table function so that its static 'analyze' method accepts the provided arguments, and then try the query again." @@ -813,7 +813,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "DUPLICATE_ROUTINE_PARAMETER_ASSIGNMENT.DOUBLE_NAMED_ARGUMENT_REFERENCE", + "condition" : "DUPLICATE_ROUTINE_PARAMETER_ASSIGNMENT.DOUBLE_NAMED_ARGUMENT_REFERENCE", "sqlState" : "4274K", "messageParameters" : { "parameterName" : "`argument`", @@ -836,7 +836,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "TABLE_VALUED_FUNCTION_FAILED_TO_ANALYZE_IN_PYTHON", + "condition" : "TABLE_VALUED_FUNCTION_FAILED_TO_ANALYZE_IN_PYTHON", "sqlState" : "38000", "messageParameters" : { "msg" : "Failed to evaluate the user-defined table function 'InvalidAnalyzeMethodWithSinglePartitionNoInputTable' because the static 'analyze' method returned an 'AnalyzeResult' object with the 'withSinglePartition' field set to 'true', but the function call did not provide any table argument. Please update the query so that it provides a table argument, or else update the table function so that its 'analyze' method returns an 'AnalyzeResult' object with the 'withSinglePartition' field set to 'false', and then try the query again." @@ -858,7 +858,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "TABLE_VALUED_FUNCTION_FAILED_TO_ANALYZE_IN_PYTHON", + "condition" : "TABLE_VALUED_FUNCTION_FAILED_TO_ANALYZE_IN_PYTHON", "sqlState" : "38000", "messageParameters" : { "msg" : "Failed to evaluate the user-defined table function 'InvalidAnalyzeMethodWithPartitionByNoInputTable' because the static 'analyze' method returned an 'AnalyzeResult' object with the 'partitionBy' list set to non-empty, but the function call did not provide any table argument. Please update the query so that it provides a table argument, or else update the table function so that its 'analyze' method returns an 'AnalyzeResult' object with the 'partitionBy' list set to empty, and then try the query again." @@ -880,7 +880,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "TABLE_VALUED_FUNCTION_FAILED_TO_ANALYZE_IN_PYTHON", + "condition" : "TABLE_VALUED_FUNCTION_FAILED_TO_ANALYZE_IN_PYTHON", "sqlState" : "38000", "messageParameters" : { "msg" : "Failed to evaluate the user-defined table function 'InvalidAnalyzeMethodReturnsNonStructTypeSchema' because the static 'analyze' method expects a result of type pyspark.sql.udtf.AnalyzeResult with a 'schema' field comprising a StructType, but the 'schema' field had the wrong type: " @@ -902,7 +902,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "TABLE_VALUED_FUNCTION_FAILED_TO_ANALYZE_IN_PYTHON", + "condition" : "TABLE_VALUED_FUNCTION_FAILED_TO_ANALYZE_IN_PYTHON", "sqlState" : "38000", "messageParameters" : { "msg" : "Failed to evaluate the user-defined table function 'InvalidAnalyzeMethodWithPartitionByListOfStrings' because the static 'analyze' method returned an 'AnalyzeResult' object with the 'partitionBy' field set to a value besides a list or tuple of 'PartitioningColumn' objects. Please update the table function and then try the query again." @@ -942,7 +942,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "TABLE_VALUED_FUNCTION_FAILED_TO_ANALYZE_IN_PYTHON", + "condition" : "TABLE_VALUED_FUNCTION_FAILED_TO_ANALYZE_IN_PYTHON", "sqlState" : "38000", "messageParameters" : { "msg" : "Failed to evaluate the user-defined table function 'UDTFWithSinglePartition' because the function arguments did not match the expected signature of the static 'analyze' method (missing a required argument: 'input_table'). Please update the query so that this table function call provides arguments matching the expected signature, or else update the table function so that its static 'analyze' method accepts the provided arguments, and then try the query again." @@ -964,7 +964,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "TABLE_VALUED_FUNCTION_FAILED_TO_ANALYZE_IN_PYTHON", + "condition" : "TABLE_VALUED_FUNCTION_FAILED_TO_ANALYZE_IN_PYTHON", "sqlState" : "38000", "messageParameters" : { "msg" : "Failed to evaluate the user-defined table function 'UDTFWithSinglePartition' because the function arguments did not match the expected signature of the static 'analyze' method (too many positional arguments). Please update the query so that this table function call provides arguments matching the expected signature, or else update the table function so that its static 'analyze' method accepts the provided arguments, and then try the query again." @@ -986,7 +986,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "TABLE_VALUED_FUNCTION_FAILED_TO_ANALYZE_IN_PYTHON", + "condition" : "TABLE_VALUED_FUNCTION_FAILED_TO_ANALYZE_IN_PYTHON", "sqlState" : "38000", "messageParameters" : { "msg" : "Failed to evaluate the user-defined table function 'UDTFWithSinglePartition' because the function arguments did not match the expected signature of the static 'analyze' method (missing a required argument: 'input_table'). Please update the query so that this table function call provides arguments matching the expected signature, or else update the table function so that its static 'analyze' method accepts the provided arguments, and then try the query again." @@ -1008,7 +1008,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "TABLE_VALUED_FUNCTION_FAILED_TO_ANALYZE_IN_PYTHON", + "condition" : "TABLE_VALUED_FUNCTION_FAILED_TO_ANALYZE_IN_PYTHON", "sqlState" : "38000", "messageParameters" : { "msg" : "Failed to evaluate the user-defined table function 'UDTFWithSinglePartition' because the function arguments did not match the expected signature of the static 'analyze' method (multiple values for argument 'initial_count'). Please update the query so that this table function call provides arguments matching the expected signature, or else update the table function so that its static 'analyze' method accepts the provided arguments, and then try the query again." @@ -1030,7 +1030,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "DUPLICATE_ROUTINE_PARAMETER_ASSIGNMENT.DOUBLE_NAMED_ARGUMENT_REFERENCE", + "condition" : "DUPLICATE_ROUTINE_PARAMETER_ASSIGNMENT.DOUBLE_NAMED_ARGUMENT_REFERENCE", "sqlState" : "4274K", "messageParameters" : { "parameterName" : "`initial_count`", @@ -1053,7 +1053,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITH_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`unparsable`", @@ -1103,7 +1103,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "UNRESOLVED_COLUMN.WITHOUT_SUGGESTION", + "condition" : "UNRESOLVED_COLUMN.WITHOUT_SUGGESTION", "sqlState" : "42703", "messageParameters" : { "objectName" : "`unresolved_column`" diff --git a/sql/core/src/test/resources/sql-tests/results/union.sql.out b/sql/core/src/test/resources/sql-tests/results/union.sql.out index d7db2163f8b8..552c27748c83 100644 --- a/sql/core/src/test/resources/sql-tests/results/union.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/union.sql.out @@ -41,7 +41,7 @@ struct<> -- !query output org.apache.spark.SparkNumberFormatException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'a'", @@ -121,7 +121,7 @@ struct<> -- !query output org.apache.spark.SparkNumberFormatException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'str'", @@ -147,7 +147,7 @@ struct<> -- !query output org.apache.spark.SparkNumberFormatException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'str'", diff --git a/sql/core/src/test/resources/sql-tests/results/url-functions.sql.out b/sql/core/src/test/resources/sql-tests/results/url-functions.sql.out index 2d1daee8500a..acc9e18a6a06 100644 --- a/sql/core/src/test/resources/sql-tests/results/url-functions.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/url-functions.sql.out @@ -102,7 +102,7 @@ struct<> -- !query output org.apache.spark.SparkIllegalArgumentException { - "errorClass" : "CANNOT_DECODE_URL", + "condition" : "CANNOT_DECODE_URL", "sqlState" : "22546", "messageParameters" : { "url" : "http%3A%2F%2spark.apache.org" diff --git a/sql/core/src/test/resources/sql-tests/results/view-schema-binding-config.sql.out b/sql/core/src/test/resources/sql-tests/results/view-schema-binding-config.sql.out index 80260f8fcaa7..d59fc412d3f5 100644 --- a/sql/core/src/test/resources/sql-tests/results/view-schema-binding-config.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/view-schema-binding-config.sql.out @@ -22,7 +22,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "FEATURE_NOT_ENABLED", + "condition" : "FEATURE_NOT_ENABLED", "sqlState" : "56038", "messageParameters" : { "configKey" : "spark.sql.legacy.viewSchemaBindingMode", @@ -46,7 +46,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "FEATURE_NOT_ENABLED", + "condition" : "FEATURE_NOT_ENABLED", "sqlState" : "56038", "messageParameters" : { "configKey" : "spark.sql.legacy.viewSchemaBindingMode", @@ -70,7 +70,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "FEATURE_NOT_ENABLED", + "condition" : "FEATURE_NOT_ENABLED", "sqlState" : "56038", "messageParameters" : { "configKey" : "spark.sql.legacy.viewSchemaBindingMode", @@ -94,7 +94,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "FEATURE_NOT_ENABLED", + "condition" : "FEATURE_NOT_ENABLED", "sqlState" : "56038", "messageParameters" : { "configKey" : "spark.sql.legacy.viewSchemaBindingMode", @@ -447,7 +447,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "CANNOT_UP_CAST_DATATYPE", + "condition" : "CANNOT_UP_CAST_DATATYPE", "sqlState" : "42846", "messageParameters" : { "details" : "The type path of the target object is:\n\nYou can either add an explicit cast to the input data or choose a higher precision type of the field in the target object", @@ -698,7 +698,7 @@ struct<> -- !query output org.apache.spark.SparkNumberFormatException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'a'", @@ -738,7 +738,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.CAST_WITHOUT_SUGGESTION", + "condition" : "DATATYPE_MISMATCH.CAST_WITHOUT_SUGGESTION", "sqlState" : "42K09", "messageParameters" : { "sqlExpr" : "\"c1\"", @@ -863,7 +863,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "INCOMPATIBLE_VIEW_SCHEMA_CHANGE", + "condition" : "INCOMPATIBLE_VIEW_SCHEMA_CHANGE", "sqlState" : "51024", "messageParameters" : { "actualCols" : "[]", @@ -921,7 +921,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "INCOMPATIBLE_VIEW_SCHEMA_CHANGE", + "condition" : "INCOMPATIBLE_VIEW_SCHEMA_CHANGE", "sqlState" : "51024", "messageParameters" : { "actualCols" : "[]", diff --git a/sql/core/src/test/resources/sql-tests/results/view-schema-binding.sql.out b/sql/core/src/test/resources/sql-tests/results/view-schema-binding.sql.out index b0311699d1b5..3e76ba1db9d0 100644 --- a/sql/core/src/test/resources/sql-tests/results/view-schema-binding.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/view-schema-binding.sql.out @@ -76,7 +76,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "CANNOT_UP_CAST_DATATYPE", + "condition" : "CANNOT_UP_CAST_DATATYPE", "sqlState" : "42846", "messageParameters" : { "details" : "The type path of the target object is:\n\nYou can either add an explicit cast to the input data or choose a higher precision type of the field in the target object", @@ -187,7 +187,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "INCOMPATIBLE_VIEW_SCHEMA_CHANGE", + "condition" : "INCOMPATIBLE_VIEW_SCHEMA_CHANGE", "sqlState" : "51024", "messageParameters" : { "actualCols" : "[]", @@ -337,7 +337,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "CANNOT_UP_CAST_DATATYPE", + "condition" : "CANNOT_UP_CAST_DATATYPE", "sqlState" : "42846", "messageParameters" : { "details" : "The type path of the target object is:\n\nYou can either add an explicit cast to the input data or choose a higher precision type of the field in the target object", diff --git a/sql/core/src/test/resources/sql-tests/results/view-schema-compensation.sql.out b/sql/core/src/test/resources/sql-tests/results/view-schema-compensation.sql.out index 94948d1a9a24..330c151b051d 100644 --- a/sql/core/src/test/resources/sql-tests/results/view-schema-compensation.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/view-schema-compensation.sql.out @@ -184,7 +184,7 @@ struct<> -- !query output org.apache.spark.SparkNumberFormatException { - "errorClass" : "CAST_INVALID_INPUT", + "condition" : "CAST_INVALID_INPUT", "sqlState" : "22018", "messageParameters" : { "expression" : "'a'", @@ -224,7 +224,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.CAST_WITHOUT_SUGGESTION", + "condition" : "DATATYPE_MISMATCH.CAST_WITHOUT_SUGGESTION", "sqlState" : "42K09", "messageParameters" : { "sqlExpr" : "\"c1\"", @@ -349,7 +349,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "INCOMPATIBLE_VIEW_SCHEMA_CHANGE", + "condition" : "INCOMPATIBLE_VIEW_SCHEMA_CHANGE", "sqlState" : "51024", "messageParameters" : { "actualCols" : "[]", @@ -407,7 +407,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "INCOMPATIBLE_VIEW_SCHEMA_CHANGE", + "condition" : "INCOMPATIBLE_VIEW_SCHEMA_CHANGE", "sqlState" : "51024", "messageParameters" : { "actualCols" : "[]", @@ -527,7 +527,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "CANNOT_UP_CAST_DATATYPE", + "condition" : "CANNOT_UP_CAST_DATATYPE", "sqlState" : "42846", "messageParameters" : { "details" : "The type path of the target object is:\n\nYou can either add an explicit cast to the input data or choose a higher precision type of the field in the target object", diff --git a/sql/core/src/test/resources/sql-tests/results/view-schema-evolution.sql.out b/sql/core/src/test/resources/sql-tests/results/view-schema-evolution.sql.out index 7410e7eaafd6..0b49aafe0493 100644 --- a/sql/core/src/test/resources/sql-tests/results/view-schema-evolution.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/view-schema-evolution.sql.out @@ -541,7 +541,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "INCOMPATIBLE_VIEW_SCHEMA_CHANGE", + "condition" : "INCOMPATIBLE_VIEW_SCHEMA_CHANGE", "sqlState" : "51024", "messageParameters" : { "actualCols" : "[]", @@ -599,7 +599,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "INCOMPATIBLE_VIEW_SCHEMA_CHANGE", + "condition" : "INCOMPATIBLE_VIEW_SCHEMA_CHANGE", "sqlState" : "51024", "messageParameters" : { "actualCols" : "[]", @@ -986,7 +986,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "COLUMN_ALREADY_EXISTS", + "condition" : "COLUMN_ALREADY_EXISTS", "sqlState" : "42711", "messageParameters" : { "columnName" : "`c1`" diff --git a/sql/core/src/test/resources/sql-tests/results/view-schema-type-evolution.sql.out b/sql/core/src/test/resources/sql-tests/results/view-schema-type-evolution.sql.out index 6254d254b39e..de0655750d50 100644 --- a/sql/core/src/test/resources/sql-tests/results/view-schema-type-evolution.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/view-schema-type-evolution.sql.out @@ -258,7 +258,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "INCOMPATIBLE_VIEW_SCHEMA_CHANGE", + "condition" : "INCOMPATIBLE_VIEW_SCHEMA_CHANGE", "sqlState" : "51024", "messageParameters" : { "actualCols" : "[]", @@ -316,7 +316,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "INCOMPATIBLE_VIEW_SCHEMA_CHANGE", + "condition" : "INCOMPATIBLE_VIEW_SCHEMA_CHANGE", "sqlState" : "51024", "messageParameters" : { "actualCols" : "[]", diff --git a/sql/core/src/test/resources/sql-tests/results/window.sql.out b/sql/core/src/test/resources/sql-tests/results/window.sql.out index ce88fb57f8aa..08c7a4321440 100644 --- a/sql/core/src/test/resources/sql-tests/results/window.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/window.sql.out @@ -103,7 +103,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.SPECIFIED_WINDOW_FRAME_UNACCEPTED_TYPE", + "condition" : "DATATYPE_MISMATCH.SPECIFIED_WINDOW_FRAME_UNACCEPTED_TYPE", "sqlState" : "42K09", "messageParameters" : { "expectedType" : "\"INT\"", @@ -341,7 +341,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.RANGE_FRAME_INVALID_TYPE", + "condition" : "DATATYPE_MISMATCH.RANGE_FRAME_INVALID_TYPE", "sqlState" : "42K09", "messageParameters" : { "orderSpecType" : "\"DATE\"", @@ -383,7 +383,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.SPECIFIED_WINDOW_FRAME_INVALID_BOUND", + "condition" : "DATATYPE_MISMATCH.SPECIFIED_WINDOW_FRAME_INVALID_BOUND", "sqlState" : "42K09", "messageParameters" : { "lower" : "\"UNBOUNDED FOLLOWING\"", @@ -408,7 +408,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.RANGE_FRAME_WITHOUT_ORDER", + "condition" : "DATATYPE_MISMATCH.RANGE_FRAME_WITHOUT_ORDER", "sqlState" : "42K09", "messageParameters" : { "sqlExpr" : "\"(PARTITION BY cate RANGE BETWEEN CURRENT ROW AND 1 FOLLOWING)\"" @@ -431,7 +431,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.RANGE_FRAME_MULTI_ORDER", + "condition" : "DATATYPE_MISMATCH.RANGE_FRAME_MULTI_ORDER", "sqlState" : "42K09", "messageParameters" : { "orderSpec" : "val#x ASC NULLS FIRST,cate#x ASC NULLS FIRST", @@ -455,7 +455,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.RANGE_FRAME_INVALID_TYPE", + "condition" : "DATATYPE_MISMATCH.RANGE_FRAME_INVALID_TYPE", "sqlState" : "42K09", "messageParameters" : { "orderSpecType" : "\"TIMESTAMP\"", @@ -480,7 +480,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.SPECIFIED_WINDOW_FRAME_WRONG_COMPARISON", + "condition" : "DATATYPE_MISMATCH.SPECIFIED_WINDOW_FRAME_WRONG_COMPARISON", "sqlState" : "42K09", "messageParameters" : { "comparison" : "less than or equal", @@ -504,7 +504,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "_LEGACY_ERROR_TEMP_0064", + "condition" : "_LEGACY_ERROR_TEMP_0064", "messageParameters" : { "msg" : "Frame bound value must be a literal." }, @@ -561,7 +561,7 @@ struct<> -- !query output org.apache.spark.SparkArithmeticException { - "errorClass" : "DIVIDE_BY_ZERO", + "condition" : "DIVIDE_BY_ZERO", "sqlState" : "22012", "messageParameters" : { "config" : "\"spark.sql.ansi.enabled\"" @@ -599,7 +599,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "_LEGACY_ERROR_TEMP_1037", + "condition" : "_LEGACY_ERROR_TEMP_1037", "messageParameters" : { "wf" : "row_number()" } @@ -676,7 +676,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "_LEGACY_ERROR_TEMP_1030" + "condition" : "_LEGACY_ERROR_TEMP_1030" } @@ -1032,7 +1032,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.parser.ParseException { - "errorClass" : "INVALID_SQL_SYNTAX.REPETITIVE_WINDOW_DEFINITION", + "condition" : "INVALID_SQL_SYNTAX.REPETITIVE_WINDOW_DEFINITION", "sqlState" : "42000", "messageParameters" : { "windowName" : "`w`" @@ -1327,7 +1327,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "MISSING_WINDOW_SPECIFICATION", + "condition" : "MISSING_WINDOW_SPECIFICATION", "sqlState" : "42P20", "messageParameters" : { "docroot" : "https://spark.apache.org/docs/latest", @@ -1346,7 +1346,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "MISSING_WINDOW_SPECIFICATION", + "condition" : "MISSING_WINDOW_SPECIFICATION", "sqlState" : "42P20", "messageParameters" : { "docroot" : "https://spark.apache.org/docs/latest", @@ -1468,7 +1468,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.RANGE_FRAME_INVALID_TYPE", + "condition" : "DATATYPE_MISMATCH.RANGE_FRAME_INVALID_TYPE", "sqlState" : "42K09", "messageParameters" : { "orderSpecType" : "\"DATE\"", @@ -1492,7 +1492,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.SPECIFIED_WINDOW_FRAME_UNACCEPTED_TYPE", + "condition" : "DATATYPE_MISMATCH.SPECIFIED_WINDOW_FRAME_UNACCEPTED_TYPE", "sqlState" : "42K09", "messageParameters" : { "expectedType" : "(\"NUMERIC\" or \"INTERVAL DAY TO SECOND\" or \"INTERVAL YEAR TO MONTH\" or \"INTERVAL\")", diff --git a/sql/core/src/test/resources/sql-tests/results/xml-functions.sql.out b/sql/core/src/test/resources/sql-tests/results/xml-functions.sql.out index 60531b2bfd50..0a6a180a6a55 100644 --- a/sql/core/src/test/resources/sql-tests/results/xml-functions.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/xml-functions.sql.out @@ -22,7 +22,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"array(named_struct(a, 1, b, 2))\"", @@ -48,7 +48,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_INPUT_TYPE", "sqlState" : "42K09", "messageParameters" : { "inputSql" : "\"map(a, 1)\"", @@ -74,7 +74,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_OPTIONS.NON_MAP_FUNCTION", + "condition" : "INVALID_OPTIONS.NON_MAP_FUNCTION", "sqlState" : "42K06", "queryContext" : [ { "objectType" : "", @@ -93,7 +93,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_OPTIONS.NON_STRING_TYPE", + "condition" : "INVALID_OPTIONS.NON_STRING_TYPE", "sqlState" : "42K06", "messageParameters" : { "mapType" : "\"MAP\"" @@ -115,7 +115,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", + "condition" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", "sqlState" : "42605", "messageParameters" : { "actualNum" : "0", @@ -156,7 +156,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_SCHEMA.NON_STRING_LITERAL", + "condition" : "INVALID_SCHEMA.NON_STRING_LITERAL", "sqlState" : "42K07", "messageParameters" : { "inputSchema" : "\"1\"" @@ -178,7 +178,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "PARSE_SYNTAX_ERROR", + "condition" : "PARSE_SYNTAX_ERROR", "sqlState" : "42601", "messageParameters" : { "error" : "'InvalidType'", @@ -201,7 +201,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_OPTIONS.NON_MAP_FUNCTION", + "condition" : "INVALID_OPTIONS.NON_MAP_FUNCTION", "sqlState" : "42K06", "queryContext" : [ { "objectType" : "", @@ -220,7 +220,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "INVALID_OPTIONS.NON_STRING_TYPE", + "condition" : "INVALID_OPTIONS.NON_STRING_TYPE", "sqlState" : "42K06", "messageParameters" : { "mapType" : "\"MAP\"" @@ -242,7 +242,7 @@ struct<> -- !query output org.apache.spark.sql.AnalysisException { - "errorClass" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", + "condition" : "WRONG_NUM_ARGS.WITHOUT_SUGGESTION", "sqlState" : "42605", "messageParameters" : { "actualNum" : "0", @@ -380,7 +380,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.UNEXPECTED_NULL", + "condition" : "DATATYPE_MISMATCH.UNEXPECTED_NULL", "sqlState" : "42K09", "messageParameters" : { "exprName" : "xml", @@ -411,7 +411,7 @@ struct<> -- !query output org.apache.spark.sql.catalyst.ExtendedAnalysisException { - "errorClass" : "DATATYPE_MISMATCH.NON_FOLDABLE_INPUT", + "condition" : "DATATYPE_MISMATCH.NON_FOLDABLE_INPUT", "sqlState" : "42K09", "messageParameters" : { "inputExpr" : "\"xmlField\"", diff --git a/sql/hive-thriftserver/src/test/scala/org/apache/spark/sql/hive/thriftserver/CliSuite.scala b/sql/hive-thriftserver/src/test/scala/org/apache/spark/sql/hive/thriftserver/CliSuite.scala index a117e7af7f63..e6d04f18586f 100644 --- a/sql/hive-thriftserver/src/test/scala/org/apache/spark/sql/hive/thriftserver/CliSuite.scala +++ b/sql/hive-thriftserver/src/test/scala/org/apache/spark/sql/hive/thriftserver/CliSuite.scala @@ -735,7 +735,7 @@ class CliSuite extends SparkFunSuite { format = ErrorMessageFormat.MINIMAL, errorMessage = """{ - | "errorClass" : "DIVIDE_BY_ZERO", + | "condition" : "DIVIDE_BY_ZERO", | "sqlState" : "22012", | "messageParameters" : { | "config" : "\"spark.sql.ansi.enabled\"" @@ -753,7 +753,7 @@ class CliSuite extends SparkFunSuite { format = ErrorMessageFormat.STANDARD, errorMessage = """{ - | "errorClass" : "DIVIDE_BY_ZERO", + | "condition" : "DIVIDE_BY_ZERO", | "messageTemplate" : "Division by zero. Use `try_divide` to tolerate divisor being 0 and return NULL instead. If necessary set to \"false\" to bypass this error.", | "sqlState" : "22012", | "messageParameters" : { diff --git a/sql/hive-thriftserver/src/test/scala/org/apache/spark/sql/hive/thriftserver/ThriftServerWithSparkContextSuite.scala b/sql/hive-thriftserver/src/test/scala/org/apache/spark/sql/hive/thriftserver/ThriftServerWithSparkContextSuite.scala index 254eda69e86e..5152c2193499 100644 --- a/sql/hive-thriftserver/src/test/scala/org/apache/spark/sql/hive/thriftserver/ThriftServerWithSparkContextSuite.scala +++ b/sql/hive-thriftserver/src/test/scala/org/apache/spark/sql/hive/thriftserver/ThriftServerWithSparkContextSuite.scala @@ -173,7 +173,7 @@ trait ThriftServerWithSparkContextSuite extends SharedThriftServer { val e2 = intercept[HiveSQLException](exec(sql)) assert(e2.getMessage === """{ - | "errorClass" : "DIVIDE_BY_ZERO", + | "condition" : "DIVIDE_BY_ZERO", | "sqlState" : "22012", | "messageParameters" : { | "config" : "\"spark.sql.ansi.enabled\"" @@ -191,7 +191,7 @@ trait ThriftServerWithSparkContextSuite extends SharedThriftServer { val e3 = intercept[HiveSQLException](exec(sql)) assert(e3.getMessage === """{ - | "errorClass" : "DIVIDE_BY_ZERO", + | "condition" : "DIVIDE_BY_ZERO", | "messageTemplate" : "Division by zero. Use `try_divide` to tolerate divisor being 0 and return NULL instead. If necessary set to \"false\" to bypass this error.", | "sqlState" : "22012", | "messageParameters" : {