diff --git a/sql/core/src/test/resources/sql-tests/inputs/datetime-parsing-invalid.sql b/sql/core/src/test/resources/sql-tests/inputs/datetime-parsing-invalid.sql index a6d743cab5480..1d1e2a5282c81 100644 --- a/sql/core/src/test/resources/sql-tests/inputs/datetime-parsing-invalid.sql +++ b/sql/core/src/test/resources/sql-tests/inputs/datetime-parsing-invalid.sql @@ -14,7 +14,8 @@ select to_timestamp('366', 'D'); select to_timestamp('9', 'DD'); -- in java 8 this case is invalid, but valid in java 11, disabled for jenkins -- select to_timestamp('100', 'DD'); -select to_timestamp('366', 'DD'); +-- The error message is changed since Java 11+ +-- select to_timestamp('366', 'DD'); select to_timestamp('9', 'DDD'); select to_timestamp('99', 'DDD'); select to_timestamp('30-365', 'dd-DDD'); diff --git a/sql/core/src/test/resources/sql-tests/results/ansi/datetime-parsing-invalid.sql.out b/sql/core/src/test/resources/sql-tests/results/ansi/datetime-parsing-invalid.sql.out index 5dc3b85b3a9eb..59761d5ac53f0 100644 --- a/sql/core/src/test/resources/sql-tests/results/ansi/datetime-parsing-invalid.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/ansi/datetime-parsing-invalid.sql.out @@ -1,5 +1,5 @@ -- Automatically generated by SQLQueryTestSuite --- Number of queries: 29 +-- Number of queries: 28 -- !query @@ -74,15 +74,6 @@ org.apache.spark.SparkUpgradeException You may get a different result due to the upgrading to Spark >= 3.0: Fail to parse '9' in the new parser. You can set spark.sql.legacy.timeParserPolicy to LEGACY to restore the behavior before Spark 3.0, or set to CORRECTED and treat it as an invalid datetime string. --- !query -select to_timestamp('366', 'DD') --- !query schema -struct<> --- !query output -java.time.format.DateTimeParseException -Text '366' could not be parsed, unparsed text found at index 2. If necessary set spark.sql.ansi.enabled to false to bypass this error. - - -- !query select to_timestamp('9', 'DDD') -- !query schema diff --git a/sql/core/src/test/resources/sql-tests/results/datetime-parsing-invalid.sql.out b/sql/core/src/test/resources/sql-tests/results/datetime-parsing-invalid.sql.out index 33504709c08ec..9fc28876a5b2a 100644 --- a/sql/core/src/test/resources/sql-tests/results/datetime-parsing-invalid.sql.out +++ b/sql/core/src/test/resources/sql-tests/results/datetime-parsing-invalid.sql.out @@ -1,5 +1,5 @@ -- Automatically generated by SQLQueryTestSuite --- Number of queries: 29 +-- Number of queries: 28 -- !query @@ -72,14 +72,6 @@ org.apache.spark.SparkUpgradeException You may get a different result due to the upgrading to Spark >= 3.0: Fail to parse '9' in the new parser. You can set spark.sql.legacy.timeParserPolicy to LEGACY to restore the behavior before Spark 3.0, or set to CORRECTED and treat it as an invalid datetime string. --- !query -select to_timestamp('366', 'DD') --- !query schema -struct --- !query output -NULL - - -- !query select to_timestamp('9', 'DDD') -- !query schema