You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@spark.apache.org by "Apache Spark (Jira)" <ji...@apache.org> on 2023/02/15 03:55:00 UTC

[jira] [Assigned] (SPARK-38324) The second range is not [0, 59] in the day time ANSI interval

     [ https://issues.apache.org/jira/browse/SPARK-38324?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]

Apache Spark reassigned SPARK-38324:
------------------------------------

    Assignee:     (was: Apache Spark)

> The second range is not [0, 59] in the day time ANSI interval
> -------------------------------------------------------------
>
>                 Key: SPARK-38324
>                 URL: https://issues.apache.org/jira/browse/SPARK-38324
>             Project: Spark
>          Issue Type: Bug
>          Components: Java API
>    Affects Versions: 3.3.0
>         Environment: Spark 3.3.0 snapshot
>            Reporter: chong
>            Priority: Major
>
> [https://spark.apache.org/docs/latest/sql-ref-datatypes.html]
>  * SECOND, seconds within minutes and possibly fractions of a second [0..59.999999]{{{{}}{}}}
> {{Doc shows SECOND is seconds within minutes, it's range should be [0, 59]}}
>  
> But testing shows 99 second is valid:
> {{>>> spark.sql("select INTERVAL '10 01:01:99' DAY TO SECOND")}}
> {{{}DataFrame[INTERVAL '10 01:02:39' DAY TO SECOND: interval day to second]{}}}{{{{}}{}}}
>  
> Meanwhile, minute range check is ok, see below:
> >>> spark.sql("select INTERVAL '10 01:60:01' DAY TO SECOND")
> requirement failed: {color:#de350b}*minute 60 outside range [0, 59]*{color}(line 1, pos 16)
> == SQL ==
> select INTERVAL '10 01:60:01' DAY TO SECOND
> ----------------^^^
>  



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscribe@spark.apache.org
For additional commands, e-mail: issues-help@spark.apache.org