You are viewing a plain text version of this content. The canonical link for it is here.
Posted to reviews@spark.apache.org by cloud-fan <gi...@git.apache.org> on 2018/09/25 07:37:42 UTC

[GitHub] spark pull request #22543: [SPARK-23715][SQL] improve document for from/to_u...

GitHub user cloud-fan opened a pull request:

    https://github.com/apache/spark/pull/22543

    [SPARK-23715][SQL] improve document for from/to_utc_timestamp

    ## What changes were proposed in this pull request?
    
    We have an agreement that the behavior of `from/to_utc_timestamp` is corrected, although the function itself doesn't make much sense in Spark: https://issues.apache.org/jira/browse/SPARK-23715
    
    This PR improves the document.
    
    ## How was this patch tested?
    
    N/A

You can merge this pull request into a Git repository by running:

    $ git pull https://github.com/cloud-fan/spark doc

Alternatively you can review and apply these changes as the patch at:

    https://github.com/apache/spark/pull/22543.patch

To close this pull request, make a commit to your master/trunk branch
with (at least) the following in the commit message:

    This closes #22543
    
----
commit b4d31f862ce7c09ebd9a431997dbc9edf4a2f6f3
Author: Wenchen Fan <we...@...>
Date:   2018-09-25T07:31:31Z

    improve doc

----


---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark issue #22543: [SPARK-23715][SQL][DOC] improve document for from/to_utc...

Posted by cloud-fan <gi...@git.apache.org>.
Github user cloud-fan commented on the issue:

    https://github.com/apache/spark/pull/22543
  
    thanks, merging to master/2.4!


---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark issue #22543: [SPARK-23715][SQL][DOC] improve document for from/to_utc...

Posted by AmplabJenkins <gi...@git.apache.org>.
Github user AmplabJenkins commented on the issue:

    https://github.com/apache/spark/pull/22543
  
    Test PASSed.
    Refer to this link for build results (access rights to CI server needed): 
    https://amplab.cs.berkeley.edu/jenkins//job/testing-k8s-prb-make-spark-distribution-unified/3464/
    Test PASSed.


---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark pull request #22543: [SPARK-23715][SQL] improve document for from/to_u...

Posted by dongjoon-hyun <gi...@git.apache.org>.
Github user dongjoon-hyun commented on a diff in the pull request:

    https://github.com/apache/spark/pull/22543#discussion_r220409796
  
    --- Diff: sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/datetimeExpressions.scala ---
    @@ -1018,9 +1018,20 @@ case class TimeAdd(start: Expression, interval: Expression, timeZoneId: Option[S
     }
     
     /**
    - * Given a timestamp like '2017-07-14 02:40:00.0', interprets it as a time in UTC, and renders
    - * that time as a timestamp in the given time zone. For example, 'GMT+1' would yield
    - * '2017-07-14 03:40:00.0'.
    + * This is a common function for databases supporting TIMESTAMP WITHOUT TIMEZONE. This function
    + * takes a timestamp which is timezone-agnostic, and interprets it as a timestamp in UTC, and
    + * renders that timestamp as a timestamp in the given time zone.
    + *
    + * However, timestamp in Spark represents number of microseconds from the Unix epoch, which is not
    + * timezone-agnostic. So in Spark this function just shift the timestamp value from UTC timezone to
    + * the given timezone.
    + *
    + * This function may return confusing result if the input is a string with timezone, e.g.
    + * '2018-03-13T06:18:23+00:00'. The reason is that, Spark firstly cast the string to timestamp
    + * according to the timezone in the string, and finally display the result by converting the
    + * timestamp to string according to the session local timezone.
    + *
    + * We may remove this function in Spark 3.0.
    --- End diff --
    
    nit, could you add `[DOC]` in the title?


---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark issue #22543: [SPARK-23715][SQL][DOC] improve document for from/to_utc...

Posted by SparkQA <gi...@git.apache.org>.
Github user SparkQA commented on the issue:

    https://github.com/apache/spark/pull/22543
  
    **[Test build #96598 has finished](https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/96598/testReport)** for PR 22543 at commit [`e033694`](https://github.com/apache/spark/commit/e033694ab62a79185cc60e0039657cd2eaea5f4d).
     * This patch **fails Spark unit tests**.
     * This patch merges cleanly.
     * This patch adds no public classes.


---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark issue #22543: [SPARK-23715][SQL][DOC] improve document for from/to_utc...

Posted by HyukjinKwon <gi...@git.apache.org>.
Github user HyukjinKwon commented on the issue:

    https://github.com/apache/spark/pull/22543
  
    retest this please


---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark pull request #22543: [SPARK-23715][SQL] improve document for from/to_u...

Posted by cloud-fan <gi...@git.apache.org>.
Github user cloud-fan commented on a diff in the pull request:

    https://github.com/apache/spark/pull/22543#discussion_r220114639
  
    --- Diff: sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/datetimeExpressions.scala ---
    @@ -1018,9 +1018,20 @@ case class TimeAdd(start: Expression, interval: Expression, timeZoneId: Option[S
     }
     
     /**
    - * Given a timestamp like '2017-07-14 02:40:00.0', interprets it as a time in UTC, and renders
    - * that time as a timestamp in the given time zone. For example, 'GMT+1' would yield
    - * '2017-07-14 03:40:00.0'.
    + * This is a common function for databases supporting TIMESTAMP WITHOUT TIMEZONE. This function
    + * takes a timestamp which is timezone-agnostic, and interprets it as a timestamp in UTC, and
    + * renders that timestamp as a timestamp in the given time zone.
    + *
    + * However, timestamp in Spark represents number of microseconds from the Unix epoch, which is not
    + * timezone-agnostic. So in Spark this function just shift the timestamp value from UTC timezone to
    + * the given timezone.
    + *
    + * This function may return confusing result if the input is a string with timezone, e.g.
    + * '2018-03-13T06:18:23+00:00'. The reason is that, Spark firstly cast the string to timestamp
    + * according to the timezone in the string, and finally display the result by converting the
    + * timestamp to string according to the session local timezone.
    + *
    + * We may remove this function in Spark 3.0.
    --- End diff --
    
    It's too late for 2.4, I'm thinking about deprecating them in 2.5. @rxin what do you think?


---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark pull request #22543: [SPARK-23715][SQL][DOC] improve document for from...

Posted by cloud-fan <gi...@git.apache.org>.
Github user cloud-fan commented on a diff in the pull request:

    https://github.com/apache/spark/pull/22543#discussion_r220411054
  
    --- Diff: sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/datetimeExpressions.scala ---
    @@ -1018,9 +1018,20 @@ case class TimeAdd(start: Expression, interval: Expression, timeZoneId: Option[S
     }
     
     /**
    - * Given a timestamp like '2017-07-14 02:40:00.0', interprets it as a time in UTC, and renders
    - * that time as a timestamp in the given time zone. For example, 'GMT+1' would yield
    - * '2017-07-14 03:40:00.0'.
    + * This is a common function for databases supporting TIMESTAMP WITHOUT TIMEZONE. This function
    + * takes a timestamp which is timezone-agnostic, and interprets it as a timestamp in UTC, and
    + * renders that timestamp as a timestamp in the given time zone.
    + *
    + * However, timestamp in Spark represents number of microseconds from the Unix epoch, which is not
    + * timezone-agnostic. So in Spark this function just shift the timestamp value from UTC timezone to
    + * the given timezone.
    + *
    + * This function may return confusing result if the input is a string with timezone, e.g.
    + * '2018-03-13T06:18:23+00:00'. The reason is that, Spark firstly cast the string to timestamp
    + * according to the timezone in the string, and finally display the result by converting the
    + * timestamp to string according to the session local timezone.
    + *
    + * We may remove this function in Spark 3.0.
    --- End diff --
    
    This may be too long to put into the SQL doc, I'll update doc in other languages.


---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark issue #22543: [SPARK-23715][SQL][DOC] improve document for from/to_utc...

Posted by SparkQA <gi...@git.apache.org>.
Github user SparkQA commented on the issue:

    https://github.com/apache/spark/pull/22543
  
    **[Test build #96599 has started](https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/96599/testReport)** for PR 22543 at commit [`e033694`](https://github.com/apache/spark/commit/e033694ab62a79185cc60e0039657cd2eaea5f4d).


---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark pull request #22543: [SPARK-23715][SQL][DOC] improve document for from...

Posted by asfgit <gi...@git.apache.org>.
Github user asfgit closed the pull request at:

    https://github.com/apache/spark/pull/22543


---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark pull request #22543: [SPARK-23715][SQL][DOC] improve document for from...

Posted by dongjoon-hyun <gi...@git.apache.org>.
Github user dongjoon-hyun commented on a diff in the pull request:

    https://github.com/apache/spark/pull/22543#discussion_r220411333
  
    --- Diff: sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/datetimeExpressions.scala ---
    @@ -1018,9 +1018,20 @@ case class TimeAdd(start: Expression, interval: Expression, timeZoneId: Option[S
     }
     
     /**
    - * Given a timestamp like '2017-07-14 02:40:00.0', interprets it as a time in UTC, and renders
    - * that time as a timestamp in the given time zone. For example, 'GMT+1' would yield
    - * '2017-07-14 03:40:00.0'.
    + * This is a common function for databases supporting TIMESTAMP WITHOUT TIMEZONE. This function
    + * takes a timestamp which is timezone-agnostic, and interprets it as a timestamp in UTC, and
    + * renders that timestamp as a timestamp in the given time zone.
    + *
    + * However, timestamp in Spark represents number of microseconds from the Unix epoch, which is not
    + * timezone-agnostic. So in Spark this function just shift the timestamp value from UTC timezone to
    + * the given timezone.
    + *
    + * This function may return confusing result if the input is a string with timezone, e.g.
    + * '2018-03-13T06:18:23+00:00'. The reason is that, Spark firstly cast the string to timestamp
    + * according to the timezone in the string, and finally display the result by converting the
    + * timestamp to string according to the session local timezone.
    + *
    + * We may remove this function in Spark 3.0.
    --- End diff --
    
    If we need to do `deprecation` for all, let's update the PR title clearly; maybe additional `[SPARK-25496]` and `Deprecation` wording at the title.


---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark issue #22543: [SPARK-23715][SQL][DOC] improve document for from/to_utc...

Posted by SparkQA <gi...@git.apache.org>.
Github user SparkQA commented on the issue:

    https://github.com/apache/spark/pull/22543
  
    **[Test build #96598 has started](https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/96598/testReport)** for PR 22543 at commit [`e033694`](https://github.com/apache/spark/commit/e033694ab62a79185cc60e0039657cd2eaea5f4d).


---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark issue #22543: [SPARK-23715][SQL][DOC] improve document for from/to_utc...

Posted by AmplabJenkins <gi...@git.apache.org>.
Github user AmplabJenkins commented on the issue:

    https://github.com/apache/spark/pull/22543
  
    Test PASSed.
    Refer to this link for build results (access rights to CI server needed): 
    https://amplab.cs.berkeley.edu/jenkins//job/testing-k8s-prb-make-spark-distribution-unified/3463/
    Test PASSed.


---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark pull request #22543: [SPARK-23715][SQL][DOC] improve document for from...

Posted by dongjoon-hyun <gi...@git.apache.org>.
Github user dongjoon-hyun commented on a diff in the pull request:

    https://github.com/apache/spark/pull/22543#discussion_r220412259
  
    --- Diff: sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/datetimeExpressions.scala ---
    @@ -1018,9 +1018,20 @@ case class TimeAdd(start: Expression, interval: Expression, timeZoneId: Option[S
     }
     
     /**
    - * Given a timestamp like '2017-07-14 02:40:00.0', interprets it as a time in UTC, and renders
    - * that time as a timestamp in the given time zone. For example, 'GMT+1' would yield
    - * '2017-07-14 03:40:00.0'.
    + * This is a common function for databases supporting TIMESTAMP WITHOUT TIMEZONE. This function
    + * takes a timestamp which is timezone-agnostic, and interprets it as a timestamp in UTC, and
    + * renders that timestamp as a timestamp in the given time zone.
    + *
    + * However, timestamp in Spark represents number of microseconds from the Unix epoch, which is not
    + * timezone-agnostic. So in Spark this function just shift the timestamp value from UTC timezone to
    + * the given timezone.
    + *
    + * This function may return confusing result if the input is a string with timezone, e.g.
    + * '2018-03-13T06:18:23+00:00'. The reason is that, Spark firstly cast the string to timestamp
    + * according to the timezone in the string, and finally display the result by converting the
    + * timestamp to string according to the session local timezone.
    + *
    + * We may remove this function in Spark 3.0.
    --- End diff --
    
    What I meant is `Deprecation` is a more formal activity requiring complete coverage. And, it's open as SPARK-25496.


---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark issue #22543: [SPARK-23715][SQL][DOC] improve document for from/to_utc...

Posted by SparkQA <gi...@git.apache.org>.
Github user SparkQA commented on the issue:

    https://github.com/apache/spark/pull/22543
  
    **[Test build #96597 has started](https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/96597/testReport)** for PR 22543 at commit [`1220888`](https://github.com/apache/spark/commit/1220888cd74e2db6a88fbe309192f7ad501e2952).


---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark issue #22543: [SPARK-23715][SQL][DOC] improve document for from/to_utc...

Posted by AmplabJenkins <gi...@git.apache.org>.
Github user AmplabJenkins commented on the issue:

    https://github.com/apache/spark/pull/22543
  
    Merged build finished. Test PASSed.


---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark pull request #22543: [SPARK-23715][SQL][DOC] improve document for from...

Posted by HyukjinKwon <gi...@git.apache.org>.
Github user HyukjinKwon commented on a diff in the pull request:

    https://github.com/apache/spark/pull/22543#discussion_r220410681
  
    --- Diff: sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/datetimeExpressions.scala ---
    @@ -1018,9 +1018,20 @@ case class TimeAdd(start: Expression, interval: Expression, timeZoneId: Option[S
     }
     
     /**
    - * Given a timestamp like '2017-07-14 02:40:00.0', interprets it as a time in UTC, and renders
    - * that time as a timestamp in the given time zone. For example, 'GMT+1' would yield
    - * '2017-07-14 03:40:00.0'.
    + * This is a common function for databases supporting TIMESTAMP WITHOUT TIMEZONE. This function
    + * takes a timestamp which is timezone-agnostic, and interprets it as a timestamp in UTC, and
    + * renders that timestamp as a timestamp in the given time zone.
    + *
    + * However, timestamp in Spark represents number of microseconds from the Unix epoch, which is not
    + * timezone-agnostic. So in Spark this function just shift the timestamp value from UTC timezone to
    + * the given timezone.
    + *
    + * This function may return confusing result if the input is a string with timezone, e.g.
    + * '2018-03-13T06:18:23+00:00'. The reason is that, Spark firstly cast the string to timestamp
    + * according to the timezone in the string, and finally display the result by converting the
    + * timestamp to string according to the session local timezone.
    + *
    + * We may remove this function in Spark 3.0.
    --- End diff --
    
    should also update other language APIs as well.


---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark issue #22543: [SPARK-23715][SQL][DOC] improve document for from/to_utc...

Posted by AmplabJenkins <gi...@git.apache.org>.
Github user AmplabJenkins commented on the issue:

    https://github.com/apache/spark/pull/22543
  
    Merged build finished. Test PASSed.


---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark issue #22543: [SPARK-23715][SQL][DOC] improve document for from/to_utc...

Posted by AmplabJenkins <gi...@git.apache.org>.
Github user AmplabJenkins commented on the issue:

    https://github.com/apache/spark/pull/22543
  
    Merged build finished. Test PASSed.


---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark issue #22543: [SPARK-23715][SQL] improve document for from/to_utc_time...

Posted by AmplabJenkins <gi...@git.apache.org>.
Github user AmplabJenkins commented on the issue:

    https://github.com/apache/spark/pull/22543
  
    Merged build finished. Test PASSed.


---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark issue #22543: [SPARK-23715][SQL] improve document for from/to_utc_time...

Posted by SparkQA <gi...@git.apache.org>.
Github user SparkQA commented on the issue:

    https://github.com/apache/spark/pull/22543
  
    **[Test build #96541 has started](https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/96541/testReport)** for PR 22543 at commit [`b4d31f8`](https://github.com/apache/spark/commit/b4d31f862ce7c09ebd9a431997dbc9edf4a2f6f3).


---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark pull request #22543: [SPARK-23715][SQL] improve document for from/to_u...

Posted by HyukjinKwon <gi...@git.apache.org>.
Github user HyukjinKwon commented on a diff in the pull request:

    https://github.com/apache/spark/pull/22543#discussion_r220112968
  
    --- Diff: sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/datetimeExpressions.scala ---
    @@ -1018,9 +1018,20 @@ case class TimeAdd(start: Expression, interval: Expression, timeZoneId: Option[S
     }
     
     /**
    - * Given a timestamp like '2017-07-14 02:40:00.0', interprets it as a time in UTC, and renders
    - * that time as a timestamp in the given time zone. For example, 'GMT+1' would yield
    - * '2017-07-14 03:40:00.0'.
    + * This is a common function for databases supporting TIMESTAMP WITHOUT TIMEZONE. This function
    + * takes a timestamp which is timezone-agnostic, and interprets it as a timestamp in UTC, and
    + * renders that timestamp as a timestamp in the given time zone.
    + *
    + * However, timestamp in Spark represents number of microseconds from the Unix epoch, which is not
    + * timezone-agnostic. So in Spark this function just shift the timestamp value from UTC timezone to
    + * the given timezone.
    + *
    + * This function may return confusing result if the input is a string with timezone, e.g.
    + * '2018-03-13T06:18:23+00:00'. The reason is that, Spark firstly cast the string to timestamp
    + * according to the timezone in the string, and finally display the result by converting the
    + * timestamp to string according to the session local timezone.
    + *
    + * We may remove this function in Spark 3.0.
    --- End diff --
    
    Hm, how about explicitly deprecating them in Scala/Python/R APIs?


---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark issue #22543: [SPARK-23715][SQL] improve document for from/to_utc_time...

Posted by cloud-fan <gi...@git.apache.org>.
Github user cloud-fan commented on the issue:

    https://github.com/apache/spark/pull/22543
  
    cc @rxin @gatorsmile 


---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark issue #22543: [SPARK-23715][SQL][DOC] improve document for from/to_utc...

Posted by AmplabJenkins <gi...@git.apache.org>.
Github user AmplabJenkins commented on the issue:

    https://github.com/apache/spark/pull/22543
  
    Merged build finished. Test FAILed.


---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark issue #22543: [SPARK-23715][SQL][DOC] improve document for from/to_utc...

Posted by cloud-fan <gi...@git.apache.org>.
Github user cloud-fan commented on the issue:

    https://github.com/apache/spark/pull/22543
  
    retest this please


---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark issue #22543: [SPARK-23715][SQL][DOC] improve document for from/to_utc...

Posted by SparkQA <gi...@git.apache.org>.
Github user SparkQA commented on the issue:

    https://github.com/apache/spark/pull/22543
  
    **[Test build #96597 has finished](https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/96597/testReport)** for PR 22543 at commit [`1220888`](https://github.com/apache/spark/commit/1220888cd74e2db6a88fbe309192f7ad501e2952).
     * This patch **fails due to an unknown error code, -9**.
     * This patch merges cleanly.
     * This patch adds no public classes.


---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark issue #22543: [SPARK-23715][SQL][DOC] improve document for from/to_utc...

Posted by AmplabJenkins <gi...@git.apache.org>.
Github user AmplabJenkins commented on the issue:

    https://github.com/apache/spark/pull/22543
  
    Merged build finished. Test PASSed.


---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark issue #22543: [SPARK-23715][SQL] improve document for from/to_utc_time...

Posted by AmplabJenkins <gi...@git.apache.org>.
Github user AmplabJenkins commented on the issue:

    https://github.com/apache/spark/pull/22543
  
    Test PASSed.
    Refer to this link for build results (access rights to CI server needed): 
    https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/96541/
    Test PASSed.


---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark issue #22543: [SPARK-23715][SQL][DOC] improve document for from/to_utc...

Posted by AmplabJenkins <gi...@git.apache.org>.
Github user AmplabJenkins commented on the issue:

    https://github.com/apache/spark/pull/22543
  
    Test PASSed.
    Refer to this link for build results (access rights to CI server needed): 
    https://amplab.cs.berkeley.edu/jenkins//job/testing-k8s-prb-make-spark-distribution-unified/3462/
    Test PASSed.


---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark issue #22543: [SPARK-23715][SQL][DOC] improve document for from/to_utc...

Posted by AmplabJenkins <gi...@git.apache.org>.
Github user AmplabJenkins commented on the issue:

    https://github.com/apache/spark/pull/22543
  
    Merged build finished. Test FAILed.


---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark issue #22543: [SPARK-23715][SQL][DOC] improve document for from/to_utc...

Posted by AmplabJenkins <gi...@git.apache.org>.
Github user AmplabJenkins commented on the issue:

    https://github.com/apache/spark/pull/22543
  
    Merged build finished. Test FAILed.


---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark issue #22543: [SPARK-23715][SQL] improve document for from/to_utc_time...

Posted by AmplabJenkins <gi...@git.apache.org>.
Github user AmplabJenkins commented on the issue:

    https://github.com/apache/spark/pull/22543
  
    Merged build finished. Test PASSed.


---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark issue #22543: [SPARK-23715][SQL][DOC] improve document for from/to_utc...

Posted by SparkQA <gi...@git.apache.org>.
Github user SparkQA commented on the issue:

    https://github.com/apache/spark/pull/22543
  
    **[Test build #96653 has started](https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/96653/testReport)** for PR 22543 at commit [`e033694`](https://github.com/apache/spark/commit/e033694ab62a79185cc60e0039657cd2eaea5f4d).


---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark issue #22543: [SPARK-23715][SQL] improve document for from/to_utc_time...

Posted by AmplabJenkins <gi...@git.apache.org>.
Github user AmplabJenkins commented on the issue:

    https://github.com/apache/spark/pull/22543
  
    Test PASSed.
    Refer to this link for build results (access rights to CI server needed): 
    https://amplab.cs.berkeley.edu/jenkins//job/testing-k8s-prb-make-spark-distribution-unified/3427/
    Test PASSed.


---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark issue #22543: [SPARK-23715][SQL][DOC] improve document for from/to_utc...

Posted by AmplabJenkins <gi...@git.apache.org>.
Github user AmplabJenkins commented on the issue:

    https://github.com/apache/spark/pull/22543
  
    Merged build finished. Test PASSed.


---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark issue #22543: [SPARK-23715][SQL][DOC] improve document for from/to_utc...

Posted by AmplabJenkins <gi...@git.apache.org>.
Github user AmplabJenkins commented on the issue:

    https://github.com/apache/spark/pull/22543
  
    Test FAILed.
    Refer to this link for build results (access rights to CI server needed): 
    https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/96598/
    Test FAILed.


---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark pull request #22543: [SPARK-23715][SQL][DOC] improve document for from...

Posted by rxin <gi...@git.apache.org>.
Github user rxin commented on a diff in the pull request:

    https://github.com/apache/spark/pull/22543#discussion_r220410457
  
    --- Diff: sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/datetimeExpressions.scala ---
    @@ -1018,9 +1018,20 @@ case class TimeAdd(start: Expression, interval: Expression, timeZoneId: Option[S
     }
     
     /**
    - * Given a timestamp like '2017-07-14 02:40:00.0', interprets it as a time in UTC, and renders
    - * that time as a timestamp in the given time zone. For example, 'GMT+1' would yield
    - * '2017-07-14 03:40:00.0'.
    + * This is a common function for databases supporting TIMESTAMP WITHOUT TIMEZONE. This function
    + * takes a timestamp which is timezone-agnostic, and interprets it as a timestamp in UTC, and
    + * renders that timestamp as a timestamp in the given time zone.
    + *
    + * However, timestamp in Spark represents number of microseconds from the Unix epoch, which is not
    + * timezone-agnostic. So in Spark this function just shift the timestamp value from UTC timezone to
    + * the given timezone.
    + *
    + * This function may return confusing result if the input is a string with timezone, e.g.
    + * '2018-03-13T06:18:23+00:00'. The reason is that, Spark firstly cast the string to timestamp
    + * according to the timezone in the string, and finally display the result by converting the
    + * timestamp to string according to the session local timezone.
    + *
    + * We may remove this function in Spark 3.0.
    --- End diff --
    
    should also update the sql doc?



---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark issue #22543: [SPARK-23715][SQL][DOC] improve document for from/to_utc...

Posted by HyukjinKwon <gi...@git.apache.org>.
Github user HyukjinKwon commented on the issue:

    https://github.com/apache/spark/pull/22543
  
    retest this please
    
    



---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark issue #22543: [SPARK-23715][SQL][DOC] improve document for from/to_utc...

Posted by AmplabJenkins <gi...@git.apache.org>.
Github user AmplabJenkins commented on the issue:

    https://github.com/apache/spark/pull/22543
  
    Test FAILed.
    Refer to this link for build results (access rights to CI server needed): 
    https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/96599/
    Test FAILed.


---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark issue #22543: [SPARK-23715][SQL][DOC] improve document for from/to_utc...

Posted by SparkQA <gi...@git.apache.org>.
Github user SparkQA commented on the issue:

    https://github.com/apache/spark/pull/22543
  
    **[Test build #96653 has finished](https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/96653/testReport)** for PR 22543 at commit [`e033694`](https://github.com/apache/spark/commit/e033694ab62a79185cc60e0039657cd2eaea5f4d).
     * This patch passes all tests.
     * This patch merges cleanly.
     * This patch adds no public classes.


---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark issue #22543: [SPARK-23715][SQL][DOC] improve document for from/to_utc...

Posted by AmplabJenkins <gi...@git.apache.org>.
Github user AmplabJenkins commented on the issue:

    https://github.com/apache/spark/pull/22543
  
    Test PASSed.
    Refer to this link for build results (access rights to CI server needed): 
    https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/96653/
    Test PASSed.


---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark issue #22543: [SPARK-23715][SQL][DOC] improve document for from/to_utc...

Posted by AmplabJenkins <gi...@git.apache.org>.
Github user AmplabJenkins commented on the issue:

    https://github.com/apache/spark/pull/22543
  
    Test PASSed.
    Refer to this link for build results (access rights to CI server needed): 
    https://amplab.cs.berkeley.edu/jenkins//job/testing-k8s-prb-make-spark-distribution-unified/3506/
    Test PASSed.


---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark pull request #22543: [SPARK-23715][SQL][DOC] improve document for from...

Posted by HyukjinKwon <gi...@git.apache.org>.
Github user HyukjinKwon commented on a diff in the pull request:

    https://github.com/apache/spark/pull/22543#discussion_r220411454
  
    --- Diff: sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/datetimeExpressions.scala ---
    @@ -1018,9 +1018,20 @@ case class TimeAdd(start: Expression, interval: Expression, timeZoneId: Option[S
     }
     
     /**
    - * Given a timestamp like '2017-07-14 02:40:00.0', interprets it as a time in UTC, and renders
    - * that time as a timestamp in the given time zone. For example, 'GMT+1' would yield
    - * '2017-07-14 03:40:00.0'.
    + * This is a common function for databases supporting TIMESTAMP WITHOUT TIMEZONE. This function
    + * takes a timestamp which is timezone-agnostic, and interprets it as a timestamp in UTC, and
    + * renders that timestamp as a timestamp in the given time zone.
    + *
    + * However, timestamp in Spark represents number of microseconds from the Unix epoch, which is not
    + * timezone-agnostic. So in Spark this function just shift the timestamp value from UTC timezone to
    + * the given timezone.
    + *
    + * This function may return confusing result if the input is a string with timezone, e.g.
    + * '2018-03-13T06:18:23+00:00'. The reason is that, Spark firstly cast the string to timestamp
    + * according to the timezone in the string, and finally display the result by converting the
    + * timestamp to string according to the session local timezone.
    + *
    + * We may remove this function in Spark 3.0.
    --- End diff --
    
    Deprecation is kind of a warning before removal .. warning for deprecation sounds like a warning for a warning .. 


---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark pull request #22543: [SPARK-23715][SQL][DOC] improve document for from...

Posted by HyukjinKwon <gi...@git.apache.org>.
Github user HyukjinKwon commented on a diff in the pull request:

    https://github.com/apache/spark/pull/22543#discussion_r220410505
  
    --- Diff: sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/datetimeExpressions.scala ---
    @@ -1018,9 +1018,20 @@ case class TimeAdd(start: Expression, interval: Expression, timeZoneId: Option[S
     }
     
     /**
    - * Given a timestamp like '2017-07-14 02:40:00.0', interprets it as a time in UTC, and renders
    - * that time as a timestamp in the given time zone. For example, 'GMT+1' would yield
    - * '2017-07-14 03:40:00.0'.
    + * This is a common function for databases supporting TIMESTAMP WITHOUT TIMEZONE. This function
    + * takes a timestamp which is timezone-agnostic, and interprets it as a timestamp in UTC, and
    + * renders that timestamp as a timestamp in the given time zone.
    + *
    + * However, timestamp in Spark represents number of microseconds from the Unix epoch, which is not
    + * timezone-agnostic. So in Spark this function just shift the timestamp value from UTC timezone to
    + * the given timezone.
    + *
    + * This function may return confusing result if the input is a string with timezone, e.g.
    + * '2018-03-13T06:18:23+00:00'. The reason is that, Spark firstly cast the string to timestamp
    + * according to the timezone in the string, and finally display the result by converting the
    + * timestamp to string according to the session local timezone.
    + *
    + * We may remove this function in Spark 3.0.
    --- End diff --
    
    Hmmm .. I think we should get rid of this sentence here and just formally deprecate it in 2.5.0 ..  don't feel strongly.


---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark pull request #22543: [SPARK-23715][SQL][DOC] improve document for from...

Posted by dongjoon-hyun <gi...@git.apache.org>.
Github user dongjoon-hyun commented on a diff in the pull request:

    https://github.com/apache/spark/pull/22543#discussion_r220411158
  
    --- Diff: sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/datetimeExpressions.scala ---
    @@ -1018,9 +1018,20 @@ case class TimeAdd(start: Expression, interval: Expression, timeZoneId: Option[S
     }
     
     /**
    - * Given a timestamp like '2017-07-14 02:40:00.0', interprets it as a time in UTC, and renders
    - * that time as a timestamp in the given time zone. For example, 'GMT+1' would yield
    - * '2017-07-14 03:40:00.0'.
    + * This is a common function for databases supporting TIMESTAMP WITHOUT TIMEZONE. This function
    + * takes a timestamp which is timezone-agnostic, and interprets it as a timestamp in UTC, and
    + * renders that timestamp as a timestamp in the given time zone.
    + *
    + * However, timestamp in Spark represents number of microseconds from the Unix epoch, which is not
    + * timezone-agnostic. So in Spark this function just shift the timestamp value from UTC timezone to
    + * the given timezone.
    + *
    + * This function may return confusing result if the input is a string with timezone, e.g.
    + * '2018-03-13T06:18:23+00:00'. The reason is that, Spark firstly cast the string to timestamp
    + * according to the timezone in the string, and finally display the result by converting the
    + * timestamp to string according to the session local timezone.
    + *
    + * We may remove this function in Spark 3.0.
    --- End diff --
    
    I thought this is a simple documentation change for SQL function doc for 2.4.0 and https://issues.apache.org/jira/browse/SPARK-25496 is aiming the official global deprecation at 2.5.0.


---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark pull request #22543: [SPARK-23715][SQL][DOC] improve document for from...

Posted by cloud-fan <gi...@git.apache.org>.
Github user cloud-fan commented on a diff in the pull request:

    https://github.com/apache/spark/pull/22543#discussion_r220420622
  
    --- Diff: sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/datetimeExpressions.scala ---
    @@ -1018,9 +1018,20 @@ case class TimeAdd(start: Expression, interval: Expression, timeZoneId: Option[S
     }
     
     /**
    - * Given a timestamp like '2017-07-14 02:40:00.0', interprets it as a time in UTC, and renders
    - * that time as a timestamp in the given time zone. For example, 'GMT+1' would yield
    - * '2017-07-14 03:40:00.0'.
    + * This is a common function for databases supporting TIMESTAMP WITHOUT TIMEZONE. This function
    + * takes a timestamp which is timezone-agnostic, and interprets it as a timestamp in UTC, and
    + * renders that timestamp as a timestamp in the given time zone.
    + *
    + * However, timestamp in Spark represents number of microseconds from the Unix epoch, which is not
    + * timezone-agnostic. So in Spark this function just shift the timestamp value from UTC timezone to
    + * the given timezone.
    + *
    + * This function may return confusing result if the input is a string with timezone, e.g.
    + * '2018-03-13T06:18:23+00:00'. The reason is that, Spark firstly cast the string to timestamp
    + * according to the timezone in the string, and finally display the result by converting the
    + * timestamp to string according to the session local timezone.
    + *
    + * We may remove this function in Spark 3.0.
    --- End diff --
    
    The decision is not made yet, anyway I won't deprecate it in this PR, we definitely need a new PR


---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark issue #22543: [SPARK-23715][SQL][DOC] improve document for from/to_utc...

Posted by AmplabJenkins <gi...@git.apache.org>.
Github user AmplabJenkins commented on the issue:

    https://github.com/apache/spark/pull/22543
  
    Test FAILed.
    Refer to this link for build results (access rights to CI server needed): 
    https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/96597/
    Test FAILed.


---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark issue #22543: [SPARK-23715][SQL][DOC] improve document for from/to_utc...

Posted by SparkQA <gi...@git.apache.org>.
Github user SparkQA commented on the issue:

    https://github.com/apache/spark/pull/22543
  
    **[Test build #96599 has finished](https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/96599/testReport)** for PR 22543 at commit [`e033694`](https://github.com/apache/spark/commit/e033694ab62a79185cc60e0039657cd2eaea5f4d).
     * This patch **fails due to an unknown error code, -9**.
     * This patch merges cleanly.
     * This patch adds no public classes.


---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark issue #22543: [SPARK-23715][SQL] improve document for from/to_utc_time...

Posted by SparkQA <gi...@git.apache.org>.
Github user SparkQA commented on the issue:

    https://github.com/apache/spark/pull/22543
  
    **[Test build #96541 has finished](https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/96541/testReport)** for PR 22543 at commit [`b4d31f8`](https://github.com/apache/spark/commit/b4d31f862ce7c09ebd9a431997dbc9edf4a2f6f3).
     * This patch passes all tests.
     * This patch merges cleanly.
     * This patch adds no public classes.


---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org