You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@arrow.apache.org by "Nick Throckmorton (JIRA)" <ji...@apache.org> on 2019/04/05 19:02:00 UTC

[jira] [Commented] (ARROW-1989) [Python] Better UX on timestamp conversion to Pandas

    [ https://issues.apache.org/jira/browse/ARROW-1989?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16811172#comment-16811172 ] 

Nick Throckmorton commented on ARROW-1989:
------------------------------------------

Agreed. I would propose this message to point users to a potential solution.
ArrowInvalid: Casting from timestamp[ns] to timestamp[us] would lose data: XX
To accept this, use `allow_truncated_timestamps=True`. 
 

> [Python] Better UX on timestamp conversion to Pandas
> ----------------------------------------------------
>
>                 Key: ARROW-1989
>                 URL: https://issues.apache.org/jira/browse/ARROW-1989
>             Project: Apache Arrow
>          Issue Type: Improvement
>          Components: Python
>            Reporter: Uwe L. Korn
>            Priority: Major
>             Fix For: 0.14.0
>
>
> Converting timestamp columns to Pandas, users often have the problem that they have dates that are larger than Pandas can represent with their nanosecond representation. Currently they simply see an Arrow exception and think that this problem is caused by Arrow. We should try to change the error from
> {code}
> ArrowInvalid: Casting from timestamp[ns] to timestamp[us] would lose data: XX
> {code}
> to something along the lines of 
> {code}
> ArrowInvalid: Casting from timestamp[ns] to timestamp[us] would lose data: XX. This conversion is needed as Pandas does only support nanosecond timestamps. Your data is likely out of the range that can be represented with nanosecond resolution.
> {code}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)