You are viewing a plain text version of this content. The canonical link for it is here.
Posted to dev@hive.apache.org by "Venki Korukanti (JIRA)" <ji...@apache.org> on 2014/02/06 23:57:20 UTC

[jira] [Commented] (HIVE-3844) Unix timestamps don't seem to be read correctly from HDFS as Timestamp column

    [ https://issues.apache.org/jira/browse/HIVE-3844?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13893934#comment-13893934 ] 

Venki Korukanti commented on HIVE-3844:
---------------------------------------

Updated the patch to use Pattern/Matcher to identify the format of timestamp. I thought about not using the BigDecimal, but the output has floating point errors.

Reg: Should we really be allowing numeric formats to be read as timestamp?
I have seen multiple customers that have logs with different timestamp formats and they want to create one table with timestamp schema and read all sources of data.

> Unix timestamps don't seem to be read correctly from HDFS as Timestamp column
> -----------------------------------------------------------------------------
>
>                 Key: HIVE-3844
>                 URL: https://issues.apache.org/jira/browse/HIVE-3844
>             Project: Hive
>          Issue Type: Bug
>          Components: Serializers/Deserializers
>    Affects Versions: 0.8.0
>            Reporter: Mark Grover
>            Assignee: Venki Korukanti
>         Attachments: HIVE-3844.1.patch.txt
>
>
> Serega Shepak pointed out that something like
> {code}
> select cast(date_occurrence as timestamp) from xvlr_data limit 10
> {code}
> where  date_occurrence has BIGINT type (timestamp in milliseconds) works. But it doesn't work if the declared type is TIMESTAMP on column. The data in the date_occurence column in unix timestamp in millis.



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)