You are viewing a plain text version of this content. The canonical link for it is here.
Posted to dev@hive.apache.org by "Zheng Shao (JIRA)" <ji...@apache.org> on 2009/08/06 09:35:14 UTC

[jira] Created: (HIVE-733) allow specifying maximum number of serde exceptions to ignore

allow specifying maximum number of serde exceptions to ignore
-------------------------------------------------------------

                 Key: HIVE-733
                 URL: https://issues.apache.org/jira/browse/HIVE-733
             Project: Hadoop Hive
          Issue Type: New Feature
          Components: Query Processor
            Reporter: Zheng Shao


Sometimes user data contains some corrupted records. We should ignore the records till we hit some limit set in configuration.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.


[jira] Updated: (HIVE-733) allow specifying maximum number of serde exceptions to ignore

Posted by "Zheng Shao (JIRA)" <ji...@apache.org>.
     [ https://issues.apache.org/jira/browse/HIVE-733?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]

Zheng Shao updated HIVE-733:
----------------------------

    Attachment: HIVE-733.1.patch

HIVE-733.1.patch: This patch shows the idea. There is no test yet. Please comment.

> allow specifying maximum number of serde exceptions to ignore
> -------------------------------------------------------------
>
>                 Key: HIVE-733
>                 URL: https://issues.apache.org/jira/browse/HIVE-733
>             Project: Hadoop Hive
>          Issue Type: New Feature
>          Components: Query Processor
>            Reporter: Zheng Shao
>         Attachments: HIVE-733.1.patch
>
>
> Sometimes user data contains some corrupted records. We should ignore the records till we hit some limit set in configuration.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.


[jira] Commented: (HIVE-733) allow specifying maximum number of serde exceptions to ignore

Posted by "Namit Jain (JIRA)" <ji...@apache.org>.
    [ https://issues.apache.org/jira/browse/HIVE-733?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12740284#action_12740284 ] 

Namit Jain commented on HIVE-733:
---------------------------------

The parameter makes complete sense - in fact, traditional loader and querying tools (sql*loader etc.) support it
However, the parameter should be system wide and not specific to a mapper - you need to aggregate the counters
instead of looking at a single mapper.

> allow specifying maximum number of serde exceptions to ignore
> -------------------------------------------------------------
>
>                 Key: HIVE-733
>                 URL: https://issues.apache.org/jira/browse/HIVE-733
>             Project: Hadoop Hive
>          Issue Type: New Feature
>          Components: Query Processor
>            Reporter: Zheng Shao
>         Attachments: HIVE-733.1.patch
>
>
> Sometimes user data contains some corrupted records. We should ignore the records till we hit some limit set in configuration.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.


[jira] Commented: (HIVE-733) allow specifying maximum number of serde exceptions to ignore

Posted by "Andraz Tori (JIRA)" <ji...@apache.org>.
    [ https://issues.apache.org/jira/browse/HIVE-733?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12739931#action_12739931 ] 

Andraz Tori commented on HIVE-733:
----------------------------------

+1 

this makes sense, I'd set the  default to larger than 0.

> allow specifying maximum number of serde exceptions to ignore
> -------------------------------------------------------------
>
>                 Key: HIVE-733
>                 URL: https://issues.apache.org/jira/browse/HIVE-733
>             Project: Hadoop Hive
>          Issue Type: New Feature
>          Components: Query Processor
>            Reporter: Zheng Shao
>         Attachments: HIVE-733.1.patch
>
>
> Sometimes user data contains some corrupted records. We should ignore the records till we hit some limit set in configuration.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.