You are viewing a plain text version of this content. The canonical link for it is here.
Posted to dev@hive.apache.org by "Raghotham Murthy (JIRA)" <ji...@apache.org> on 2009/03/11 02:48:50 UTC

[jira] Resolved: (HIVE-295) Handles error input

     [ https://issues.apache.org/jira/browse/HIVE-295?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]

Raghotham Murthy resolved HIVE-295.
-----------------------------------

    Resolution: Duplicate

Fixing HIVE-293 should also fix this issue. Parse exceptions thrown by serdes should be counted and then the query should fail based on a threshold.

> Handles error input
> -------------------
>
>                 Key: HIVE-295
>                 URL: https://issues.apache.org/jira/browse/HIVE-295
>             Project: Hadoop Hive
>          Issue Type: New Feature
>            Reporter: qing yan
>
> It is common for Hive to encounter bad records when processing massive data. 
> Currently the forked hadoop job would throw exception and cause the whole job to fail ultimately. 
> Hive should handle error input systematically instead of treat it as an exception.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.