You are viewing a plain text version of this content. The canonical link for it is here.
Posted to dev@hive.apache.org by "Navis (JIRA)" <ji...@apache.org> on 2012/08/27 07:48:07 UTC
[jira] [Updated] (HIVE-3387) meta data file size exceeds limit
[ https://issues.apache.org/jira/browse/HIVE-3387?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
Navis updated HIVE-3387:
------------------------
Status: Patch Available (was: Open)
You're right. This should be fixed.
https://reviews.facebook.net/D4899
> meta data file size exceeds limit
> ---------------------------------
>
> Key: HIVE-3387
> URL: https://issues.apache.org/jira/browse/HIVE-3387
> Project: Hive
> Issue Type: Bug
> Affects Versions: 0.7.1
> Reporter: Alexander Alten-Lorenz
> Fix For: 0.9.1
>
>
> The cause is certainly that we use an array list instead of a set structure in the split locations API. Looks like a bug in Hive's CombineFileInputFormat.
> Reproduce:
> Set mapreduce.jobtracker.split.metainfo.maxsize=100000000 when submitting the Hive query. Run a big hive query that write data into a partitioned table. Due to the large number of splits, you encounter an exception on the job submitted to Hadoop and the exception said:
> meta data size exceeds 100000000.
--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira