You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@hive.apache.org by "Wang Zhiqiang (JIRA)" <ji...@apache.org> on 2016/09/07 09:52:20 UTC
[jira] [Commented] (HIVE-11414) Fix OOM in MapTask with many input
partitions with RCFile
[ https://issues.apache.org/jira/browse/HIVE-11414?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15470173#comment-15470173 ]
Wang Zhiqiang commented on HIVE-11414:
--------------------------------------
We met the same issue when processing too many small RCFiles. Any one is working on this?
Set mapred.min.split.size.per.node to a small value can solve it, but we want to set mapred.min.split.size.per.node to a bigger default value(eg, 32M), for reducing map overhead.
> Fix OOM in MapTask with many input partitions with RCFile
> ---------------------------------------------------------
>
> Key: HIVE-11414
> URL: https://issues.apache.org/jira/browse/HIVE-11414
> Project: Hive
> Issue Type: Improvement
> Components: File Formats, Serializers/Deserializers
> Affects Versions: 0.11.0, 0.12.0, 0.14.0, 0.13.1, 1.2.0
> Reporter: Zheng Shao
> Priority: Minor
>
> MapTask hit OOM in the following situation in our production environment:
> * src: 2048 partitions, each with 1 file of about 2MB using RCFile format
> * query: INSERT OVERWRITE TABLE tgt SELECT * FROM src
> * Hadoop version: Both on CDH 4.7 using MR1 and CDH 5.4.1 using YARN.
> * MapTask memory Xmx: 1.5GB
> By analyzing the heap dump using jhat, we realized that the problem is:
> * One single mapper is processing many partitions (because of CombineHiveInputFormat)
> * Each input path (equivalent to partition here) will construct its own SerDe
> * Each SerDe will do its own caching of deserialized object (and try to reuse it), but will never release it (in this case, the serde2.columnar.ColumnarSerDeBase has a field cachedLazyStruct which can take a lot of space - pretty much the last N rows of a file where N is the number of rows in a columnar block).
> * This problem may exist in other SerDe as well, but columnar file format are affected the most because they need bigger cache for the last N rows instead of 1 row.
> Proposed solution:
> * Remove cachedLazyStruct in serde2.columnar.ColumnarSerDeBase. The cost saving of not recreating a single object is too small compared to processing N rows.
> Alternative solutions:
> * We can also free up the whole SerDe after processing a block/file. The problem with that is that the input splits may contain multiple blocks/files that maps to the same SerDe, and recreating a SerDe is a much bigger change to the code.
> * We can also move the SerDe creation/free-up to the place when input file changes. But that requires a much bigger change to the code.
> * We can also add a "cleanup()" method to SerDe interface that release the cached object, but that change is not backward compatible with many SerDes that people have wrote.
> * We can make cachedLazyStruct in serde2.columnar.ColumnarSerDeBase a weakly referenced object, but that feels like an overkill.
--
This message was sent by Atlassian JIRA
(v6.3.4#6332)