You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@hive.apache.org by "Wang Zhiqiang (JIRA)" <ji...@apache.org> on 2016/09/07 09:52:20 UTC

[jira] [Commented] (HIVE-11414) Fix OOM in MapTask with many input partitions with RCFile

    [ https://issues.apache.org/jira/browse/HIVE-11414?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15470173#comment-15470173 ] 

Wang Zhiqiang commented on HIVE-11414:
--------------------------------------

We met the same issue when processing too many small RCFiles. Any one is working on this? 
Set mapred.min.split.size.per.node to a small value can solve it, but we want to set mapred.min.split.size.per.node to a bigger default value(eg, 32M), for reducing map overhead.

> Fix OOM in MapTask with many input partitions with RCFile
> ---------------------------------------------------------
>
>                 Key: HIVE-11414
>                 URL: https://issues.apache.org/jira/browse/HIVE-11414
>             Project: Hive
>          Issue Type: Improvement
>          Components: File Formats, Serializers/Deserializers
>    Affects Versions: 0.11.0, 0.12.0, 0.14.0, 0.13.1, 1.2.0
>            Reporter: Zheng Shao
>            Priority: Minor
>
> MapTask hit OOM in the following situation in our production environment:
> * src: 2048 partitions, each with 1 file of about 2MB using RCFile format
> * query: INSERT OVERWRITE TABLE tgt SELECT * FROM src
> * Hadoop version: Both on CDH 4.7 using MR1 and CDH 5.4.1 using YARN.
> * MapTask memory Xmx: 1.5GB
> By analyzing the heap dump using jhat, we realized that the problem is:
> * One single mapper is processing many partitions (because of CombineHiveInputFormat)
> * Each input path (equivalent to partition here) will construct its own SerDe
> * Each SerDe will do its own caching of deserialized object (and try to reuse it), but will never release it (in this case, the serde2.columnar.ColumnarSerDeBase has a field cachedLazyStruct which can take a lot of space - pretty much the last N rows of a file where N is the number of rows in a columnar block).
> * This problem may exist in other SerDe as well, but columnar file format are affected the most because they need bigger cache for the last N rows instead of 1 row.
> Proposed solution:
> * Remove cachedLazyStruct in serde2.columnar.ColumnarSerDeBase.  The cost saving of not recreating a single object is too small compared to processing N rows.
> Alternative solutions:
> * We can also free up the whole SerDe after processing a block/file.  The problem with that is that the input splits may contain multiple blocks/files that maps to the same SerDe, and recreating a SerDe is a much bigger change to the code.
> * We can also move the SerDe creation/free-up to the place when input file changes.  But that requires a much bigger change to the code.
> * We can also add a "cleanup()" method to SerDe interface that release the cached object, but that change is not backward compatible with many SerDes that people have wrote.
> * We can make cachedLazyStruct in serde2.columnar.ColumnarSerDeBase a weakly referenced object, but that feels like an overkill.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)