You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@kylin.apache.org by "Liu Shaohui (JIRA)" <ji...@apache.org> on 2019/06/17 11:26:00 UTC
[jira] [Created] (KYLIN-4048) Too long spark cube building time for
too many eviction and loading for dict slices
Liu Shaohui created KYLIN-4048:
----------------------------------
Summary: Too long spark cube building time for too many eviction and loading for dict slices
Key: KYLIN-4048
URL: https://issues.apache.org/jira/browse/KYLIN-4048
Project: Kylin
Issue Type: Improvement
Affects Versions: v2.5.2
Reporter: Liu Shaohui
In our cluster, a cube building costs too long. In the log for spark, we found there are too many eviction and loading for dict slices in AppendTrieDictionary.
{code:java}
$ grep "read slice from" spark.log | wc -l
119721
$ grep "Evict slice with key" spark.log| wc -l
119634
{code}
The reason is that the memory of spark executor(4G) is not enough to hold all the slices of dict(3.3G in hdfs) in memory, which cause the bad performance of cube building.
--
This message was sent by Atlassian JIRA
(v7.6.3#76005)