You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@spark.apache.org by "ZemingZhao (JIRA)" <ji...@apache.org> on 2015/08/20 09:27:46 UTC

[jira] [Created] (SPARK-10132) daemon crash caused by memory leak

ZemingZhao created SPARK-10132:
----------------------------------

             Summary: daemon crash caused by memory leak
                 Key: SPARK-10132
                 URL: https://issues.apache.org/jira/browse/SPARK-10132
             Project: Spark
          Issue Type: Bug
          Components: Spark Core
    Affects Versions: 1.4.1, 1.3.1
         Environment: 1. Cluster: 7 Redhat notes cluster, each has 32 cores

2. OS type: Red Hat Enterprise Linux Server release 7.1 (Maipo)

3. Java version:   tried both Oracle jdk 1.6 and 1.7 
java version "1.6.0_13"
Java(TM) SE Runtime Environment (build 1.6.0_13-b03)
Java HotSpot(TM) 64-Bit Server VM (build 11.3-b02, mixed mode)
java version "1.7.0"
Java(TM) SE Runtime Environment (build 1.7.0-b147)
Java HotSpot(TM) 64-Bit Server VM (build 21.0-b17, mixed mode)

4. JVM Option on spark-env.sh, 
Notes: SPARK_DAEMON_MEMORY was set to 300M to speed up the crash process
SPARK_DAEMON_JAVA_OPTS="-Xloggc:/root/spark/oracle_gclog"
SPARK_DAEMON_MEMORY=300m
            Reporter: ZemingZhao
            Priority: Critical


constantly submit short batch workload onto spark. 
spark master and worker will crash casued by memory leak.
according to the gclog and jmap info, this leak should be related to Akka but cannot find the root cause by now.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscribe@spark.apache.org
For additional commands, e-mail: issues-help@spark.apache.org