You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@spark.apache.org by "AKSonic (JIRA)" <ji...@apache.org> on 2018/09/25 06:17:00 UTC

[jira] [Updated] (SPARK-25520) The state of executors is KILLED on standalone

     [ https://issues.apache.org/jira/browse/SPARK-25520?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]

AKSonic updated SPARK-25520:
----------------------------
    Attachment: Spark.zip

> The state of executors is  KILLED on standalone
> -----------------------------------------------
>
>                 Key: SPARK-25520
>                 URL: https://issues.apache.org/jira/browse/SPARK-25520
>             Project: Spark
>          Issue Type: Bug
>          Components: Web UI
>    Affects Versions: 2.2.2, 2.3.1
>         Environment: spark 2.3.1
> spark 2.2.2
> Java 1.8.0_131
> scala 2.11.8
>  
>            Reporter: AKSonic
>            Priority: Major
>         Attachments: Spark.zip
>
>
> I create spark standalone cluster (4 servers) by using spark 2.3.1. The job can be finished on Completed Drivers. I also can get the result by driver log. But the status of all executors show KILLED state. The log show the following error.
> 2018-09-25 00:47:37 INFO CoarseGrainedExecutorBackend:54 - Driver commanded a shutdown
> 2018-09-25 00:47:37 ERROR CoarseGrainedExecutorBackend:43 - RECEIVED SIGNAL TERM utdown
> I also try spark 2.2.2. I see the same issues on the GUI. All executors are KILLED status.
> Is it right? what is the problem?
> -------------------------------------------- Config ----------------------------------------------------------------
> *spark-env.sh:*
> export SPARK_PUBLIC_DNS=hostname1
> export SCALA_HOME=/opt/gpf/bigdata/scala-2.11.8
> export JAVA_HOME=/usr/java/jdk1.8.0_131
> export HADOOP_HOME=/opt/bigdata/hadoop-2.6.5
> export HADOOP_CONF_DIR=$HADOOP_HOME/etc/hadoop
> export SPARK_HISTORY_OPTS="-Dspark.history.fs.logDirectory=file:///spark/spark-event-dir -Dspark.history.ui.port=16066 -Dspark.history.retainedApplications=30 -Dspark.history.fs.cleaner.enabled=true -Dspark.history.fs.cleaner.interval=1d -Dspark.history.fs.cleaner.maxAge=7d"
> export SPARK_DAEMON_JAVA_OPTS="-Dspark.deploy.recoveryMode=ZOOKEEPER -Dspark.deploy.zookeeper.url=hostname1:2181,hostname2:2181,hostname3:2181 -Dspark.deploy.zookeeper.dir=/opt/bigdata/spark-2.3.1/zk-recovery-dir"
> SPARK_LOCAL_DIRS=/opt/bigdata/spark-2.3.1/local-dir
> SPARK_DRIVER_MEMORY=1G
> *spark-defaults.conf:*
> spark.eventLog.enabled true
> spark.eventLog.compress true
> spark.eventLog.dir file:///spark/spark-event-dir
> *slaves:*
> hostname1
> hostname2
> hostname3
> hostname4
> ------------------------------------------------Testing ------------------------------------------------------------
> *Testing program1 (Java):*
> public class JDBCApp {
>  private static final String DB_OLAP_UAT_URL = "jdbc:sqlserver://dbhost";
>  private static final String DB_DRIVER = "com.microsoft.sqlserver.jdbc.SQLServerDriver";
>  private static final String SQL_TEXT = "select top 10 * from table1";
>  private static final String DB_OLAP_UAT_USR = "";
>  private static final String DB_OLAP_UAT_PWD = "";
>  public static void main(String[] args) {
>  System.setProperty("spark.sql.warehouse.dir","file:///bigdata/spark/spark-warehouse");
> // Logger.getLogger("org.apache.spark").setLevel(Level.DEBUG);
>  SparkSession spark = SparkSession
>  .builder()
>  .appName("JDBCApp")
>  .getOrCreate();
>  Dataset<Row> jdbcDF = spark.read()
>  .format("jdbc")
>  .option("driver", DB_DRIVER)
>  .option("url", DB_OLAP_UAT_URL)
>  .option("dbtable", "(" + SQL_TEXT + ") tmp")
>  .option("user", DB_OLAP_UAT_USR)
>  .option("password", DB_OLAP_UAT_PWD)
>  .load();
>  jdbcDF.show();
>  }
> }
> *Testing program2 (Java):*
> public class SimpleApp {
>  public static void main(String[] args) {
>  String filePath = args[0];
>  Logger logger = Logger.getLogger("org.apache.spark");
> // logger.setLevel(Level.DEBUG);
>  SparkSession spark = SparkSession.builder()
>  .appName("Simple Application")
>  .getOrCreate();
>  Dataset<String> logData = spark.read().textFile(filePath).cache();
>  long numAs = logData.filter((FilterFunction<String>) s -> s.contains("e")).count();
>  long numBs = logData.filter((FilterFunction<String>) s -> s.contains("r")).count();
>  logger.info("Lines with a: " + numAs + ", lines with b: " + numBs);
>  spark.stop();
>  }
> }
> You can run above 2 testing programs, and then you can see the state of executors are KILLED.
> Another issue is that I can't see the stdout/stderr of driver on Executors tab on spark history server. I only can see the stdout/stderr of executors. It causes after restart spark standalone cluster, I can't see the result of driver on both spark UI and spark history server. 



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscribe@spark.apache.org
For additional commands, e-mail: issues-help@spark.apache.org