You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@spark.apache.org by "Nishant Ranjan (JIRA)" <ji...@apache.org> on 2019/04/10 09:14:01 UTC
[jira] [Created] (SPARK-27426) SparkAppHandle states not getting
updated in Kubernetes
Nishant Ranjan created SPARK-27426:
--------------------------------------
Summary: SparkAppHandle states not getting updated in Kubernetes
Key: SPARK-27426
URL: https://issues.apache.org/jira/browse/SPARK-27426
Project: Spark
Issue Type: Bug
Components: Kubernetes
Affects Versions: 2.4.0
Environment: # Cent os 7
# Kubernetes 1.14
# Spark 2.4.0
Reporter: Nishant Ranjan
While launching Spark application through "startApplication()" , SparkAppHandle state is not getting updated.
sparkLaunch = new SparkLauncher()
.setSparkHome("/root/test/spark-2.4.0-bin-hadoop2.7")
.setMaster("k8s://https://172.16.23.30:6443")
.setVerbose(true)
.addSparkArg("--verbose")
.setAppResource("local:///opt/spark/examples/jars/spark-examples_2.11-2.4.0.jar")
.setConf("spark.app.name","spark-pi")
.setMainClass("org.apache.spark.examples.SparkPi")
.setConf("spark.executor.instances","5")
.setConf("spark.kubernetes.container.image","registry.renovite.com/spark:v2")
.setConf("spark.kubernetes.driver.pod.name","spark-pi-driver")
.setConf("spark.kubernetes.container.image.pullSecrets","dev-registry-key")
.setConf("spark.kubernetes.authenticate.driver.serviceAccountName","spark")
.setDeployMode("cluster")
;
SparkAppHandle handle = sparkLaunch.startApplication();
Observations:
# Now, I tried listeners etc but handle.getState() returns UNKNOWN and when Spark application is completed. state changes to LOST.
# SparkAppHandle is not null
# handle.getAppId() is always null.
My best guess is that communication is not working properly between listener and Spark driver in Kubernetes.
--
This message was sent by Atlassian JIRA
(v7.6.3#76005)
---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscribe@spark.apache.org
For additional commands, e-mail: issues-help@spark.apache.org