You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@spark.apache.org by "Nishant Ranjan (JIRA)" <ji...@apache.org> on 2019/04/10 09:14:01 UTC

[jira] [Created] (SPARK-27426) SparkAppHandle states not getting updated in Kubernetes

Nishant Ranjan created SPARK-27426:
--------------------------------------

             Summary: SparkAppHandle states not getting updated in Kubernetes
                 Key: SPARK-27426
                 URL: https://issues.apache.org/jira/browse/SPARK-27426
             Project: Spark
          Issue Type: Bug
          Components: Kubernetes
    Affects Versions: 2.4.0
         Environment: # Cent os 7
 # Kubernetes 1.14
 # Spark 2.4.0
            Reporter: Nishant Ranjan


While launching Spark application through "startApplication()" , SparkAppHandle state is not getting updated. 

sparkLaunch = new SparkLauncher()
 .setSparkHome("/root/test/spark-2.4.0-bin-hadoop2.7")
 .setMaster("k8s://https://172.16.23.30:6443")
 .setVerbose(true)
 .addSparkArg("--verbose")
 .setAppResource("local:///opt/spark/examples/jars/spark-examples_2.11-2.4.0.jar")
 .setConf("spark.app.name","spark-pi")
 .setMainClass("org.apache.spark.examples.SparkPi")
 .setConf("spark.executor.instances","5")
 .setConf("spark.kubernetes.container.image","registry.renovite.com/spark:v2")
 .setConf("spark.kubernetes.driver.pod.name","spark-pi-driver")
 .setConf("spark.kubernetes.container.image.pullSecrets","dev-registry-key")
 .setConf("spark.kubernetes.authenticate.driver.serviceAccountName","spark")
 .setDeployMode("cluster")
 ;



SparkAppHandle handle = sparkLaunch.startApplication(); 



Observations:
 # Now, I tried listeners etc but handle.getState() returns UNKNOWN and when Spark application is completed. state changes to LOST.
 # SparkAppHandle is not null
 # handle.getAppId() is always null.

My best guess is that communication is not working properly between listener and Spark driver in Kubernetes. 

 



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscribe@spark.apache.org
For additional commands, e-mail: issues-help@spark.apache.org