You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@spark.apache.org by "Manohar Chamaraju (Jira)" <ji...@apache.org> on 2021/02/02 07:53:00 UTC

[jira] [Commented] (SPARK-34293) kubernetes executor pod unable to access secure hdfs

    [ https://issues.apache.org/jira/browse/SPARK-34293?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17276913#comment-17276913 ] 

Manohar Chamaraju commented on SPARK-34293:
-------------------------------------------

Update:
 # In client mode by adding fs.defaultFS in core-site.xml fixed the issue for me.
 # what do to work is usage of hadoop-conf configmap in client mode.

> kubernetes executor pod unable to access secure hdfs
> ----------------------------------------------------
>
>                 Key: SPARK-34293
>                 URL: https://issues.apache.org/jira/browse/SPARK-34293
>             Project: Spark
>          Issue Type: Bug
>          Components: Kubernetes
>    Affects Versions: 3.0.1
>            Reporter: Manohar Chamaraju
>            Priority: Major
>         Attachments: driver.log, executor.log, image-2021-01-30-00-13-18-234.png, image-2021-01-30-00-14-14-329.png, image-2021-01-30-00-14-45-335.png, image-2021-01-30-00-20-54-620.png, image-2021-01-30-00-33-02-109.png, image-2021-01-30-00-34-05-946.png
>
>
> Steps to reproduce
>  # Configure secure HDFS(kerberos) cluster running as containers in kubernetes.
>  # Configure KDC on centos and create keytab for user principal hdfs, in hdfsuser.keytab.
>  # Genearte spark image(v3.0.1), to spawn as container out of spark image.
>  # Inside spark container, run export HADOOP_CONF_DIR=/etc/hadoop/conf/ with core-site.xml configuration as below 
>  !image-2021-01-30-00-13-18-234.png!
>  # Create configmap kbr-conf 
>  !image-2021-01-30-00-14-14-329.png!
>  # Run the command /opt/spark/bin/spark-submit \
>  --deploy-mode client \
>  --executor-memory 1g\
>  --executor-memory 1g\
>  --executor-cores 1\
>  --class org.apache.spark.examples.HdfsTest \
>  --conf spark.kubernetes.namespace=arcsight-installer-lh7fm\
>  --master k8s://[https://172.17.17.1:443|https://172.17.17.1/] \
>  --conf spark.kubernetes.authenticate.driver.serviceAccountName=spark \
>  --conf spark.app.name=spark-hdfs \
>  --conf spark.executer.instances=1 \
>  --conf spark.kubernetes.node.selector.spark=yes\
>  --conf spark.kubernetes.node.selector.Worker=label\
>  --conf spark.kubernetes.container.image=manohar/spark:v3.0.1 \
>  --conf spark.kubernetes.kerberos.enabled=true \
>  --conf spark.kubernetes.kerberos.krb5.configMapName=krb5-conf \
>  --conf spark.kerberos.keytab=/data/hdfsuser.keytab \
>  --conf spark.kerberos.principal=hdfs@DOM047600.LAB \
>  local:///opt/spark/examples/jars/spark-examples_2.12-3.0.1.jar \
>  hdfs://hdfs-namenode:30820/staging-directory.
>  # On running this command driver is able to connect hdfs with kerberos but execurtor fails to connect to secure hdfs and below is the logs 
> !image-2021-01-30-00-34-05-946.png!
>  # Some of observation
>  ## In Client mode, --conf spark.kubernetes.hadoop.configMapName=hadoop-conf as not effect only works after HADOOP_CONF_DIR is set. Below was the contents of hadoop-conf configmap.
>  !image-2021-01-30-00-20-54-620.png!
>  ## Ran the command in cluster mode as well, in cluster mode also executor could not connect to secure hdfs.
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscribe@spark.apache.org
For additional commands, e-mail: issues-help@spark.apache.org