You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@spark.apache.org by "Martin Andersson (Jira)" <ji...@apache.org> on 2021/11/04 18:11:00 UTC

[jira] [Comment Edited] (SPARK-31726) Make spark.files available in driver with cluster deploy mode on kubernetes

    [ https://issues.apache.org/jira/browse/SPARK-31726?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17438862#comment-17438862 ] 

Martin Andersson edited comment on SPARK-31726 at 11/4/21, 6:10 PM:
--------------------------------------------------------------------

I've also run into this issue. Trying to simply include logging and application configuration using {{--files}}, but I suppose I'll have to include those in the docker image instead until this issue gets fixed.

EDIT: I tried using the --jars option as well to include a logging config file as well. Files included in this fashion should also be added to the classpath, but that doesn't seem to be the case as well.


was (Author: beregon87):
I've also run into this issue. Trying to simply include logging and application configuration using {{--files}}, but I suppose I'll have to include those in the docker image instead until this issue gets fixed.

> Make spark.files available in driver with cluster deploy mode on kubernetes
> ---------------------------------------------------------------------------
>
>                 Key: SPARK-31726
>                 URL: https://issues.apache.org/jira/browse/SPARK-31726
>             Project: Spark
>          Issue Type: Improvement
>          Components: Kubernetes, Spark Core
>    Affects Versions: 3.0.0
>            Reporter: koert kuipers
>            Priority: Minor
>
> currently on yarn with cluster deploy mode --files makes the files available for driver and executors and also put them on classpath for driver and executors.
> on k8s with cluster deploy mode --files makes the files available on executors but they are not on classpath. it does not make the files available on driver and they are not on driver classpath.
> it would be nice if the k8s behavior was consistent with yarn, or at least makes the files available on driver. once the files are available there is a simple workaround to get them on classpath using spark.driver.extraClassPath="./"
> background:
> we recently started testing kubernetes for spark. our main platform is yarn on which we use client deploy mode. our first experience was that client deploy mode was difficult to use on k8s (we dont launch from inside a pod). so we switched to cluster deploy mode, which seems to behave well on k8s. but then we realized that our program rely on reading files on classpath (application.conf, log4j.properties etc.) that are on the client but now are no longer on the driver (since driver is no longer on client). an easy fix for this seems to be to ship the files using --files to make them available on driver, but we could not get this to work.
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscribe@spark.apache.org
For additional commands, e-mail: issues-help@spark.apache.org