You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@spark.apache.org by "Lianhui Wang (JIRA)" <ji...@apache.org> on 2015/01/29 05:59:35 UTC

[jira] [Issue Comment Deleted] (SPARK-5162) Python yarn-cluster mode

     [ https://issues.apache.org/jira/browse/SPARK-5162?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]

Lianhui Wang updated SPARK-5162:
--------------------------------
    Comment: was deleted

(was: The previous problem is that packages.egg is not found. but from your commonds, I donnot see packages.egg from your spark-submit and there are only package1.egg and package1.egg.so i think you need to add packages.egg to  --py-files and you can try it. 
Or firstly you can run on yarn-client mode, if it is ok then try to run on yarn cluste mode. That can help us to find some problems.)

> Python yarn-cluster mode
> ------------------------
>
>                 Key: SPARK-5162
>                 URL: https://issues.apache.org/jira/browse/SPARK-5162
>             Project: Spark
>          Issue Type: New Feature
>          Components: PySpark, YARN
>            Reporter: Dana Klassen
>              Labels: cluster, python, yarn
>
> Running pyspark in yarn is currently limited to ‘yarn-client’ mode. It would be great to be able to submit python applications to the cluster and (just like java classes) have the resource manager setup an AM on any node in the cluster. Does anyone know the issues blocking this feature? I was snooping around with enabling python apps:
> Removing the logic stopping python and yarn-cluster from sparkSubmit.scala
> ...
>     // The following modes are not supported or applicable
>     (clusterManager, deployMode) match {
>       ...
>       case (_, CLUSTER) if args.isPython =>
>         printErrorAndExit("Cluster deploy mode is currently not supported for python applications.")
>       ...
>     }
> …
> and submitting application via:
> HADOOP_CONF_DIR={{insert conf dir}} ./bin/spark-submit --master yarn-cluster --num-executors 2  —-py-files {{insert location of egg here}} --executor-cores 1  ../tools/canary.py
> Everything looks to run alright, pythonRunner is picked up as main class, resources get setup, yarn client gets launched but falls flat on its face:
> 2015-01-08 18:48:03,444 INFO org.apache.hadoop.yarn.server.nodemanager.containermanager.localizer.ResourceLocalizationService: DEBUG: FAILED { {{redacted}}/.sparkStaging/application_1420594669313_4687/canary.py, 1420742868009, FILE, null }, Resource {{redacted}}/.sparkStaging/application_1420594669313_4687/canary.py changed on src filesystem (expected 1420742868009, was 1420742869284
> and
> 2015-01-08 18:48:03,446 INFO org.apache.hadoop.yarn.server.nodemanager.containermanager.localizer.LocalizedResource: Resource {{redacted}}/.sparkStaging/application_1420594669313_4687/canary.py(->/data/4/yarn/nm/usercache/klassen/filecache/11/canary.py) transitioned from DOWNLOADING to FAILED
> Tracked this down to the apache hadoop code(FSDownload.java line 249) related to container localization of files upon downloading. At this point thought it would be best to raise the issue here and get input.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscribe@spark.apache.org
For additional commands, e-mail: issues-help@spark.apache.org