You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@datafu.apache.org by ey...@apache.org on 2019/04/03 13:37:34 UTC

[datafu] branch spark-tmp updated: Text cleanup

This is an automated email from the ASF dual-hosted git repository.

eyal pushed a commit to branch spark-tmp
in repository https://gitbox.apache.org/repos/asf/datafu.git


The following commit(s) were added to refs/heads/spark-tmp by this push:
     new 3b4e142  Text cleanup
3b4e142 is described below

commit 3b4e142c635810b5813ed2b375f9b92a707d1b51
Author: Eyal Allweil <ey...@apache.org>
AuthorDate: Wed Apr 3 16:37:10 2019 +0300

    Text cleanup
---
 datafu-spark/src/main/scala/datafu/spark/PythonPathsManager.scala | 4 +---
 1 file changed, 1 insertion(+), 3 deletions(-)

diff --git a/datafu-spark/src/main/scala/datafu/spark/PythonPathsManager.scala b/datafu-spark/src/main/scala/datafu/spark/PythonPathsManager.scala
index b451846..9c16898 100644
--- a/datafu-spark/src/main/scala/datafu/spark/PythonPathsManager.scala
+++ b/datafu-spark/src/main/scala/datafu/spark/PythonPathsManager.scala
@@ -53,12 +53,10 @@ abstract class PythonResource(val resourcePath: String, val isAbsolutePath: Bool
   *
   * 1) When launching spark:
   *   the files need to be added to spark.executorEnv.PYTHONPATH
-  *   on cluster mode this is handled by com.paypal.risk.ars.bigdata.execution_fw.hadoop.spark.SparkTask
-  *   and on test mode by com.paypal.risk.ars.bigdata.hadoop.spark.utils.SparkTestManager
   *
   * 2) When executing python file via bridge:
   *   the files need to be added to the process PYTHONPATH. This is different than the previous phase because
-  *   this python process is spawn by zonkey, not by spark, and always on the driver.
+  *   this python process is spawned by datafu-spark, not by spark, and always on the driver.
   */
 object PythonPathsManager {