You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@spark.apache.org by "Josh Rosen (JIRA)" <ji...@apache.org> on 2014/08/20 19:22:27 UTC

[jira] [Commented] (SPARK-3140) PySpark start-up throws confusing exception

    [ https://issues.apache.org/jira/browse/SPARK-3140?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14104153#comment-14104153 ] 

Josh Rosen commented on SPARK-3140:
-----------------------------------

Is it the gateway server JVM -> PySpark driver communication that's getting messed up (the step where the Python driver's Java child process launches with some ephemeral port and communicates that port number back to the Python driver)?  Wouldn't that imply that the GatewayServer is has some extra logging to stdout that's being printed before it writes the port number?


> PySpark start-up throws confusing exception
> -------------------------------------------
>
>                 Key: SPARK-3140
>                 URL: https://issues.apache.org/jira/browse/SPARK-3140
>             Project: Spark
>          Issue Type: Bug
>          Components: PySpark
>    Affects Versions: 1.0.2
>            Reporter: Andrew Or
>            Priority: Critical
>
> Currently we read the pyspark port through stdout of the spark-submit subprocess. However, if there is stdout interference, e.g. spark-submit echoes something unexpected to stdout, we print the following:
> {code}
> Exception: Launching GatewayServer failed! (Warning: unexpected output detected.)
> {code}
> This condition is fine. However, we actually throw the same exception if there is *no* output from the subprocess as well. This is very confusing because it implies that the subprocess is outputting something (possibly whitespace, which is not visible) when it's actually not.



--
This message was sent by Atlassian JIRA
(v6.2#6252)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscribe@spark.apache.org
For additional commands, e-mail: issues-help@spark.apache.org