You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@spark.apache.org by "Russell Spitzer (Jira)" <ji...@apache.org> on 2020/09/30 21:55:00 UTC

[jira] [Updated] (SPARK-33041) Better error messages when PySpark Java Gateway Crashes

     [ https://issues.apache.org/jira/browse/SPARK-33041?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]

Russell Spitzer updated SPARK-33041:
------------------------------------
    Summary: Better error messages when PySpark Java Gateway Crashes  (was: Better error messages when PySpark Java Gateway Fails to Start or Crashes)

> Better error messages when PySpark Java Gateway Crashes
> -------------------------------------------------------
>
>                 Key: SPARK-33041
>                 URL: https://issues.apache.org/jira/browse/SPARK-33041
>             Project: Spark
>          Issue Type: Improvement
>          Components: PySpark
>    Affects Versions: 2.4.7
>            Reporter: Russell Spitzer
>            Priority: Major
>
> Currently the startup works by opening the Gateway process and waiting until the the process has written the conn_info_file. Once the conn_file is written it proceeds to attempt to connect to the port.
> This connection can succeed and the process can start normally, but if the gateway process dies or is killed the error that the user ends up getting is a confusing "connection_failed" style error like
> {code}
> Traceback (most recent call last):
>   File "/usr/lib/spark-packages/spark2.4.4/python/lib/py4j-0.10.7-src.zip/py4j/java_gateway.py", line 929, in _get_connection
>     connection = self.deque.pop()
> IndexError: pop from an empty deque
> {code}
> Since we have a handle on the py4j process, we should probably check whether it has terminated before surfacing any exceptions like this. 
> CC [~holden]



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscribe@spark.apache.org
For additional commands, e-mail: issues-help@spark.apache.org