You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@spark.apache.org by "Russell Spitzer (Jira)" <ji...@apache.org> on 2020/09/30 21:55:00 UTC
[jira] [Created] (SPARK-33041) Better error messages when PySpark
Java Gateway Fails to Start or Crashes
Russell Spitzer created SPARK-33041:
---------------------------------------
Summary: Better error messages when PySpark Java Gateway Fails to Start or Crashes
Key: SPARK-33041
URL: https://issues.apache.org/jira/browse/SPARK-33041
Project: Spark
Issue Type: Improvement
Components: PySpark
Affects Versions: 2.4.7
Reporter: Russell Spitzer
Currently the startup works by opening the Gateway process and waiting until the the process has written the conn_info_file. Once the conn_file is written it proceeds to attempt to connect to the port.
This connection can succeed and the process can start normally, but if the gateway process dies or is killed the error that the user ends up getting is a confusing "connection_failed" style error like
{code}
Traceback (most recent call last):
File "/usr/lib/spark-packages/spark2.4.4/python/lib/py4j-0.10.7-src.zip/py4j/java_gateway.py", line 929, in _get_connection
connection = self.deque.pop()
IndexError: pop from an empty deque
{code}
Since we have a handle on the py4j process, we should probably check whether it has terminated before surfacing any exceptions like this.
CC [~holden]
--
This message was sent by Atlassian Jira
(v8.3.4#803005)
---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscribe@spark.apache.org
For additional commands, e-mail: issues-help@spark.apache.org