You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@spark.apache.org by "Jim Huang (Jira)" <ji...@apache.org> on 2020/06/23 14:27:00 UTC
[jira] [Created] (SPARK-32076) Structured Streaming application
continuity when encountering streaming query task level error
Jim Huang created SPARK-32076:
---------------------------------
Summary: Structured Streaming application continuity when encountering streaming query task level error
Key: SPARK-32076
URL: https://issues.apache.org/jira/browse/SPARK-32076
Project: Spark
Issue Type: Improvement
Components: Structured Streaming
Affects Versions: 2.4.5
Environment: Apache Spark 2.4.5 Scala 2.11 without Hadoop
Hadoop 2.7.3 - YARN cluster
delta-core_ 2.11:0.6.1
Reporter: Jim Huang
From the Spark Structured Streaming application continuity perspective, the thread that ran this task was terminated with ERROR SPARK-31995 but to YARN it is still an active running job even though this instance of the Spark Structured Streaming job is no longer making any further processing. If the monitoring of the Spark Structured Streaming job is done only from the YARN job perspective, it may provide a false status. In this situation, should the Spark Structure Streaming application fail hard and completely (fail by Spark framework or Application exception handling)? Or should the developer investigate and develop some ideal monitoring implementation that has the right level of specificity to detect Spark Structured Streaming *task* level failures? Any references on these topics are much appreciated.
--
This message was sent by Atlassian Jira
(v8.3.4#803005)
---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscribe@spark.apache.org
For additional commands, e-mail: issues-help@spark.apache.org