You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@spark.apache.org by "Liangchang Zhu (JIRA)" <ji...@apache.org> on 2019/02/01 06:04:00 UTC

[jira] [Closed] (SPARK-24404) Increase currentEpoch when meet a EpochMarker in ContinuousQueuedDataReader.next() in CP mode based on PR #21353 #21332 #21293 and the latest master

     [ https://issues.apache.org/jira/browse/SPARK-24404?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]

Liangchang Zhu closed SPARK-24404.
----------------------------------

> Increase currentEpoch when meet a EpochMarker in ContinuousQueuedDataReader.next()  in CP mode based  on PR #21353 #21332 #21293 and the latest master
> ------------------------------------------------------------------------------------------------------------------------------------------------------
>
>                 Key: SPARK-24404
>                 URL: https://issues.apache.org/jira/browse/SPARK-24404
>             Project: Spark
>          Issue Type: Bug
>          Components: Structured Streaming
>    Affects Versions: 2.3.0
>            Reporter: Liangchang Zhu
>            Priority: Major
>
> In CP mode,  based on PR #21353 #21332 #21293 and the latest master ContinuousQueuedDataReader.next() will be invoked by ContinuousDataSourceRDD.compute to return UnsafeRow. When currentEntry polled from ArrayBlockingQueue is a EpochMarker, ContinuousQueuedDataReader will send `ReportPartitionOffset` message to epochCoordinator with currentEpoch of EpochTracker. The currentEpoch is a ThreadLocal variable, but now no place invoke `incrementCurrentEpoch` to increase currentEpoch in its thread, so `getCurrentEpoch` will return `None` all the time(because currentEpoch is -1). This will cause exception when invoke `None.get`. At the same time, in order to make the `ReportPartitionOffset` have correct semantics, we need increase currentEpoch before send this message 



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscribe@spark.apache.org
For additional commands, e-mail: issues-help@spark.apache.org