You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@flink.apache.org by "jocean.shi (Jira)" <ji...@apache.org> on 2021/10/27 09:11:00 UTC

[jira] [Commented] (FLINK-24014) Filters pushed through ChangelogNormalize cause incorrect results

    [ https://issues.apache.org/jira/browse/FLINK-24014?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17434748#comment-17434748 ] 

jocean.shi commented on FLINK-24014:
------------------------------------

Hi [~airblader] I don't understand this case, the pre record is +I[key1, a=9], I think the next recore will be -D[key1, a=9], not the -D[key1, a=10].

if the next record is -D[key1, a=10] it will be error even not pushdown.

> Filters pushed through ChangelogNormalize cause incorrect results
> -----------------------------------------------------------------
>
>                 Key: FLINK-24014
>                 URL: https://issues.apache.org/jira/browse/FLINK-24014
>             Project: Flink
>          Issue Type: Bug
>          Components: Table SQL / Planner
>    Affects Versions: 1.13.2
>            Reporter: Ingo Bürk
>            Assignee: Ingo Bürk
>            Priority: Major
>
> If a source implements SupportsFilterPushDown, all filters of a query get pushed through to the source, including past a ChangelogNormalize.
> However, this can cause incorrect results as pushing non-PK filters is incorrect. For example, consider a filter a < 10 and records +I[key1, a=9], -D[key1, a=10].
> A strategy to fix it could be a physical rule which deoptimizes the filter pushdown, however in this case we should investigate more how strict we need to be. For now we could also simply document this in the SupportsFilterPushDown interface and warn implementations of this possibility.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)