You are viewing a plain text version of this content. The canonical link for it is here.
Posted to reviews@spark.apache.org by GitBox <gi...@apache.org> on 2020/07/12 21:09:25 UTC

[GitHub] [spark] dongjoon-hyun commented on pull request #29075: [SPARK-32284][SQL] Avoid expanding too many CNF predicates in partition pruning

dongjoon-hyun commented on pull request #29075:
URL: https://github.com/apache/spark/pull/29075#issuecomment-657275435


   Thank you for updating, @gengliangwang . Shall we adjust this test case name accordingly together?
   ```
   test("SPARK-32284: Avoid pushing down too many predicates in partition pruning") {
   ```
   
   BTW, in the test case, since `20` looks reasonably a small number in the Spark world. Could you use more functional word to describe the change? For example, this PR is not limiting based on the number of predicate like `10 is possible, but 20 is not allowed`. Apache Spark still will hit the HMS issue when we have a long long SQL query with `too many predicates` after this PR. So, this PR doesn't fix `Avoid pushing down too many predicates in partition pruning`. Instead, this PR looks like mitigating the regression due to the previous improvement PR.
   ```
   We should push down the convertible original queries as they are, instead of converting all predicates into CNF
   We can skip grouping expressions so that we can stop the CNF conversion when the predicates becoming too long.
   ```


----------------------------------------------------------------
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
users@infra.apache.org



---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org