You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hive.apache.org by na...@apache.org on 2012/07/21 05:38:14 UTC
svn commit: r1364037 - in /hive/trunk/ql/src:
java/org/apache/hadoop/hive/ql/parse/SemanticAnalyzer.java
test/queries/clientpositive/mergejoins.q
test/results/clientpositive/mergejoins.q.out
Author: namit
Date: Sat Jul 21 03:38:14 2012
New Revision: 1364037
URL: http://svn.apache.org/viewvc?rev=1364037&view=rev
Log:
HIVE-3070 Filter on outer join condition removed while merging join tree
(Navis via namit)
Modified:
hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/parse/SemanticAnalyzer.java
hive/trunk/ql/src/test/queries/clientpositive/mergejoins.q
hive/trunk/ql/src/test/results/clientpositive/mergejoins.q.out
Modified: hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/parse/SemanticAnalyzer.java
URL: http://svn.apache.org/viewvc/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/parse/SemanticAnalyzer.java?rev=1364037&r1=1364036&r2=1364037&view=diff
==============================================================================
--- hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/parse/SemanticAnalyzer.java (original)
+++ hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/parse/SemanticAnalyzer.java Sat Jul 21 03:38:14 2012
@@ -5615,6 +5615,11 @@ public class SemanticAnalyzer extends Ba
filters.add(node.getFilters().get(i + 1));
}
+ if (node.getFilters().get(0).size() != 0) {
+ ArrayList<ASTNode> filterPos = filters.get(pos);
+ filterPos.addAll(node.getFilters().get(0));
+ }
+
ArrayList<ArrayList<ASTNode>> filter = target.getFiltersForPushing();
for (int i = 0; i < nodeRightAliases.length; i++) {
filter.add(node.getFiltersForPushing().get(i + 1));
Modified: hive/trunk/ql/src/test/queries/clientpositive/mergejoins.q
URL: http://svn.apache.org/viewvc/hive/trunk/ql/src/test/queries/clientpositive/mergejoins.q?rev=1364037&r1=1364036&r2=1364037&view=diff
==============================================================================
--- hive/trunk/ql/src/test/queries/clientpositive/mergejoins.q (original)
+++ hive/trunk/ql/src/test/queries/clientpositive/mergejoins.q Sat Jul 21 03:38:14 2012
@@ -4,4 +4,7 @@ create table c (val1 int, val2 int);
create table d (val1 int, val2 int);
create table e (val1 int, val2 int);
-explain select * from a join b on a.val1=b.val1 join c on a.val1=c.val1 join d on a.val1=d.val1 join e on a.val2=e.val2;
\ No newline at end of file
+explain select * from a join b on a.val1=b.val1 join c on a.val1=c.val1 join d on a.val1=d.val1 join e on a.val2=e.val2;
+
+--HIVE-3070 filter on outer join condition removed while merging join tree
+explain select * from src a join src b on a.key=b.key left outer join src c on b.key=c.key and b.key<10;
Modified: hive/trunk/ql/src/test/results/clientpositive/mergejoins.q.out
URL: http://svn.apache.org/viewvc/hive/trunk/ql/src/test/results/clientpositive/mergejoins.q.out?rev=1364037&r1=1364036&r2=1364037&view=diff
==============================================================================
--- hive/trunk/ql/src/test/results/clientpositive/mergejoins.q.out (original)
+++ hive/trunk/ql/src/test/results/clientpositive/mergejoins.q.out Sat Jul 21 03:38:14 2012
@@ -218,3 +218,113 @@ STAGE PLANS:
limit: -1
+PREHOOK: query: --HIVE-3070 filter on outer join condition removed while merging join tree
+explain select * from src a join src b on a.key=b.key left outer join src c on b.key=c.key and b.key<10
+PREHOOK: type: QUERY
+POSTHOOK: query: --HIVE-3070 filter on outer join condition removed while merging join tree
+explain select * from src a join src b on a.key=b.key left outer join src c on b.key=c.key and b.key<10
+POSTHOOK: type: QUERY
+ABSTRACT SYNTAX TREE:
+ (TOK_QUERY (TOK_FROM (TOK_LEFTOUTERJOIN (TOK_JOIN (TOK_TABREF (TOK_TABNAME src) a) (TOK_TABREF (TOK_TABNAME src) b) (= (. (TOK_TABLE_OR_COL a) key) (. (TOK_TABLE_OR_COL b) key))) (TOK_TABREF (TOK_TABNAME src) c) (and (= (. (TOK_TABLE_OR_COL b) key) (. (TOK_TABLE_OR_COL c) key)) (< (. (TOK_TABLE_OR_COL b) key) 10)))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR TOK_ALLCOLREF))))
+
+STAGE DEPENDENCIES:
+ Stage-1 is a root stage
+ Stage-0 is a root stage
+
+STAGE PLANS:
+ Stage: Stage-1
+ Map Reduce
+ Alias -> Map Operator Tree:
+ a
+ TableScan
+ alias: a
+ Reduce Output Operator
+ key expressions:
+ expr: key
+ type: string
+ sort order: +
+ Map-reduce partition columns:
+ expr: key
+ type: string
+ tag: 0
+ value expressions:
+ expr: key
+ type: string
+ expr: value
+ type: string
+ b
+ TableScan
+ alias: b
+ Reduce Output Operator
+ key expressions:
+ expr: key
+ type: string
+ sort order: +
+ Map-reduce partition columns:
+ expr: key
+ type: string
+ tag: 1
+ value expressions:
+ expr: key
+ type: string
+ expr: value
+ type: string
+ c
+ TableScan
+ alias: c
+ Reduce Output Operator
+ key expressions:
+ expr: key
+ type: string
+ sort order: +
+ Map-reduce partition columns:
+ expr: key
+ type: string
+ tag: 2
+ value expressions:
+ expr: key
+ type: string
+ expr: value
+ type: string
+ Reduce Operator Tree:
+ Join Operator
+ condition map:
+ Inner Join 0 to 1
+ Left Outer Join1 to 2
+ condition expressions:
+ 0 {VALUE._col0} {VALUE._col1}
+ 1 {VALUE._col0} {VALUE._col1}
+ 2 {VALUE._col0} {VALUE._col1}
+ filter predicates:
+ 0
+ 1 {(VALUE._col0 < 10.0)}
+ 2
+ handleSkewJoin: false
+ outputColumnNames: _col0, _col1, _col4, _col5, _col8, _col9
+ Select Operator
+ expressions:
+ expr: _col0
+ type: string
+ expr: _col1
+ type: string
+ expr: _col4
+ type: string
+ expr: _col5
+ type: string
+ expr: _col8
+ type: string
+ expr: _col9
+ type: string
+ outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5
+ File Output Operator
+ compressed: false
+ GlobalTableId: 0
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+
+ Stage: Stage-0
+ Fetch Operator
+ limit: -1
+
+