You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hive.apache.org by ha...@apache.org on 2013/03/29 23:59:41 UTC

svn commit: r1462670 [12/14] - in /hive/branches/ptf-windowing: ./ common/src/java/org/apache/hadoop/hive/conf/ conf/ contrib/src/java/org/apache/hadoop/hive/contrib/serde2/ contrib/src/java/org/apache/hadoop/hive/contrib/serde2/s3/ data/conf/ data/fil...

Modified: hive/branches/ptf-windowing/ql/src/test/results/clientpositive/join_1to1.q.out
URL: http://svn.apache.org/viewvc/hive/branches/ptf-windowing/ql/src/test/results/clientpositive/join_1to1.q.out?rev=1462670&r1=1462669&r2=1462670&view=diff
==============================================================================
--- hive/branches/ptf-windowing/ql/src/test/results/clientpositive/join_1to1.q.out (original)
+++ hive/branches/ptf-windowing/ql/src/test/results/clientpositive/join_1to1.q.out Fri Mar 29 22:58:43 2013
@@ -701,17 +701,15 @@ POSTHOOK: Input: default@join_1to1_2
 NULL	NULL	NULL	NULL	NULL	66
 NULL	NULL	NULL	NULL	10050	66
 NULL	NULL	NULL	10	10010	66
+NULL	NULL	NULL	25	10025	66
 NULL	NULL	NULL	30	10030	88
 NULL	NULL	NULL	35	10035	88
 NULL	NULL	NULL	40	10040	88
-NULL	NULL	NULL	40	10040	88
-NULL	NULL	NULL	50	10050	88
-NULL	NULL	NULL	50	10050	88
 NULL	NULL	NULL	50	10050	88
 NULL	NULL	NULL	70	10040	88
 NULL	NULL	NULL	70	10040	88
-NULL	NULL	NULL	70	10040	88
-NULL	NULL	NULL	70	10040	88
+NULL	NULL	NULL	80	10040	66
+NULL	NULL	NULL	80	10040	66
 NULL	NULL	66	NULL	NULL	NULL
 NULL	10050	66	NULL	NULL	NULL
 5	10005	66	5	10005	66
@@ -720,29 +718,18 @@ NULL	10050	66	NULL	NULL	NULL
 25	10025	88	NULL	NULL	NULL
 30	10030	66	NULL	NULL	NULL
 35	10035	88	NULL	NULL	NULL
-40	10040	66	NULL	NULL	NULL
 40	10040	66	40	10040	66
 40	10040	88	NULL	NULL	NULL
-40	10040	88	NULL	NULL	NULL
-50	10050	66	NULL	NULL	NULL
 50	10050	66	50	10050	66
 50	10050	66	50	10050	66
 50	10050	88	NULL	NULL	NULL
 50	10050	88	NULL	NULL	NULL
-50	10050	88	NULL	NULL	NULL
-50	10050	88	NULL	NULL	NULL
-50	10050	88	NULL	NULL	NULL
-50	10050	88	NULL	NULL	NULL
 60	10040	66	60	10040	66
 60	10040	66	60	10040	66
 60	10040	66	60	10040	66
 60	10040	66	60	10040	66
 70	10040	66	NULL	NULL	NULL
 70	10040	66	NULL	NULL	NULL
-70	10040	66	NULL	NULL	NULL
-70	10040	66	NULL	NULL	NULL
-80	10040	88	NULL	NULL	NULL
-80	10040	88	NULL	NULL	NULL
 80	10040	88	NULL	NULL	NULL
 80	10040	88	NULL	NULL	NULL
 PREHOOK: query: SELECT * FROM join_1to1_1 a full outer join join_1to1_2 b on a.key1 = b.key1 and a.key2 = b.key2 ORDER BY a.key1 ASC, a.key2 ASC, a.value ASC, b.key1 ASC, b.key2 ASC, b.value ASC
@@ -804,17 +791,15 @@ POSTHOOK: Input: default@join_1to1_2
 NULL	NULL	NULL	NULL	NULL	66
 NULL	NULL	NULL	NULL	10050	66
 NULL	NULL	NULL	10	10010	66
+NULL	NULL	NULL	25	10025	66
 NULL	NULL	NULL	30	10030	88
 NULL	NULL	NULL	35	10035	88
 NULL	NULL	NULL	40	10040	88
-NULL	NULL	NULL	40	10040	88
-NULL	NULL	NULL	50	10050	88
 NULL	NULL	NULL	50	10050	88
-NULL	NULL	NULL	50	10050	88
-NULL	NULL	NULL	70	10040	88
-NULL	NULL	NULL	70	10040	88
 NULL	NULL	NULL	70	10040	88
 NULL	NULL	NULL	70	10040	88
+NULL	NULL	NULL	80	10040	66
+NULL	NULL	NULL	80	10040	66
 NULL	NULL	66	NULL	NULL	NULL
 NULL	10050	66	NULL	NULL	NULL
 5	10005	66	5	10005	66
@@ -823,29 +808,18 @@ NULL	10050	66	NULL	NULL	NULL
 25	10025	88	NULL	NULL	NULL
 30	10030	66	NULL	NULL	NULL
 35	10035	88	NULL	NULL	NULL
-40	10040	66	NULL	NULL	NULL
 40	10040	66	40	10040	66
 40	10040	88	NULL	NULL	NULL
-40	10040	88	NULL	NULL	NULL
-50	10050	66	NULL	NULL	NULL
 50	10050	66	50	10050	66
 50	10050	66	50	10050	66
 50	10050	88	NULL	NULL	NULL
 50	10050	88	NULL	NULL	NULL
-50	10050	88	NULL	NULL	NULL
-50	10050	88	NULL	NULL	NULL
-50	10050	88	NULL	NULL	NULL
-50	10050	88	NULL	NULL	NULL
 60	10040	66	60	10040	66
 60	10040	66	60	10040	66
 60	10040	66	60	10040	66
 60	10040	66	60	10040	66
 70	10040	66	NULL	NULL	NULL
 70	10040	66	NULL	NULL	NULL
-70	10040	66	NULL	NULL	NULL
-70	10040	66	NULL	NULL	NULL
-80	10040	88	NULL	NULL	NULL
-80	10040	88	NULL	NULL	NULL
 80	10040	88	NULL	NULL	NULL
 80	10040	88	NULL	NULL	NULL
 PREHOOK: query: SELECT * FROM join_1to1_1 a join join_1to1_2 b on a.key1 = b.key1 ORDER BY a.key1 ASC, a.key2 ASC, a.value ASC, b.key1 ASC, b.key2 ASC, b.value ASC
@@ -947,17 +921,15 @@ POSTHOOK: Input: default@join_1to1_2
 NULL	NULL	NULL	NULL	NULL	66
 NULL	NULL	NULL	NULL	10050	66
 NULL	NULL	NULL	10	10010	66
+NULL	NULL	NULL	25	10025	66
 NULL	NULL	NULL	30	10030	88
 NULL	NULL	NULL	35	10035	88
 NULL	NULL	NULL	40	10040	88
-NULL	NULL	NULL	40	10040	88
-NULL	NULL	NULL	50	10050	88
-NULL	NULL	NULL	50	10050	88
 NULL	NULL	NULL	50	10050	88
 NULL	NULL	NULL	70	10040	88
 NULL	NULL	NULL	70	10040	88
-NULL	NULL	NULL	70	10040	88
-NULL	NULL	NULL	70	10040	88
+NULL	NULL	NULL	80	10040	66
+NULL	NULL	NULL	80	10040	66
 NULL	NULL	66	NULL	NULL	NULL
 NULL	10050	66	NULL	NULL	NULL
 5	10005	66	5	10005	66
@@ -966,29 +938,20 @@ NULL	10050	66	NULL	NULL	NULL
 25	10025	88	NULL	NULL	NULL
 30	10030	66	NULL	NULL	NULL
 35	10035	88	NULL	NULL	NULL
-40	10040	66	NULL	NULL	NULL
 40	10040	66	40	10040	66
 40	10040	88	NULL	NULL	NULL
-40	10040	88	NULL	NULL	NULL
-50	10050	66	NULL	NULL	NULL
 50	10050	66	50	10050	66
 50	10050	66	50	10050	66
 50	10050	88	NULL	NULL	NULL
 50	10050	88	NULL	NULL	NULL
 50	10050	88	NULL	NULL	NULL
 50	10050	88	NULL	NULL	NULL
-50	10050	88	NULL	NULL	NULL
-50	10050	88	NULL	NULL	NULL
 60	10040	66	60	10040	66
 60	10040	66	60	10040	66
 60	10040	66	60	10040	66
 60	10040	66	60	10040	66
 70	10040	66	NULL	NULL	NULL
 70	10040	66	NULL	NULL	NULL
-70	10040	66	NULL	NULL	NULL
-70	10040	66	NULL	NULL	NULL
-80	10040	88	NULL	NULL	NULL
-80	10040	88	NULL	NULL	NULL
 80	10040	88	NULL	NULL	NULL
 80	10040	88	NULL	NULL	NULL
 PREHOOK: query: SELECT * FROM join_1to1_1 a full outer join join_1to1_2 b on a.key1 = b.key1 and a.key2 = b.key2 ORDER BY a.key1 ASC, a.key2 ASC, a.value ASC, b.key1 ASC, b.key2 ASC, b.value ASC
@@ -1050,17 +1013,15 @@ POSTHOOK: Input: default@join_1to1_2
 NULL	NULL	NULL	NULL	NULL	66
 NULL	NULL	NULL	NULL	10050	66
 NULL	NULL	NULL	10	10010	66
+NULL	NULL	NULL	25	10025	66
 NULL	NULL	NULL	30	10030	88
 NULL	NULL	NULL	35	10035	88
 NULL	NULL	NULL	40	10040	88
-NULL	NULL	NULL	40	10040	88
-NULL	NULL	NULL	50	10050	88
-NULL	NULL	NULL	50	10050	88
 NULL	NULL	NULL	50	10050	88
 NULL	NULL	NULL	70	10040	88
 NULL	NULL	NULL	70	10040	88
-NULL	NULL	NULL	70	10040	88
-NULL	NULL	NULL	70	10040	88
+NULL	NULL	NULL	80	10040	66
+NULL	NULL	NULL	80	10040	66
 NULL	NULL	66	NULL	NULL	NULL
 NULL	10050	66	NULL	NULL	NULL
 5	10005	66	5	10005	66
@@ -1069,29 +1030,20 @@ NULL	10050	66	NULL	NULL	NULL
 25	10025	88	NULL	NULL	NULL
 30	10030	66	NULL	NULL	NULL
 35	10035	88	NULL	NULL	NULL
-40	10040	66	NULL	NULL	NULL
 40	10040	66	40	10040	66
 40	10040	88	NULL	NULL	NULL
-40	10040	88	NULL	NULL	NULL
-50	10050	66	NULL	NULL	NULL
 50	10050	66	50	10050	66
 50	10050	66	50	10050	66
 50	10050	88	NULL	NULL	NULL
 50	10050	88	NULL	NULL	NULL
 50	10050	88	NULL	NULL	NULL
 50	10050	88	NULL	NULL	NULL
-50	10050	88	NULL	NULL	NULL
-50	10050	88	NULL	NULL	NULL
 60	10040	66	60	10040	66
 60	10040	66	60	10040	66
 60	10040	66	60	10040	66
 60	10040	66	60	10040	66
 70	10040	66	NULL	NULL	NULL
 70	10040	66	NULL	NULL	NULL
-70	10040	66	NULL	NULL	NULL
-70	10040	66	NULL	NULL	NULL
-80	10040	88	NULL	NULL	NULL
-80	10040	88	NULL	NULL	NULL
 80	10040	88	NULL	NULL	NULL
 80	10040	88	NULL	NULL	NULL
 PREHOOK: query: SELECT * FROM join_1to1_1 a join join_1to1_2 b on a.key1 = b.key1 ORDER BY a.key1 ASC, a.key2 ASC, a.value ASC, b.key1 ASC, b.key2 ASC, b.value ASC
@@ -1193,17 +1145,15 @@ POSTHOOK: Input: default@join_1to1_2
 NULL	NULL	NULL	NULL	NULL	66
 NULL	NULL	NULL	NULL	10050	66
 NULL	NULL	NULL	10	10010	66
+NULL	NULL	NULL	25	10025	66
 NULL	NULL	NULL	30	10030	88
 NULL	NULL	NULL	35	10035	88
 NULL	NULL	NULL	40	10040	88
-NULL	NULL	NULL	40	10040	88
-NULL	NULL	NULL	50	10050	88
 NULL	NULL	NULL	50	10050	88
-NULL	NULL	NULL	50	10050	88
-NULL	NULL	NULL	70	10040	88
-NULL	NULL	NULL	70	10040	88
 NULL	NULL	NULL	70	10040	88
 NULL	NULL	NULL	70	10040	88
+NULL	NULL	NULL	80	10040	66
+NULL	NULL	NULL	80	10040	66
 NULL	NULL	66	NULL	NULL	NULL
 NULL	10050	66	NULL	NULL	NULL
 5	10005	66	5	10005	66
@@ -1216,7 +1166,6 @@ NULL	10050	66	NULL	NULL	NULL
 40	10040	66	40	10040	66
 40	10040	88	NULL	NULL	NULL
 40	10040	88	NULL	NULL	NULL
-50	10050	66	NULL	NULL	NULL
 50	10050	66	50	10050	66
 50	10050	66	50	10050	66
 50	10050	88	NULL	NULL	NULL
@@ -1296,17 +1245,15 @@ POSTHOOK: Input: default@join_1to1_2
 NULL	NULL	NULL	NULL	NULL	66
 NULL	NULL	NULL	NULL	10050	66
 NULL	NULL	NULL	10	10010	66
+NULL	NULL	NULL	25	10025	66
 NULL	NULL	NULL	30	10030	88
 NULL	NULL	NULL	35	10035	88
 NULL	NULL	NULL	40	10040	88
-NULL	NULL	NULL	40	10040	88
 NULL	NULL	NULL	50	10050	88
-NULL	NULL	NULL	50	10050	88
-NULL	NULL	NULL	50	10050	88
-NULL	NULL	NULL	70	10040	88
-NULL	NULL	NULL	70	10040	88
 NULL	NULL	NULL	70	10040	88
 NULL	NULL	NULL	70	10040	88
+NULL	NULL	NULL	80	10040	66
+NULL	NULL	NULL	80	10040	66
 NULL	NULL	66	NULL	NULL	NULL
 NULL	10050	66	NULL	NULL	NULL
 5	10005	66	5	10005	66
@@ -1319,7 +1266,6 @@ NULL	10050	66	NULL	NULL	NULL
 40	10040	66	40	10040	66
 40	10040	88	NULL	NULL	NULL
 40	10040	88	NULL	NULL	NULL
-50	10050	66	NULL	NULL	NULL
 50	10050	66	50	10050	66
 50	10050	66	50	10050	66
 50	10050	88	NULL	NULL	NULL

Modified: hive/branches/ptf-windowing/ql/src/test/results/clientpositive/join_filters.q.out
URL: http://svn.apache.org/viewvc/hive/branches/ptf-windowing/ql/src/test/results/clientpositive/join_filters.q.out?rev=1462670&r1=1462669&r2=1462670&view=diff
==============================================================================
--- hive/branches/ptf-windowing/ql/src/test/results/clientpositive/join_filters.q.out (original)
+++ hive/branches/ptf-windowing/ql/src/test/results/clientpositive/join_filters.q.out Fri Mar 29 22:58:43 2013
@@ -51,32 +51,11 @@ POSTHOOK: type: QUERY
 POSTHOOK: Input: default@myinput1
 #### A masked pattern was here ####
 NULL	NULL	NULL	40
-NULL	NULL	NULL	40
-NULL	NULL	NULL	40
-NULL	NULL	NULL	40
-NULL	NULL	12	35
-NULL	NULL	12	35
 NULL	NULL	12	35
-NULL	NULL	12	35
-NULL	NULL	48	NULL
-NULL	NULL	48	NULL
-NULL	NULL	48	NULL
 NULL	NULL	48	NULL
 NULL	40	NULL	NULL
-NULL	40	NULL	NULL
-NULL	40	NULL	NULL
-NULL	40	NULL	NULL
-12	35	NULL	NULL
-12	35	NULL	NULL
 12	35	NULL	NULL
-12	35	NULL	NULL
-48	NULL	NULL	NULL
-48	NULL	NULL	NULL
-48	NULL	NULL	NULL
 48	NULL	NULL	NULL
-100	100	NULL	NULL
-100	100	NULL	NULL
-100	100	NULL	NULL
 100	100	100	100
 PREHOOK: query: SELECT * FROM myinput1 a JOIN myinput1 b ON a.key = b.value AND a.key > 40 AND a.value > 50 AND a.key = a.value AND b.key > 40 AND b.value > 50 AND b.key = b.value ORDER BY a.key ASC, a.value ASC, b.key ASC, b.value ASC
 PREHOOK: type: QUERY

Modified: hive/branches/ptf-windowing/ql/src/test/results/clientpositive/join_filters_overlap.q.out
URL: http://svn.apache.org/viewvc/hive/branches/ptf-windowing/ql/src/test/results/clientpositive/join_filters_overlap.q.out?rev=1462670&r1=1462669&r2=1462670&view=diff
==============================================================================
--- hive/branches/ptf-windowing/ql/src/test/results/clientpositive/join_filters_overlap.q.out (original)
+++ hive/branches/ptf-windowing/ql/src/test/results/clientpositive/join_filters_overlap.q.out Fri Mar 29 22:58:43 2013
@@ -847,19 +847,9 @@ POSTHOOK: Input: default@a
 #### A masked pattern was here ####
 100	40	NULL	NULL	NULL	NULL	100	40
 NULL	NULL	100	40	NULL	NULL	NULL	NULL
-100	40	NULL	NULL	NULL	NULL	100	40
-100	40	NULL	NULL	NULL	NULL	100	40
 NULL	NULL	100	60	100	60	NULL	NULL
-100	50	NULL	NULL	NULL	NULL	NULL	NULL
-NULL	NULL	100	40	NULL	NULL	NULL	NULL
 100	50	100	50	NULL	NULL	NULL	NULL
-100	50	NULL	NULL	NULL	NULL	NULL	NULL
-NULL	NULL	100	60	100	60	NULL	NULL
 100	60	NULL	NULL	NULL	NULL	NULL	NULL
-NULL	NULL	100	40	NULL	NULL	NULL	NULL
-100	60	NULL	NULL	NULL	NULL	NULL	NULL
-100	60	NULL	NULL	NULL	NULL	NULL	NULL
-NULL	NULL	100	60	100	60	NULL	NULL
 PREHOOK: query: -- triple overlap on a
 explain extended select * from a left outer join a b on (a.key=b.key AND a.value=50 AND b.value=50) left outer join a c on (a.key=c.key AND a.value=60 AND c.value=60) left outer join a d on (a.key=d.key AND a.value=40 AND d.value=40)
 PREHOOK: type: QUERY

Modified: hive/branches/ptf-windowing/ql/src/test/results/clientpositive/join_reorder4.q.out
URL: http://svn.apache.org/viewvc/hive/branches/ptf-windowing/ql/src/test/results/clientpositive/join_reorder4.q.out?rev=1462670&r1=1462669&r2=1462670&view=diff
==============================================================================
--- hive/branches/ptf-windowing/ql/src/test/results/clientpositive/join_reorder4.q.out (original)
+++ hive/branches/ptf-windowing/ql/src/test/results/clientpositive/join_reorder4.q.out Fri Mar 29 22:58:43 2013
@@ -39,105 +39,12 @@ ABSTRACT SYNTAX TREE:
   (TOK_QUERY (TOK_FROM (TOK_JOIN (TOK_JOIN (TOK_TABREF (TOK_TABNAME T1) a) (TOK_TABREF (TOK_TABNAME T2) b) (= (. (TOK_TABLE_OR_COL a) key1) (. (TOK_TABLE_OR_COL b) key2))) (TOK_TABREF (TOK_TABNAME T3) c) (= (. (TOK_TABLE_OR_COL a) key1) (. (TOK_TABLE_OR_COL c) key3)))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_HINTLIST (TOK_HINT TOK_STREAMTABLE (TOK_HINTARGLIST a))) (TOK_SELEXPR (TOK_ALLCOLREF (TOK_TABNAME a))) (TOK_SELEXPR (TOK_ALLCOLREF (TOK_TABNAME b))) (TOK_SELEXPR (TOK_ALLCOLREF (TOK_TABNAME c))))))
 
 STAGE DEPENDENCIES:
-  Stage-7 is a root stage , consists of Stage-8, Stage-9, Stage-10, Stage-1
-  Stage-8 has a backup stage: Stage-1
-  Stage-4 depends on stages: Stage-8
-  Stage-9 has a backup stage: Stage-1
-  Stage-5 depends on stages: Stage-9
-  Stage-10 has a backup stage: Stage-1
-  Stage-6 depends on stages: Stage-10
-  Stage-1
+  Stage-5 is a root stage
+  Stage-4 depends on stages: Stage-5
   Stage-0 is a root stage
 
 STAGE PLANS:
-  Stage: Stage-7
-    Conditional Operator
-
-  Stage: Stage-8
-    Map Reduce Local Work
-      Alias -> Map Local Tables:
-        b 
-          Fetch Operator
-            limit: -1
-        c 
-          Fetch Operator
-            limit: -1
-      Alias -> Map Local Operator Tree:
-        b 
-          TableScan
-            alias: b
-            HashTable Sink Operator
-              condition expressions:
-                0 {key1} {val1}
-                1 {key2} {val2}
-                2 {key3} {val3}
-              handleSkewJoin: false
-              keys:
-                0 [Column[key1]]
-                1 [Column[key2]]
-                2 [Column[key3]]
-              Position of Big Table: 0
-        c 
-          TableScan
-            alias: c
-            HashTable Sink Operator
-              condition expressions:
-                0 {key1} {val1}
-                1 {key2} {val2}
-                2 {key3} {val3}
-              handleSkewJoin: false
-              keys:
-                0 [Column[key1]]
-                1 [Column[key2]]
-                2 [Column[key3]]
-              Position of Big Table: 0
-
-  Stage: Stage-4
-    Map Reduce
-      Alias -> Map Operator Tree:
-        a 
-          TableScan
-            alias: a
-            Map Join Operator
-              condition map:
-                   Inner Join 0 to 1
-                   Inner Join 0 to 2
-              condition expressions:
-                0 {key1} {val1}
-                1 {key2} {val2}
-                2 {key3} {val3}
-              handleSkewJoin: false
-              keys:
-                0 [Column[key1]]
-                1 [Column[key2]]
-                2 [Column[key3]]
-              outputColumnNames: _col0, _col1, _col4, _col5, _col8, _col9
-              Position of Big Table: 0
-              Select Operator
-                expressions:
-                      expr: _col0
-                      type: string
-                      expr: _col1
-                      type: string
-                      expr: _col4
-                      type: string
-                      expr: _col5
-                      type: string
-                      expr: _col8
-                      type: string
-                      expr: _col9
-                      type: string
-                outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5
-                File Output Operator
-                  compressed: false
-                  GlobalTableId: 0
-                  table:
-                      input format: org.apache.hadoop.mapred.TextInputFormat
-                      output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-      Local Work:
-        Map Reduce Local Work
-
-  Stage: Stage-9
+  Stage: Stage-5
     Map Reduce Local Work
       Alias -> Map Local Tables:
         a 
@@ -176,7 +83,7 @@ STAGE PLANS:
                 2 [Column[key3]]
               Position of Big Table: 1
 
-  Stage: Stage-5
+  Stage: Stage-4
     Map Reduce
       Alias -> Map Operator Tree:
         b 
@@ -221,177 +128,6 @@ STAGE PLANS:
       Local Work:
         Map Reduce Local Work
 
-  Stage: Stage-10
-    Map Reduce Local Work
-      Alias -> Map Local Tables:
-        a 
-          Fetch Operator
-            limit: -1
-        b 
-          Fetch Operator
-            limit: -1
-      Alias -> Map Local Operator Tree:
-        a 
-          TableScan
-            alias: a
-            HashTable Sink Operator
-              condition expressions:
-                0 {key1} {val1}
-                1 {key2} {val2}
-                2 {key3} {val3}
-              handleSkewJoin: false
-              keys:
-                0 [Column[key1]]
-                1 [Column[key2]]
-                2 [Column[key3]]
-              Position of Big Table: 2
-        b 
-          TableScan
-            alias: b
-            HashTable Sink Operator
-              condition expressions:
-                0 {key1} {val1}
-                1 {key2} {val2}
-                2 {key3} {val3}
-              handleSkewJoin: false
-              keys:
-                0 [Column[key1]]
-                1 [Column[key2]]
-                2 [Column[key3]]
-              Position of Big Table: 2
-
-  Stage: Stage-6
-    Map Reduce
-      Alias -> Map Operator Tree:
-        c 
-          TableScan
-            alias: c
-            Map Join Operator
-              condition map:
-                   Inner Join 0 to 1
-                   Inner Join 0 to 2
-              condition expressions:
-                0 {key1} {val1}
-                1 {key2} {val2}
-                2 {key3} {val3}
-              handleSkewJoin: false
-              keys:
-                0 [Column[key1]]
-                1 [Column[key2]]
-                2 [Column[key3]]
-              outputColumnNames: _col0, _col1, _col4, _col5, _col8, _col9
-              Position of Big Table: 2
-              Select Operator
-                expressions:
-                      expr: _col0
-                      type: string
-                      expr: _col1
-                      type: string
-                      expr: _col4
-                      type: string
-                      expr: _col5
-                      type: string
-                      expr: _col8
-                      type: string
-                      expr: _col9
-                      type: string
-                outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5
-                File Output Operator
-                  compressed: false
-                  GlobalTableId: 0
-                  table:
-                      input format: org.apache.hadoop.mapred.TextInputFormat
-                      output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-      Local Work:
-        Map Reduce Local Work
-
-  Stage: Stage-1
-    Map Reduce
-      Alias -> Map Operator Tree:
-        a 
-          TableScan
-            alias: a
-            Reduce Output Operator
-              key expressions:
-                    expr: key1
-                    type: string
-              sort order: +
-              Map-reduce partition columns:
-                    expr: key1
-                    type: string
-              tag: 2
-              value expressions:
-                    expr: key1
-                    type: string
-                    expr: val1
-                    type: string
-        b 
-          TableScan
-            alias: b
-            Reduce Output Operator
-              key expressions:
-                    expr: key2
-                    type: string
-              sort order: +
-              Map-reduce partition columns:
-                    expr: key2
-                    type: string
-              tag: 1
-              value expressions:
-                    expr: key2
-                    type: string
-                    expr: val2
-                    type: string
-        c 
-          TableScan
-            alias: c
-            Reduce Output Operator
-              key expressions:
-                    expr: key3
-                    type: string
-              sort order: +
-              Map-reduce partition columns:
-                    expr: key3
-                    type: string
-              tag: 0
-              value expressions:
-                    expr: key3
-                    type: string
-                    expr: val3
-                    type: string
-      Reduce Operator Tree:
-        Join Operator
-          condition map:
-               Inner Join 0 to 1
-               Inner Join 0 to 2
-          condition expressions:
-            0 {VALUE._col0} {VALUE._col1}
-            1 {VALUE._col0} {VALUE._col1}
-            2 {VALUE._col0} {VALUE._col1}
-          handleSkewJoin: false
-          outputColumnNames: _col0, _col1, _col4, _col5, _col8, _col9
-          Select Operator
-            expressions:
-                  expr: _col0
-                  type: string
-                  expr: _col1
-                  type: string
-                  expr: _col4
-                  type: string
-                  expr: _col5
-                  type: string
-                  expr: _col8
-                  type: string
-                  expr: _col9
-                  type: string
-            outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5
-            File Output Operator
-              compressed: false
-              GlobalTableId: 0
-              table:
-                  input format: org.apache.hadoop.mapred.TextInputFormat
-                  output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-
   Stage: Stage-0
     Fetch Operator
       limit: -1
@@ -418,33 +154,24 @@ ABSTRACT SYNTAX TREE:
   (TOK_QUERY (TOK_FROM (TOK_JOIN (TOK_JOIN (TOK_TABREF (TOK_TABNAME T1) a) (TOK_TABREF (TOK_TABNAME T2) b) (= (. (TOK_TABLE_OR_COL a) key1) (. (TOK_TABLE_OR_COL b) key2))) (TOK_TABREF (TOK_TABNAME T3) c) (= (. (TOK_TABLE_OR_COL a) key1) (. (TOK_TABLE_OR_COL c) key3)))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_HINTLIST (TOK_HINT TOK_STREAMTABLE (TOK_HINTARGLIST b))) (TOK_SELEXPR (TOK_ALLCOLREF (TOK_TABNAME a))) (TOK_SELEXPR (TOK_ALLCOLREF (TOK_TABNAME b))) (TOK_SELEXPR (TOK_ALLCOLREF (TOK_TABNAME c))))))
 
 STAGE DEPENDENCIES:
-  Stage-7 is a root stage , consists of Stage-8, Stage-9, Stage-10, Stage-1
-  Stage-8 has a backup stage: Stage-1
-  Stage-4 depends on stages: Stage-8
-  Stage-9 has a backup stage: Stage-1
-  Stage-5 depends on stages: Stage-9
-  Stage-10 has a backup stage: Stage-1
-  Stage-6 depends on stages: Stage-10
-  Stage-1
+  Stage-5 is a root stage
+  Stage-4 depends on stages: Stage-5
   Stage-0 is a root stage
 
 STAGE PLANS:
-  Stage: Stage-7
-    Conditional Operator
-
-  Stage: Stage-8
+  Stage: Stage-5
     Map Reduce Local Work
       Alias -> Map Local Tables:
-        b 
+        a 
           Fetch Operator
             limit: -1
         c 
           Fetch Operator
             limit: -1
       Alias -> Map Local Operator Tree:
-        b 
+        a 
           TableScan
-            alias: b
+            alias: a
             HashTable Sink Operator
               condition expressions:
                 0 {key1} {val1}
@@ -455,7 +182,7 @@ STAGE PLANS:
                 0 [Column[key1]]
                 1 [Column[key2]]
                 2 [Column[key3]]
-              Position of Big Table: 0
+              Position of Big Table: 1
         c 
           TableScan
             alias: c
@@ -469,14 +196,14 @@ STAGE PLANS:
                 0 [Column[key1]]
                 1 [Column[key2]]
                 2 [Column[key3]]
-              Position of Big Table: 0
+              Position of Big Table: 1
 
   Stage: Stage-4
     Map Reduce
       Alias -> Map Operator Tree:
-        a 
+        b 
           TableScan
-            alias: a
+            alias: b
             Map Join Operator
               condition map:
                    Inner Join 0 to 1
@@ -491,7 +218,7 @@ STAGE PLANS:
                 1 [Column[key2]]
                 2 [Column[key3]]
               outputColumnNames: _col0, _col1, _col4, _col5, _col8, _col9
-              Position of Big Table: 0
+              Position of Big Table: 1
               Select Operator
                 expressions:
                       expr: _col0
@@ -516,7 +243,38 @@ STAGE PLANS:
       Local Work:
         Map Reduce Local Work
 
-  Stage: Stage-9
+  Stage: Stage-0
+    Fetch Operator
+      limit: -1
+
+
+PREHOOK: query: select /*+ STREAMTABLE(b) */ a.*, b.*, c.* from T1 a join T2 b on a.key1=b.key2 join T3 c on a.key1=c.key3
+PREHOOK: type: QUERY
+PREHOOK: Input: default@t1
+PREHOOK: Input: default@t2
+PREHOOK: Input: default@t3
+#### A masked pattern was here ####
+POSTHOOK: query: select /*+ STREAMTABLE(b) */ a.*, b.*, c.* from T1 a join T2 b on a.key1=b.key2 join T3 c on a.key1=c.key3
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@t1
+POSTHOOK: Input: default@t2
+POSTHOOK: Input: default@t3
+#### A masked pattern was here ####
+2	12	2	22	2	12
+PREHOOK: query: explain select /*+ STREAMTABLE(c) */ a.*, b.*, c.* from T1 a join T2 b on a.key1=b.key2 join T3 c on a.key1=c.key3
+PREHOOK: type: QUERY
+POSTHOOK: query: explain select /*+ STREAMTABLE(c) */ a.*, b.*, c.* from T1 a join T2 b on a.key1=b.key2 join T3 c on a.key1=c.key3
+POSTHOOK: type: QUERY
+ABSTRACT SYNTAX TREE:
+  (TOK_QUERY (TOK_FROM (TOK_JOIN (TOK_JOIN (TOK_TABREF (TOK_TABNAME T1) a) (TOK_TABREF (TOK_TABNAME T2) b) (= (. (TOK_TABLE_OR_COL a) key1) (. (TOK_TABLE_OR_COL b) key2))) (TOK_TABREF (TOK_TABNAME T3) c) (= (. (TOK_TABLE_OR_COL a) key1) (. (TOK_TABLE_OR_COL c) key3)))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_HINTLIST (TOK_HINT TOK_STREAMTABLE (TOK_HINTARGLIST c))) (TOK_SELEXPR (TOK_ALLCOLREF (TOK_TABNAME a))) (TOK_SELEXPR (TOK_ALLCOLREF (TOK_TABNAME b))) (TOK_SELEXPR (TOK_ALLCOLREF (TOK_TABNAME c))))))
+
+STAGE DEPENDENCIES:
+  Stage-5 is a root stage
+  Stage-4 depends on stages: Stage-5
+  Stage-0 is a root stage
+
+STAGE PLANS:
+  Stage: Stage-5
     Map Reduce Local Work
       Alias -> Map Local Tables:
         a 
@@ -555,7 +313,7 @@ STAGE PLANS:
                 2 [Column[key3]]
               Position of Big Table: 1
 
-  Stage: Stage-5
+  Stage: Stage-4
     Map Reduce
       Alias -> Map Operator Tree:
         b 
@@ -600,556 +358,6 @@ STAGE PLANS:
       Local Work:
         Map Reduce Local Work
 
-  Stage: Stage-10
-    Map Reduce Local Work
-      Alias -> Map Local Tables:
-        a 
-          Fetch Operator
-            limit: -1
-        b 
-          Fetch Operator
-            limit: -1
-      Alias -> Map Local Operator Tree:
-        a 
-          TableScan
-            alias: a
-            HashTable Sink Operator
-              condition expressions:
-                0 {key1} {val1}
-                1 {key2} {val2}
-                2 {key3} {val3}
-              handleSkewJoin: false
-              keys:
-                0 [Column[key1]]
-                1 [Column[key2]]
-                2 [Column[key3]]
-              Position of Big Table: 2
-        b 
-          TableScan
-            alias: b
-            HashTable Sink Operator
-              condition expressions:
-                0 {key1} {val1}
-                1 {key2} {val2}
-                2 {key3} {val3}
-              handleSkewJoin: false
-              keys:
-                0 [Column[key1]]
-                1 [Column[key2]]
-                2 [Column[key3]]
-              Position of Big Table: 2
-
-  Stage: Stage-6
-    Map Reduce
-      Alias -> Map Operator Tree:
-        c 
-          TableScan
-            alias: c
-            Map Join Operator
-              condition map:
-                   Inner Join 0 to 1
-                   Inner Join 0 to 2
-              condition expressions:
-                0 {key1} {val1}
-                1 {key2} {val2}
-                2 {key3} {val3}
-              handleSkewJoin: false
-              keys:
-                0 [Column[key1]]
-                1 [Column[key2]]
-                2 [Column[key3]]
-              outputColumnNames: _col0, _col1, _col4, _col5, _col8, _col9
-              Position of Big Table: 2
-              Select Operator
-                expressions:
-                      expr: _col0
-                      type: string
-                      expr: _col1
-                      type: string
-                      expr: _col4
-                      type: string
-                      expr: _col5
-                      type: string
-                      expr: _col8
-                      type: string
-                      expr: _col9
-                      type: string
-                outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5
-                File Output Operator
-                  compressed: false
-                  GlobalTableId: 0
-                  table:
-                      input format: org.apache.hadoop.mapred.TextInputFormat
-                      output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-      Local Work:
-        Map Reduce Local Work
-
-  Stage: Stage-1
-    Map Reduce
-      Alias -> Map Operator Tree:
-        a 
-          TableScan
-            alias: a
-            Reduce Output Operator
-              key expressions:
-                    expr: key1
-                    type: string
-              sort order: +
-              Map-reduce partition columns:
-                    expr: key1
-                    type: string
-              tag: 0
-              value expressions:
-                    expr: key1
-                    type: string
-                    expr: val1
-                    type: string
-        b 
-          TableScan
-            alias: b
-            Reduce Output Operator
-              key expressions:
-                    expr: key2
-                    type: string
-              sort order: +
-              Map-reduce partition columns:
-                    expr: key2
-                    type: string
-              tag: 2
-              value expressions:
-                    expr: key2
-                    type: string
-                    expr: val2
-                    type: string
-        c 
-          TableScan
-            alias: c
-            Reduce Output Operator
-              key expressions:
-                    expr: key3
-                    type: string
-              sort order: +
-              Map-reduce partition columns:
-                    expr: key3
-                    type: string
-              tag: 1
-              value expressions:
-                    expr: key3
-                    type: string
-                    expr: val3
-                    type: string
-      Reduce Operator Tree:
-        Join Operator
-          condition map:
-               Inner Join 0 to 1
-               Inner Join 0 to 2
-          condition expressions:
-            0 {VALUE._col0} {VALUE._col1}
-            1 {VALUE._col0} {VALUE._col1}
-            2 {VALUE._col0} {VALUE._col1}
-          handleSkewJoin: false
-          outputColumnNames: _col0, _col1, _col4, _col5, _col8, _col9
-          Select Operator
-            expressions:
-                  expr: _col0
-                  type: string
-                  expr: _col1
-                  type: string
-                  expr: _col4
-                  type: string
-                  expr: _col5
-                  type: string
-                  expr: _col8
-                  type: string
-                  expr: _col9
-                  type: string
-            outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5
-            File Output Operator
-              compressed: false
-              GlobalTableId: 0
-              table:
-                  input format: org.apache.hadoop.mapred.TextInputFormat
-                  output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-
-  Stage: Stage-0
-    Fetch Operator
-      limit: -1
-
-
-PREHOOK: query: select /*+ STREAMTABLE(b) */ a.*, b.*, c.* from T1 a join T2 b on a.key1=b.key2 join T3 c on a.key1=c.key3
-PREHOOK: type: QUERY
-PREHOOK: Input: default@t1
-PREHOOK: Input: default@t2
-PREHOOK: Input: default@t3
-#### A masked pattern was here ####
-POSTHOOK: query: select /*+ STREAMTABLE(b) */ a.*, b.*, c.* from T1 a join T2 b on a.key1=b.key2 join T3 c on a.key1=c.key3
-POSTHOOK: type: QUERY
-POSTHOOK: Input: default@t1
-POSTHOOK: Input: default@t2
-POSTHOOK: Input: default@t3
-#### A masked pattern was here ####
-2	12	2	22	2	12
-PREHOOK: query: explain select /*+ STREAMTABLE(c) */ a.*, b.*, c.* from T1 a join T2 b on a.key1=b.key2 join T3 c on a.key1=c.key3
-PREHOOK: type: QUERY
-POSTHOOK: query: explain select /*+ STREAMTABLE(c) */ a.*, b.*, c.* from T1 a join T2 b on a.key1=b.key2 join T3 c on a.key1=c.key3
-POSTHOOK: type: QUERY
-ABSTRACT SYNTAX TREE:
-  (TOK_QUERY (TOK_FROM (TOK_JOIN (TOK_JOIN (TOK_TABREF (TOK_TABNAME T1) a) (TOK_TABREF (TOK_TABNAME T2) b) (= (. (TOK_TABLE_OR_COL a) key1) (. (TOK_TABLE_OR_COL b) key2))) (TOK_TABREF (TOK_TABNAME T3) c) (= (. (TOK_TABLE_OR_COL a) key1) (. (TOK_TABLE_OR_COL c) key3)))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_HINTLIST (TOK_HINT TOK_STREAMTABLE (TOK_HINTARGLIST c))) (TOK_SELEXPR (TOK_ALLCOLREF (TOK_TABNAME a))) (TOK_SELEXPR (TOK_ALLCOLREF (TOK_TABNAME b))) (TOK_SELEXPR (TOK_ALLCOLREF (TOK_TABNAME c))))))
-
-STAGE DEPENDENCIES:
-  Stage-7 is a root stage , consists of Stage-8, Stage-9, Stage-10, Stage-1
-  Stage-8 has a backup stage: Stage-1
-  Stage-4 depends on stages: Stage-8
-  Stage-9 has a backup stage: Stage-1
-  Stage-5 depends on stages: Stage-9
-  Stage-10 has a backup stage: Stage-1
-  Stage-6 depends on stages: Stage-10
-  Stage-1
-  Stage-0 is a root stage
-
-STAGE PLANS:
-  Stage: Stage-7
-    Conditional Operator
-
-  Stage: Stage-8
-    Map Reduce Local Work
-      Alias -> Map Local Tables:
-        b 
-          Fetch Operator
-            limit: -1
-        c 
-          Fetch Operator
-            limit: -1
-      Alias -> Map Local Operator Tree:
-        b 
-          TableScan
-            alias: b
-            HashTable Sink Operator
-              condition expressions:
-                0 {key1} {val1}
-                1 {key2} {val2}
-                2 {key3} {val3}
-              handleSkewJoin: false
-              keys:
-                0 [Column[key1]]
-                1 [Column[key2]]
-                2 [Column[key3]]
-              Position of Big Table: 0
-        c 
-          TableScan
-            alias: c
-            HashTable Sink Operator
-              condition expressions:
-                0 {key1} {val1}
-                1 {key2} {val2}
-                2 {key3} {val3}
-              handleSkewJoin: false
-              keys:
-                0 [Column[key1]]
-                1 [Column[key2]]
-                2 [Column[key3]]
-              Position of Big Table: 0
-
-  Stage: Stage-4
-    Map Reduce
-      Alias -> Map Operator Tree:
-        a 
-          TableScan
-            alias: a
-            Map Join Operator
-              condition map:
-                   Inner Join 0 to 1
-                   Inner Join 0 to 2
-              condition expressions:
-                0 {key1} {val1}
-                1 {key2} {val2}
-                2 {key3} {val3}
-              handleSkewJoin: false
-              keys:
-                0 [Column[key1]]
-                1 [Column[key2]]
-                2 [Column[key3]]
-              outputColumnNames: _col0, _col1, _col4, _col5, _col8, _col9
-              Position of Big Table: 0
-              Select Operator
-                expressions:
-                      expr: _col0
-                      type: string
-                      expr: _col1
-                      type: string
-                      expr: _col4
-                      type: string
-                      expr: _col5
-                      type: string
-                      expr: _col8
-                      type: string
-                      expr: _col9
-                      type: string
-                outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5
-                File Output Operator
-                  compressed: false
-                  GlobalTableId: 0
-                  table:
-                      input format: org.apache.hadoop.mapred.TextInputFormat
-                      output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-      Local Work:
-        Map Reduce Local Work
-
-  Stage: Stage-9
-    Map Reduce Local Work
-      Alias -> Map Local Tables:
-        a 
-          Fetch Operator
-            limit: -1
-        c 
-          Fetch Operator
-            limit: -1
-      Alias -> Map Local Operator Tree:
-        a 
-          TableScan
-            alias: a
-            HashTable Sink Operator
-              condition expressions:
-                0 {key1} {val1}
-                1 {key2} {val2}
-                2 {key3} {val3}
-              handleSkewJoin: false
-              keys:
-                0 [Column[key1]]
-                1 [Column[key2]]
-                2 [Column[key3]]
-              Position of Big Table: 1
-        c 
-          TableScan
-            alias: c
-            HashTable Sink Operator
-              condition expressions:
-                0 {key1} {val1}
-                1 {key2} {val2}
-                2 {key3} {val3}
-              handleSkewJoin: false
-              keys:
-                0 [Column[key1]]
-                1 [Column[key2]]
-                2 [Column[key3]]
-              Position of Big Table: 1
-
-  Stage: Stage-5
-    Map Reduce
-      Alias -> Map Operator Tree:
-        b 
-          TableScan
-            alias: b
-            Map Join Operator
-              condition map:
-                   Inner Join 0 to 1
-                   Inner Join 0 to 2
-              condition expressions:
-                0 {key1} {val1}
-                1 {key2} {val2}
-                2 {key3} {val3}
-              handleSkewJoin: false
-              keys:
-                0 [Column[key1]]
-                1 [Column[key2]]
-                2 [Column[key3]]
-              outputColumnNames: _col0, _col1, _col4, _col5, _col8, _col9
-              Position of Big Table: 1
-              Select Operator
-                expressions:
-                      expr: _col0
-                      type: string
-                      expr: _col1
-                      type: string
-                      expr: _col4
-                      type: string
-                      expr: _col5
-                      type: string
-                      expr: _col8
-                      type: string
-                      expr: _col9
-                      type: string
-                outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5
-                File Output Operator
-                  compressed: false
-                  GlobalTableId: 0
-                  table:
-                      input format: org.apache.hadoop.mapred.TextInputFormat
-                      output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-      Local Work:
-        Map Reduce Local Work
-
-  Stage: Stage-10
-    Map Reduce Local Work
-      Alias -> Map Local Tables:
-        a 
-          Fetch Operator
-            limit: -1
-        b 
-          Fetch Operator
-            limit: -1
-      Alias -> Map Local Operator Tree:
-        a 
-          TableScan
-            alias: a
-            HashTable Sink Operator
-              condition expressions:
-                0 {key1} {val1}
-                1 {key2} {val2}
-                2 {key3} {val3}
-              handleSkewJoin: false
-              keys:
-                0 [Column[key1]]
-                1 [Column[key2]]
-                2 [Column[key3]]
-              Position of Big Table: 2
-        b 
-          TableScan
-            alias: b
-            HashTable Sink Operator
-              condition expressions:
-                0 {key1} {val1}
-                1 {key2} {val2}
-                2 {key3} {val3}
-              handleSkewJoin: false
-              keys:
-                0 [Column[key1]]
-                1 [Column[key2]]
-                2 [Column[key3]]
-              Position of Big Table: 2
-
-  Stage: Stage-6
-    Map Reduce
-      Alias -> Map Operator Tree:
-        c 
-          TableScan
-            alias: c
-            Map Join Operator
-              condition map:
-                   Inner Join 0 to 1
-                   Inner Join 0 to 2
-              condition expressions:
-                0 {key1} {val1}
-                1 {key2} {val2}
-                2 {key3} {val3}
-              handleSkewJoin: false
-              keys:
-                0 [Column[key1]]
-                1 [Column[key2]]
-                2 [Column[key3]]
-              outputColumnNames: _col0, _col1, _col4, _col5, _col8, _col9
-              Position of Big Table: 2
-              Select Operator
-                expressions:
-                      expr: _col0
-                      type: string
-                      expr: _col1
-                      type: string
-                      expr: _col4
-                      type: string
-                      expr: _col5
-                      type: string
-                      expr: _col8
-                      type: string
-                      expr: _col9
-                      type: string
-                outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5
-                File Output Operator
-                  compressed: false
-                  GlobalTableId: 0
-                  table:
-                      input format: org.apache.hadoop.mapred.TextInputFormat
-                      output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-      Local Work:
-        Map Reduce Local Work
-
-  Stage: Stage-1
-    Map Reduce
-      Alias -> Map Operator Tree:
-        a 
-          TableScan
-            alias: a
-            Reduce Output Operator
-              key expressions:
-                    expr: key1
-                    type: string
-              sort order: +
-              Map-reduce partition columns:
-                    expr: key1
-                    type: string
-              tag: 0
-              value expressions:
-                    expr: key1
-                    type: string
-                    expr: val1
-                    type: string
-        b 
-          TableScan
-            alias: b
-            Reduce Output Operator
-              key expressions:
-                    expr: key2
-                    type: string
-              sort order: +
-              Map-reduce partition columns:
-                    expr: key2
-                    type: string
-              tag: 1
-              value expressions:
-                    expr: key2
-                    type: string
-                    expr: val2
-                    type: string
-        c 
-          TableScan
-            alias: c
-            Reduce Output Operator
-              key expressions:
-                    expr: key3
-                    type: string
-              sort order: +
-              Map-reduce partition columns:
-                    expr: key3
-                    type: string
-              tag: 2
-              value expressions:
-                    expr: key3
-                    type: string
-                    expr: val3
-                    type: string
-      Reduce Operator Tree:
-        Join Operator
-          condition map:
-               Inner Join 0 to 1
-               Inner Join 0 to 2
-          condition expressions:
-            0 {VALUE._col0} {VALUE._col1}
-            1 {VALUE._col0} {VALUE._col1}
-            2 {VALUE._col0} {VALUE._col1}
-          handleSkewJoin: false
-          outputColumnNames: _col0, _col1, _col4, _col5, _col8, _col9
-          Select Operator
-            expressions:
-                  expr: _col0
-                  type: string
-                  expr: _col1
-                  type: string
-                  expr: _col4
-                  type: string
-                  expr: _col5
-                  type: string
-                  expr: _col8
-                  type: string
-                  expr: _col9
-                  type: string
-            outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5
-            File Output Operator
-              compressed: false
-              GlobalTableId: 0
-              table:
-                  input format: org.apache.hadoop.mapred.TextInputFormat
-                  output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
-
   Stage: Stage-0
     Fetch Operator
       limit: -1

Modified: hive/branches/ptf-windowing/ql/src/test/results/clientpositive/load_dyn_part1.q.out
URL: http://svn.apache.org/viewvc/hive/branches/ptf-windowing/ql/src/test/results/clientpositive/load_dyn_part1.q.out?rev=1462670&r1=1462669&r2=1462670&view=diff
==============================================================================
--- hive/branches/ptf-windowing/ql/src/test/results/clientpositive/load_dyn_part1.q.out (original)
+++ hive/branches/ptf-windowing/ql/src/test/results/clientpositive/load_dyn_part1.q.out Fri Mar 29 22:58:43 2013
@@ -293,11 +293,13 @@ ds=2008-12-31/hr=11
 ds=2008-12-31/hr=12
 PREHOOK: query: select * from nzhang_part1 where ds is not null and hr is not null
 PREHOOK: type: QUERY
+PREHOOK: Input: default@nzhang_part1
 PREHOOK: Input: default@nzhang_part1@ds=2008-04-08/hr=11
 PREHOOK: Input: default@nzhang_part1@ds=2008-04-08/hr=12
 #### A masked pattern was here ####
 POSTHOOK: query: select * from nzhang_part1 where ds is not null and hr is not null
 POSTHOOK: type: QUERY
+POSTHOOK: Input: default@nzhang_part1
 POSTHOOK: Input: default@nzhang_part1@ds=2008-04-08/hr=11
 POSTHOOK: Input: default@nzhang_part1@ds=2008-04-08/hr=12
 #### A masked pattern was here ####
@@ -1311,11 +1313,13 @@ POSTHOOK: Lineage: nzhang_part2 PARTITIO
 97	val_97	2008-04-08	12
 PREHOOK: query: select * from nzhang_part2 where ds is not null and hr is not null
 PREHOOK: type: QUERY
+PREHOOK: Input: default@nzhang_part2
 PREHOOK: Input: default@nzhang_part2@ds=2008-12-31/hr=11
 PREHOOK: Input: default@nzhang_part2@ds=2008-12-31/hr=12
 #### A masked pattern was here ####
 POSTHOOK: query: select * from nzhang_part2 where ds is not null and hr is not null
 POSTHOOK: type: QUERY
+POSTHOOK: Input: default@nzhang_part2
 POSTHOOK: Input: default@nzhang_part2@ds=2008-12-31/hr=11
 POSTHOOK: Input: default@nzhang_part2@ds=2008-12-31/hr=12
 #### A masked pattern was here ####

Modified: hive/branches/ptf-windowing/ql/src/test/results/clientpositive/load_dyn_part10.q.out
URL: http://svn.apache.org/viewvc/hive/branches/ptf-windowing/ql/src/test/results/clientpositive/load_dyn_part10.q.out?rev=1462670&r1=1462669&r2=1462670&view=diff
==============================================================================
--- hive/branches/ptf-windowing/ql/src/test/results/clientpositive/load_dyn_part10.q.out (original)
+++ hive/branches/ptf-windowing/ql/src/test/results/clientpositive/load_dyn_part10.q.out Fri Mar 29 22:58:43 2013
@@ -118,11 +118,13 @@ ds=2008-12-31/hr=11
 ds=2008-12-31/hr=12
 PREHOOK: query: select * from nzhang_part10 where ds is not null and hr is not null
 PREHOOK: type: QUERY
+PREHOOK: Input: default@nzhang_part10
 PREHOOK: Input: default@nzhang_part10@ds=2008-12-31/hr=11
 PREHOOK: Input: default@nzhang_part10@ds=2008-12-31/hr=12
 #### A masked pattern was here ####
 POSTHOOK: query: select * from nzhang_part10 where ds is not null and hr is not null
 POSTHOOK: type: QUERY
+POSTHOOK: Input: default@nzhang_part10
 POSTHOOK: Input: default@nzhang_part10@ds=2008-12-31/hr=11
 POSTHOOK: Input: default@nzhang_part10@ds=2008-12-31/hr=12
 #### A masked pattern was here ####

Modified: hive/branches/ptf-windowing/ql/src/test/results/clientpositive/load_dyn_part11.q.out
URL: http://svn.apache.org/viewvc/hive/branches/ptf-windowing/ql/src/test/results/clientpositive/load_dyn_part11.q.out?rev=1462670&r1=1462669&r2=1462670&view=diff
==============================================================================
--- hive/branches/ptf-windowing/ql/src/test/results/clientpositive/load_dyn_part11.q.out (original)
+++ hive/branches/ptf-windowing/ql/src/test/results/clientpositive/load_dyn_part11.q.out Fri Mar 29 22:58:43 2013
@@ -52,10 +52,12 @@ POSTHOOK: Lineage: nzhang_part PARTITION
 POSTHOOK: Lineage: nzhang_part PARTITION(ds=2010-03-03,hr=12).value SIMPLE [(srcpart)srcpart.FieldSchema(name:value, type:string, comment:default), ]
 PREHOOK: query: select * from nzhang_part where ds = '2010-03-03' and hr = '11'
 PREHOOK: type: QUERY
+PREHOOK: Input: default@nzhang_part
 PREHOOK: Input: default@nzhang_part@ds=2010-03-03/hr=11
 #### A masked pattern was here ####
 POSTHOOK: query: select * from nzhang_part where ds = '2010-03-03' and hr = '11'
 POSTHOOK: type: QUERY
+POSTHOOK: Input: default@nzhang_part
 POSTHOOK: Input: default@nzhang_part@ds=2010-03-03/hr=11
 #### A masked pattern was here ####
 POSTHOOK: Lineage: nzhang_part PARTITION(ds=2010-03-03,hr=11).key SIMPLE [(srcpart)srcpart.FieldSchema(name:key, type:string, comment:default), ]
@@ -1064,10 +1066,12 @@ POSTHOOK: Lineage: nzhang_part PARTITION
 97	val_97	2010-03-03	11
 PREHOOK: query: select * from nzhang_part where ds = '2010-03-03' and hr = '12'
 PREHOOK: type: QUERY
+PREHOOK: Input: default@nzhang_part
 PREHOOK: Input: default@nzhang_part@ds=2010-03-03/hr=12
 #### A masked pattern was here ####
 POSTHOOK: query: select * from nzhang_part where ds = '2010-03-03' and hr = '12'
 POSTHOOK: type: QUERY
+POSTHOOK: Input: default@nzhang_part
 POSTHOOK: Input: default@nzhang_part@ds=2010-03-03/hr=12
 #### A masked pattern was here ####
 POSTHOOK: Lineage: nzhang_part PARTITION(ds=2010-03-03,hr=11).key SIMPLE [(srcpart)srcpart.FieldSchema(name:key, type:string, comment:default), ]

Modified: hive/branches/ptf-windowing/ql/src/test/results/clientpositive/load_dyn_part12.q.out
URL: http://svn.apache.org/viewvc/hive/branches/ptf-windowing/ql/src/test/results/clientpositive/load_dyn_part12.q.out?rev=1462670&r1=1462669&r2=1462670&view=diff
==============================================================================
--- hive/branches/ptf-windowing/ql/src/test/results/clientpositive/load_dyn_part12.q.out (original)
+++ hive/branches/ptf-windowing/ql/src/test/results/clientpositive/load_dyn_part12.q.out Fri Mar 29 22:58:43 2013
@@ -62,11 +62,13 @@ ds=2010-03-03/hr=22
 ds=2010-03-03/hr=24
 PREHOOK: query: select * from nzhang_part12 where ds is not null and hr is not null
 PREHOOK: type: QUERY
+PREHOOK: Input: default@nzhang_part12
 PREHOOK: Input: default@nzhang_part12@ds=2010-03-03/hr=22
 PREHOOK: Input: default@nzhang_part12@ds=2010-03-03/hr=24
 #### A masked pattern was here ####
 POSTHOOK: query: select * from nzhang_part12 where ds is not null and hr is not null
 POSTHOOK: type: QUERY
+POSTHOOK: Input: default@nzhang_part12
 POSTHOOK: Input: default@nzhang_part12@ds=2010-03-03/hr=22
 POSTHOOK: Input: default@nzhang_part12@ds=2010-03-03/hr=24
 #### A masked pattern was here ####

Modified: hive/branches/ptf-windowing/ql/src/test/results/clientpositive/load_dyn_part13.q.out
URL: http://svn.apache.org/viewvc/hive/branches/ptf-windowing/ql/src/test/results/clientpositive/load_dyn_part13.q.out?rev=1462670&r1=1462669&r2=1462670&view=diff
==============================================================================
--- hive/branches/ptf-windowing/ql/src/test/results/clientpositive/load_dyn_part13.q.out (original)
+++ hive/branches/ptf-windowing/ql/src/test/results/clientpositive/load_dyn_part13.q.out Fri Mar 29 22:58:43 2013
@@ -190,11 +190,13 @@ ds=2010-03-03/hr=22
 ds=2010-03-03/hr=33
 PREHOOK: query: select * from nzhang_part13 where ds is not null and hr is not null
 PREHOOK: type: QUERY
+PREHOOK: Input: default@nzhang_part13
 PREHOOK: Input: default@nzhang_part13@ds=2010-03-03/hr=22
 PREHOOK: Input: default@nzhang_part13@ds=2010-03-03/hr=33
 #### A masked pattern was here ####
 POSTHOOK: query: select * from nzhang_part13 where ds is not null and hr is not null
 POSTHOOK: type: QUERY
+POSTHOOK: Input: default@nzhang_part13
 POSTHOOK: Input: default@nzhang_part13@ds=2010-03-03/hr=22
 POSTHOOK: Input: default@nzhang_part13@ds=2010-03-03/hr=33
 #### A masked pattern was here ####

Modified: hive/branches/ptf-windowing/ql/src/test/results/clientpositive/load_dyn_part3.q.out
URL: http://svn.apache.org/viewvc/hive/branches/ptf-windowing/ql/src/test/results/clientpositive/load_dyn_part3.q.out?rev=1462670&r1=1462669&r2=1462670&view=diff
==============================================================================
--- hive/branches/ptf-windowing/ql/src/test/results/clientpositive/load_dyn_part3.q.out (original)
+++ hive/branches/ptf-windowing/ql/src/test/results/clientpositive/load_dyn_part3.q.out Fri Mar 29 22:58:43 2013
@@ -116,6 +116,7 @@ POSTHOOK: Lineage: nzhang_part3 PARTITIO
 POSTHOOK: Lineage: nzhang_part3 PARTITION(ds=2008-04-09,hr=12).value SIMPLE [(srcpart)srcpart.FieldSchema(name:value, type:string, comment:default), ]
 PREHOOK: query: select * from nzhang_part3 where ds is not null and hr is not null
 PREHOOK: type: QUERY
+PREHOOK: Input: default@nzhang_part3
 PREHOOK: Input: default@nzhang_part3@ds=2008-04-08/hr=11
 PREHOOK: Input: default@nzhang_part3@ds=2008-04-08/hr=12
 PREHOOK: Input: default@nzhang_part3@ds=2008-04-09/hr=11
@@ -123,6 +124,7 @@ PREHOOK: Input: default@nzhang_part3@ds=
 #### A masked pattern was here ####
 POSTHOOK: query: select * from nzhang_part3 where ds is not null and hr is not null
 POSTHOOK: type: QUERY
+POSTHOOK: Input: default@nzhang_part3
 POSTHOOK: Input: default@nzhang_part3@ds=2008-04-08/hr=11
 POSTHOOK: Input: default@nzhang_part3@ds=2008-04-08/hr=12
 POSTHOOK: Input: default@nzhang_part3@ds=2008-04-09/hr=11

Modified: hive/branches/ptf-windowing/ql/src/test/results/clientpositive/load_dyn_part4.q.out
URL: http://svn.apache.org/viewvc/hive/branches/ptf-windowing/ql/src/test/results/clientpositive/load_dyn_part4.q.out?rev=1462670&r1=1462669&r2=1462670&view=diff
==============================================================================
--- hive/branches/ptf-windowing/ql/src/test/results/clientpositive/load_dyn_part4.q.out (original)
+++ hive/branches/ptf-windowing/ql/src/test/results/clientpositive/load_dyn_part4.q.out Fri Mar 29 22:58:43 2013
@@ -149,12 +149,14 @@ ds=2008-04-09/hr=11
 ds=2008-04-09/hr=12
 PREHOOK: query: select * from nzhang_part4 where ds='2008-04-08' and hr is not null
 PREHOOK: type: QUERY
+PREHOOK: Input: default@nzhang_part4
 PREHOOK: Input: default@nzhang_part4@ds=2008-04-08/hr=11
 PREHOOK: Input: default@nzhang_part4@ds=2008-04-08/hr=12
 PREHOOK: Input: default@nzhang_part4@ds=2008-04-08/hr=existing_value
 #### A masked pattern was here ####
 POSTHOOK: query: select * from nzhang_part4 where ds='2008-04-08' and hr is not null
 POSTHOOK: type: QUERY
+POSTHOOK: Input: default@nzhang_part4
 POSTHOOK: Input: default@nzhang_part4@ds=2008-04-08/hr=11
 POSTHOOK: Input: default@nzhang_part4@ds=2008-04-08/hr=12
 POSTHOOK: Input: default@nzhang_part4@ds=2008-04-08/hr=existing_value
@@ -1671,6 +1673,7 @@ POSTHOOK: Lineage: nzhang_part4 PARTITIO
 97	val_97	2008-04-08	existing_value
 PREHOOK: query: select * from nzhang_part4 where ds is not null and hr is not null
 PREHOOK: type: QUERY
+PREHOOK: Input: default@nzhang_part4
 PREHOOK: Input: default@nzhang_part4@ds=2008-04-08/hr=11
 PREHOOK: Input: default@nzhang_part4@ds=2008-04-08/hr=12
 PREHOOK: Input: default@nzhang_part4@ds=2008-04-08/hr=existing_value
@@ -1679,6 +1682,7 @@ PREHOOK: Input: default@nzhang_part4@ds=
 #### A masked pattern was here ####
 POSTHOOK: query: select * from nzhang_part4 where ds is not null and hr is not null
 POSTHOOK: type: QUERY
+POSTHOOK: Input: default@nzhang_part4
 POSTHOOK: Input: default@nzhang_part4@ds=2008-04-08/hr=11
 POSTHOOK: Input: default@nzhang_part4@ds=2008-04-08/hr=12
 POSTHOOK: Input: default@nzhang_part4@ds=2008-04-08/hr=existing_value

Modified: hive/branches/ptf-windowing/ql/src/test/results/clientpositive/load_dyn_part5.q.out
URL: http://svn.apache.org/viewvc/hive/branches/ptf-windowing/ql/src/test/results/clientpositive/load_dyn_part5.q.out?rev=1462670&r1=1462669&r2=1462670&view=diff
==============================================================================
--- hive/branches/ptf-windowing/ql/src/test/results/clientpositive/load_dyn_part5.q.out (original)
+++ hive/branches/ptf-windowing/ql/src/test/results/clientpositive/load_dyn_part5.q.out Fri Mar 29 22:58:43 2013
@@ -1320,10 +1320,12 @@ value=val_97
 value=val_98
 PREHOOK: query: select * from nzhang_part5 where value='val_0'
 PREHOOK: type: QUERY
+PREHOOK: Input: default@nzhang_part5
 PREHOOK: Input: default@nzhang_part5@value=val_0
 #### A masked pattern was here ####
 POSTHOOK: query: select * from nzhang_part5 where value='val_0'
 POSTHOOK: type: QUERY
+POSTHOOK: Input: default@nzhang_part5
 POSTHOOK: Input: default@nzhang_part5@value=val_0
 #### A masked pattern was here ####
 POSTHOOK: Lineage: nzhang_part5 PARTITION(value=val_0).key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ]
@@ -1640,10 +1642,12 @@ POSTHOOK: Lineage: nzhang_part5 PARTITIO
 0	val_0
 PREHOOK: query: select * from nzhang_part5 where value='val_2'
 PREHOOK: type: QUERY
+PREHOOK: Input: default@nzhang_part5
 PREHOOK: Input: default@nzhang_part5@value=val_2
 #### A masked pattern was here ####
 POSTHOOK: query: select * from nzhang_part5 where value='val_2'
 POSTHOOK: type: QUERY
+POSTHOOK: Input: default@nzhang_part5
 POSTHOOK: Input: default@nzhang_part5@value=val_2
 #### A masked pattern was here ####
 POSTHOOK: Lineage: nzhang_part5 PARTITION(value=val_0).key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ]

Modified: hive/branches/ptf-windowing/ql/src/test/results/clientpositive/load_dyn_part6.q.out
URL: http://svn.apache.org/viewvc/hive/branches/ptf-windowing/ql/src/test/results/clientpositive/load_dyn_part6.q.out?rev=1462670&r1=1462669&r2=1462670&view=diff
==============================================================================
--- hive/branches/ptf-windowing/ql/src/test/results/clientpositive/load_dyn_part6.q.out (original)
+++ hive/branches/ptf-windowing/ql/src/test/results/clientpositive/load_dyn_part6.q.out Fri Mar 29 22:58:43 2013
@@ -52,10 +52,12 @@ POSTHOOK: Lineage: nzhang_part6 PARTITIO
 POSTHOOK: Lineage: nzhang_part6 PARTITION(ds=2010-03-03,hr=12).value SIMPLE [(srcpart)srcpart.FieldSchema(name:value, type:string, comment:default), ]
 PREHOOK: query: select * from nzhang_part6 where ds = '2010-03-03' and hr = '11'
 PREHOOK: type: QUERY
+PREHOOK: Input: default@nzhang_part6
 PREHOOK: Input: default@nzhang_part6@ds=2010-03-03/hr=11
 #### A masked pattern was here ####
 POSTHOOK: query: select * from nzhang_part6 where ds = '2010-03-03' and hr = '11'
 POSTHOOK: type: QUERY
+POSTHOOK: Input: default@nzhang_part6
 POSTHOOK: Input: default@nzhang_part6@ds=2010-03-03/hr=11
 #### A masked pattern was here ####
 POSTHOOK: Lineage: nzhang_part6 PARTITION(ds=2010-03-03,hr=11).key SIMPLE [(srcpart)srcpart.FieldSchema(name:key, type:string, comment:default), ]
@@ -1064,10 +1066,12 @@ POSTHOOK: Lineage: nzhang_part6 PARTITIO
 97	val_97	2010-03-03	11
 PREHOOK: query: select * from nzhang_part6 where ds = '2010-03-03' and hr = '12'
 PREHOOK: type: QUERY
+PREHOOK: Input: default@nzhang_part6
 PREHOOK: Input: default@nzhang_part6@ds=2010-03-03/hr=12
 #### A masked pattern was here ####
 POSTHOOK: query: select * from nzhang_part6 where ds = '2010-03-03' and hr = '12'
 POSTHOOK: type: QUERY
+POSTHOOK: Input: default@nzhang_part6
 POSTHOOK: Input: default@nzhang_part6@ds=2010-03-03/hr=12
 #### A masked pattern was here ####
 POSTHOOK: Lineage: nzhang_part6 PARTITION(ds=2010-03-03,hr=11).key SIMPLE [(srcpart)srcpart.FieldSchema(name:key, type:string, comment:default), ]

Modified: hive/branches/ptf-windowing/ql/src/test/results/clientpositive/load_dyn_part7.q.out
URL: http://svn.apache.org/viewvc/hive/branches/ptf-windowing/ql/src/test/results/clientpositive/load_dyn_part7.q.out?rev=1462670&r1=1462669&r2=1462670&view=diff
==============================================================================
--- hive/branches/ptf-windowing/ql/src/test/results/clientpositive/load_dyn_part7.q.out (original)
+++ hive/branches/ptf-windowing/ql/src/test/results/clientpositive/load_dyn_part7.q.out Fri Mar 29 22:58:43 2013
@@ -50,10 +50,12 @@ POSTHOOK: Lineage: nzhang_part7 PARTITIO
 ds=2010-03-03/hr=12
 PREHOOK: query: select * from nzhang_part7 where ds is not null and hr is not null
 PREHOOK: type: QUERY
+PREHOOK: Input: default@nzhang_part7
 PREHOOK: Input: default@nzhang_part7@ds=2010-03-03/hr=12
 #### A masked pattern was here ####
 POSTHOOK: query: select * from nzhang_part7 where ds is not null and hr is not null
 POSTHOOK: type: QUERY
+POSTHOOK: Input: default@nzhang_part7
 POSTHOOK: Input: default@nzhang_part7@ds=2010-03-03/hr=12
 #### A masked pattern was here ####
 POSTHOOK: Lineage: nzhang_part7 PARTITION(ds=2010-03-03,hr=12).key SIMPLE [(srcpart)srcpart.FieldSchema(name:key, type:string, comment:default), ]

Modified: hive/branches/ptf-windowing/ql/src/test/results/clientpositive/load_dyn_part8.q.out
URL: http://svn.apache.org/viewvc/hive/branches/ptf-windowing/ql/src/test/results/clientpositive/load_dyn_part8.q.out?rev=1462670&r1=1462669&r2=1462670&view=diff
==============================================================================
--- hive/branches/ptf-windowing/ql/src/test/results/clientpositive/load_dyn_part8.q.out (original)
+++ hive/branches/ptf-windowing/ql/src/test/results/clientpositive/load_dyn_part8.q.out Fri Mar 29 22:58:43 2013
@@ -443,6 +443,7 @@ ds=2008-12-31/hr=11
 ds=2008-12-31/hr=12
 PREHOOK: query: select * from nzhang_part8 where ds is not null and hr is not null
 PREHOOK: type: QUERY
+PREHOOK: Input: default@nzhang_part8
 PREHOOK: Input: default@nzhang_part8@ds=2008-04-08/hr=11
 PREHOOK: Input: default@nzhang_part8@ds=2008-04-08/hr=12
 PREHOOK: Input: default@nzhang_part8@ds=2008-12-31/hr=11
@@ -450,6 +451,7 @@ PREHOOK: Input: default@nzhang_part8@ds=
 #### A masked pattern was here ####
 POSTHOOK: query: select * from nzhang_part8 where ds is not null and hr is not null
 POSTHOOK: type: QUERY
+POSTHOOK: Input: default@nzhang_part8
 POSTHOOK: Input: default@nzhang_part8@ds=2008-04-08/hr=11
 POSTHOOK: Input: default@nzhang_part8@ds=2008-04-08/hr=12
 POSTHOOK: Input: default@nzhang_part8@ds=2008-12-31/hr=11

Modified: hive/branches/ptf-windowing/ql/src/test/results/clientpositive/load_dyn_part9.q.out
URL: http://svn.apache.org/viewvc/hive/branches/ptf-windowing/ql/src/test/results/clientpositive/load_dyn_part9.q.out?rev=1462670&r1=1462669&r2=1462670&view=diff
==============================================================================
--- hive/branches/ptf-windowing/ql/src/test/results/clientpositive/load_dyn_part9.q.out (original)
+++ hive/branches/ptf-windowing/ql/src/test/results/clientpositive/load_dyn_part9.q.out Fri Mar 29 22:58:43 2013
@@ -120,11 +120,13 @@ ds=2008-04-08/hr=11
 ds=2008-04-08/hr=12
 PREHOOK: query: select * from nzhang_part9 where ds is not null and hr is not null
 PREHOOK: type: QUERY
+PREHOOK: Input: default@nzhang_part9
 PREHOOK: Input: default@nzhang_part9@ds=2008-04-08/hr=11
 PREHOOK: Input: default@nzhang_part9@ds=2008-04-08/hr=12
 #### A masked pattern was here ####
 POSTHOOK: query: select * from nzhang_part9 where ds is not null and hr is not null
 POSTHOOK: type: QUERY
+POSTHOOK: Input: default@nzhang_part9
 POSTHOOK: Input: default@nzhang_part9@ds=2008-04-08/hr=11
 POSTHOOK: Input: default@nzhang_part9@ds=2008-04-08/hr=12
 #### A masked pattern was here ####

Modified: hive/branches/ptf-windowing/ql/src/test/results/clientpositive/loadpart1.q.out
URL: http://svn.apache.org/viewvc/hive/branches/ptf-windowing/ql/src/test/results/clientpositive/loadpart1.q.out?rev=1462670&r1=1462669&r2=1462670&view=diff
==============================================================================
--- hive/branches/ptf-windowing/ql/src/test/results/clientpositive/loadpart1.q.out (original)
+++ hive/branches/ptf-windowing/ql/src/test/results/clientpositive/loadpart1.q.out Fri Mar 29 22:58:43 2013
@@ -25,10 +25,12 @@ POSTHOOK: Output: default@hive_test_dst@
 POSTHOOK: Lineage: hive_test_dst PARTITION(pcol1=test_part,pcol2=test_Part).col1 SIMPLE [(hive_test_src)hive_test_src.FieldSchema(name:col1, type:string, comment:null), ]
 PREHOOK: query: select * from hive_test_dst where pcol1='test_part' and pcol2='test_Part'
 PREHOOK: type: QUERY
+PREHOOK: Input: default@hive_test_dst
 PREHOOK: Input: default@hive_test_dst@pcol1=test_part/pcol2=test_Part
 #### A masked pattern was here ####
 POSTHOOK: query: select * from hive_test_dst where pcol1='test_part' and pcol2='test_Part'
 POSTHOOK: type: QUERY
+POSTHOOK: Input: default@hive_test_dst
 POSTHOOK: Input: default@hive_test_dst@pcol1=test_part/pcol2=test_Part
 #### A masked pattern was here ####
 POSTHOOK: Lineage: hive_test_dst PARTITION(pcol1=test_part,pcol2=test_Part).col1 SIMPLE [(hive_test_src)hive_test_src.FieldSchema(name:col1, type:string, comment:null), ]
@@ -50,18 +52,22 @@ POSTHOOK: Lineage: hive_test_dst PARTITI
 POSTHOOK: Lineage: hive_test_dst PARTITION(pcol1=test_part,pcol2=test_Part).col1 SIMPLE [(hive_test_src)hive_test_src.FieldSchema(name:col1, type:string, comment:null), ]
 PREHOOK: query: select * from hive_test_dst where pcol1='test_part' and pcol2='test_part'
 PREHOOK: type: QUERY
+PREHOOK: Input: default@hive_test_dst
 #### A masked pattern was here ####
 POSTHOOK: query: select * from hive_test_dst where pcol1='test_part' and pcol2='test_part'
 POSTHOOK: type: QUERY
+POSTHOOK: Input: default@hive_test_dst
 #### A masked pattern was here ####
 POSTHOOK: Lineage: hive_test_dst PARTITION(pcol1=test_part,pcol2=test_Part).col1 SIMPLE [(hive_test_src)hive_test_src.FieldSchema(name:col1, type:string, comment:null), ]
 POSTHOOK: Lineage: hive_test_dst PARTITION(pcol1=test_part,pcol2=test_Part).col1 SIMPLE [(hive_test_src)hive_test_src.FieldSchema(name:col1, type:string, comment:null), ]
 PREHOOK: query: select * from hive_test_dst where pcol1='test_part'
 PREHOOK: type: QUERY
+PREHOOK: Input: default@hive_test_dst
 PREHOOK: Input: default@hive_test_dst@pcol1=test_part/pcol2=test_Part
 #### A masked pattern was here ####
 POSTHOOK: query: select * from hive_test_dst where pcol1='test_part'
 POSTHOOK: type: QUERY
+POSTHOOK: Input: default@hive_test_dst
 POSTHOOK: Input: default@hive_test_dst@pcol1=test_part/pcol2=test_Part
 #### A masked pattern was here ####
 POSTHOOK: Lineage: hive_test_dst PARTITION(pcol1=test_part,pcol2=test_Part).col1 SIMPLE [(hive_test_src)hive_test_src.FieldSchema(name:col1, type:string, comment:null), ]
@@ -74,17 +80,21 @@ POSTHOOK: Lineage: hive_test_dst PARTITI
 6	test_part	test_Part
 PREHOOK: query: select * from hive_test_dst where pcol1='test_part' and pcol2='test_part'
 PREHOOK: type: QUERY
+PREHOOK: Input: default@hive_test_dst
 #### A masked pattern was here ####
 POSTHOOK: query: select * from hive_test_dst where pcol1='test_part' and pcol2='test_part'
 POSTHOOK: type: QUERY
+POSTHOOK: Input: default@hive_test_dst
 #### A masked pattern was here ####
 POSTHOOK: Lineage: hive_test_dst PARTITION(pcol1=test_part,pcol2=test_Part).col1 SIMPLE [(hive_test_src)hive_test_src.FieldSchema(name:col1, type:string, comment:null), ]
 POSTHOOK: Lineage: hive_test_dst PARTITION(pcol1=test_part,pcol2=test_Part).col1 SIMPLE [(hive_test_src)hive_test_src.FieldSchema(name:col1, type:string, comment:null), ]
 PREHOOK: query: select * from hive_test_dst where pcol1='test_Part'
 PREHOOK: type: QUERY
+PREHOOK: Input: default@hive_test_dst
 #### A masked pattern was here ####
 POSTHOOK: query: select * from hive_test_dst where pcol1='test_Part'
 POSTHOOK: type: QUERY
+POSTHOOK: Input: default@hive_test_dst
 #### A masked pattern was here ####
 POSTHOOK: Lineage: hive_test_dst PARTITION(pcol1=test_part,pcol2=test_Part).col1 SIMPLE [(hive_test_src)hive_test_src.FieldSchema(name:col1, type:string, comment:null), ]
 POSTHOOK: Lineage: hive_test_dst PARTITION(pcol1=test_part,pcol2=test_Part).col1 SIMPLE [(hive_test_src)hive_test_src.FieldSchema(name:col1, type:string, comment:null), ]

Modified: hive/branches/ptf-windowing/ql/src/test/results/clientpositive/mapjoin1.q.out
URL: http://svn.apache.org/viewvc/hive/branches/ptf-windowing/ql/src/test/results/clientpositive/mapjoin1.q.out?rev=1462670&r1=1462669&r2=1462670&view=diff
==============================================================================
--- hive/branches/ptf-windowing/ql/src/test/results/clientpositive/mapjoin1.q.out (original)
+++ hive/branches/ptf-windowing/ql/src/test/results/clientpositive/mapjoin1.q.out Fri Mar 29 22:58:43 2013
@@ -222,15 +222,15 @@ POSTHOOK: type: QUERY
 POSTHOOK: Input: default@src
 #### A masked pattern was here ####
 NULL	NULL	238	val_238
-NULL	NULL	238	val_238
 86	val_86	86	val_86
 NULL	NULL	311	val_311
-NULL	NULL	311	val_311
-NULL	NULL	311	val_311
 27	val_27	27	val_27
 NULL	NULL	165	val_165
-NULL	NULL	165	val_165
 NULL	NULL	409	val_409
+NULL	NULL	255	val_255
+NULL	NULL	278	val_278
+98	val_98	98	val_98
+98	val_98	98	val_98
 PREHOOK: query: -- field filter on outer join
 EXPLAIN
 SELECT /*+ MAPJOIN(a) */ * FROM src a RIGHT OUTER JOIN
@@ -345,7 +345,7 @@ NULL	NULL	86	{"key":"86","value":"val_86
 311	val_311	311	{"key":"311","value":"val_311"}
 NULL	NULL	27	{"key":"27","value":"val_27"}
 NULL	NULL	165	{"key":"165","value":"val_165"}
-NULL	NULL	165	{"key":"165","value":"val_165"}
+409	val_409	409	{"key":"409","value":"val_409"}
 409	val_409	409	{"key":"409","value":"val_409"}
 PREHOOK: query: EXPLAIN
 SELECT /*+ MAPJOIN(a) */ * FROM src a RIGHT OUTER JOIN src b on a.key=b.key AND true limit 10

Modified: hive/branches/ptf-windowing/ql/src/test/results/clientpositive/mapjoin_filter_on_outerjoin.q.out
URL: http://svn.apache.org/viewvc/hive/branches/ptf-windowing/ql/src/test/results/clientpositive/mapjoin_filter_on_outerjoin.q.out?rev=1462670&r1=1462669&r2=1462670&view=diff
==============================================================================
--- hive/branches/ptf-windowing/ql/src/test/results/clientpositive/mapjoin_filter_on_outerjoin.q.out (original)
+++ hive/branches/ptf-windowing/ql/src/test/results/clientpositive/mapjoin_filter_on_outerjoin.q.out Fri Mar 29 22:58:43 2013
@@ -253,26 +253,19 @@ ABSTRACT SYNTAX TREE:
   (TOK_QUERY (TOK_FROM (TOK_JOIN (TOK_RIGHTOUTERJOIN (TOK_TABREF (TOK_TABNAME src1)) (TOK_TABREF (TOK_TABNAME src1) src2) (AND (AND (= (. (TOK_TABLE_OR_COL src1) key) (. (TOK_TABLE_OR_COL src2) key)) (< (. (TOK_TABLE_OR_COL src1) key) 10)) (> (. (TOK_TABLE_OR_COL src2) key) 10))) (TOK_TABREF (TOK_TABNAME src) src3) (AND (= (. (TOK_TABLE_OR_COL src2) key) (. (TOK_TABLE_OR_COL src3) key)) (< (. (TOK_TABLE_OR_COL src3) key) 300)))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR TOK_ALLCOLREF)) (TOK_SORTBY (TOK_TABSORTCOLNAMEASC (. (TOK_TABLE_OR_COL src1) key)) (TOK_TABSORTCOLNAMEASC (. (TOK_TABLE_OR_COL src2) key)) (TOK_TABSORTCOLNAMEASC (. (TOK_TABLE_OR_COL src3) key)))))
 
 STAGE DEPENDENCIES:
-  Stage-7 is a root stage , consists of Stage-8, Stage-9, Stage-1
-  Stage-8 has a backup stage: Stage-1
-  Stage-5 depends on stages: Stage-8
-  Stage-2 depends on stages: Stage-1, Stage-5, Stage-6
-  Stage-9 has a backup stage: Stage-1
-  Stage-6 depends on stages: Stage-9
-  Stage-1
+  Stage-6 is a root stage
+  Stage-5 depends on stages: Stage-6
+  Stage-2 depends on stages: Stage-5
   Stage-0 is a root stage
 
 STAGE PLANS:
-  Stage: Stage-7
-    Conditional Operator
-
-  Stage: Stage-8
+  Stage: Stage-6
     Map Reduce Local Work
       Alias -> Map Local Tables:
         src1 
           Fetch Operator
             limit: -1
-        src3 
+        src2 
           Fetch Operator
             limit: -1
       Alias -> Map Local Operator Tree:
@@ -297,10 +290,10 @@ STAGE PLANS:
                   0 [Column[key]]
                   1 [Column[key]]
                   2 [Column[key]]
-                Position of Big Table: 1
-        src3 
+                Position of Big Table: 2
+        src2 
           TableScan
-            alias: src3
+            alias: src2
             Filter Operator
               predicate:
                   expr: (key < 300.0)
@@ -319,14 +312,14 @@ STAGE PLANS:
                   0 [Column[key]]
                   1 [Column[key]]
                   2 [Column[key]]
-                Position of Big Table: 1
+                Position of Big Table: 2
 
   Stage: Stage-5
     Map Reduce
       Alias -> Map Operator Tree:
-        src2 
+        src3 
           TableScan
-            alias: src2
+            alias: src3
             Filter Operator
               predicate:
                   expr: (key < 300.0)
@@ -349,7 +342,7 @@ STAGE PLANS:
                   1 [Column[key]]
                   2 [Column[key]]
                 outputColumnNames: _col0, _col1, _col4, _col5, _col8, _col9
-                Position of Big Table: 1
+                Position of Big Table: 2
                 Select Operator
                   expressions:
                         expr: _col0
@@ -410,217 +403,6 @@ STAGE PLANS:
                 input format: org.apache.hadoop.mapred.TextInputFormat
                 output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
 
-  Stage: Stage-9
-    Map Reduce Local Work
-      Alias -> Map Local Tables:
-        src1 
-          Fetch Operator
-            limit: -1
-        src2 
-          Fetch Operator
-            limit: -1
-      Alias -> Map Local Operator Tree:
-        src1 
-          TableScan
-            alias: src1
-            Filter Operator
-              predicate:
-                  expr: ((key < 300.0) and (key < 10.0))
-                  type: boolean
-              HashTable Sink Operator
-                condition expressions:
-                  0 {key} {value}
-                  1 {key} {value}
-                  2 {key} {value}
-                filter predicates:
-                  0 
-                  1 {(key > 10.0)}
-                  2 
-                handleSkewJoin: false
-                keys:
-                  0 [Column[key]]
-                  1 [Column[key]]
-                  2 [Column[key]]
-                Position of Big Table: 2
-        src2 
-          TableScan
-            alias: src2
-            Filter Operator
-              predicate:
-                  expr: (key < 300.0)
-                  type: boolean
-              HashTable Sink Operator
-                condition expressions:
-                  0 {key} {value}
-                  1 {key} {value}
-                  2 {key} {value}
-                filter predicates:
-                  0 
-                  1 {(key > 10.0)}
-                  2 
-                handleSkewJoin: false
-                keys:
-                  0 [Column[key]]
-                  1 [Column[key]]
-                  2 [Column[key]]
-                Position of Big Table: 2
-
-  Stage: Stage-6
-    Map Reduce
-      Alias -> Map Operator Tree:
-        src3 
-          TableScan
-            alias: src3
-            Filter Operator
-              predicate:
-                  expr: (key < 300.0)
-                  type: boolean
-              Map Join Operator
-                condition map:
-                     Right Outer Join0 to 1
-                     Inner Join 1 to 2
-                condition expressions:
-                  0 {key} {value}
-                  1 {key} {value}
-                  2 {key} {value}
-                filter predicates:
-                  0 
-                  1 {(key > 10.0)}
-                  2 
-                handleSkewJoin: false
-                keys:
-                  0 [Column[key]]
-                  1 [Column[key]]
-                  2 [Column[key]]
-                outputColumnNames: _col0, _col1, _col4, _col5, _col8, _col9
-                Position of Big Table: 2
-                Select Operator
-                  expressions:
-                        expr: _col0
-                        type: string
-                        expr: _col1
-                        type: string
-                        expr: _col4
-                        type: string
-                        expr: _col5
-                        type: string
-                        expr: _col8
-                        type: string
-                        expr: _col9
-                        type: string
-                  outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5
-                  File Output Operator
-                    compressed: false
-                    GlobalTableId: 0
-                    table:
-                        input format: org.apache.hadoop.mapred.SequenceFileInputFormat
-                        output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
-      Local Work:
-        Map Reduce Local Work
-
-  Stage: Stage-1
-    Map Reduce
-      Alias -> Map Operator Tree:
-        src1 
-          TableScan
-            alias: src1
-            Filter Operator
-              predicate:
-                  expr: ((key < 300.0) and (key < 10.0))
-                  type: boolean
-              Reduce Output Operator
-                key expressions:
-                      expr: key
-                      type: string
-                sort order: +
-                Map-reduce partition columns:
-                      expr: key
-                      type: string
-                tag: 0
-                value expressions:
-                      expr: key
-                      type: string
-                      expr: value
-                      type: string
-        src2 
-          TableScan
-            alias: src2
-            Filter Operator
-              predicate:
-                  expr: (key < 300.0)
-                  type: boolean
-              Reduce Output Operator
-                key expressions:
-                      expr: key
-                      type: string
-                sort order: +
-                Map-reduce partition columns:
-                      expr: key
-                      type: string
-                tag: 1
-                value expressions:
-                      expr: key
-                      type: string
-                      expr: value
-                      type: string
-        src3 
-          TableScan
-            alias: src3
-            Filter Operator
-              predicate:
-                  expr: (key < 300.0)
-                  type: boolean
-              Reduce Output Operator
-                key expressions:
-                      expr: key
-                      type: string
-                sort order: +
-                Map-reduce partition columns:
-                      expr: key
-                      type: string
-                tag: 2
-                value expressions:
-                      expr: key
-                      type: string
-                      expr: value
-                      type: string
-      Reduce Operator Tree:
-        Join Operator
-          condition map:
-               Right Outer Join0 to 1
-               Inner Join 1 to 2
-          condition expressions:
-            0 {VALUE._col0} {VALUE._col1}
-            1 {VALUE._col0} {VALUE._col1}
-            2 {VALUE._col0} {VALUE._col1}
-          filter predicates:
-            0 
-            1 {(VALUE._col0 > 10.0)}
-            2 
-          handleSkewJoin: false
-          outputColumnNames: _col0, _col1, _col4, _col5, _col8, _col9
-          Select Operator
-            expressions:
-                  expr: _col0
-                  type: string
-                  expr: _col1
-                  type: string
-                  expr: _col4
-                  type: string
-                  expr: _col5
-                  type: string
-                  expr: _col8
-                  type: string
-                  expr: _col9
-                  type: string
-            outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5
-            File Output Operator
-              compressed: false
-              GlobalTableId: 0
-              table:
-                  input format: org.apache.hadoop.mapred.SequenceFileInputFormat
-                  output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
-
   Stage: Stage-0
     Fetch Operator
       limit: -1

Modified: hive/branches/ptf-windowing/ql/src/test/results/clientpositive/mapjoin_hook.q.out
URL: http://svn.apache.org/viewvc/hive/branches/ptf-windowing/ql/src/test/results/clientpositive/mapjoin_hook.q.out?rev=1462670&r1=1462669&r2=1462670&view=diff
==============================================================================
--- hive/branches/ptf-windowing/ql/src/test/results/clientpositive/mapjoin_hook.q.out (original)
+++ hive/branches/ptf-windowing/ql/src/test/results/clientpositive/mapjoin_hook.q.out Fri Mar 29 22:58:43 2013
@@ -14,7 +14,7 @@ INSERT OVERWRITE TABLE dest1 SELECT src1
 PREHOOK: type: QUERY
 PREHOOK: Input: default@src
 PREHOOK: Output: default@dest1
-[MapJoinCounter PostHook] CONVERTED_LOCAL_MAPJOIN: 1 CONVERTED_MAPJOIN: 1 LOCAL_MAPJOIN: 0 COMMON_JOIN: 0 BACKUP_COMMON_JOIN: 0
+[MapJoinCounter PostHook] CONVERTED_LOCAL_MAPJOIN: 0 CONVERTED_MAPJOIN: 0 LOCAL_MAPJOIN: 1 COMMON_JOIN: 0 BACKUP_COMMON_JOIN: 0
 PREHOOK: query: FROM srcpart src1 JOIN src src2 ON (src1.key = src2.key)
 INSERT OVERWRITE TABLE dest1 SELECT src1.key, src2.value 
 where (src1.ds = '2008-04-08' or src1.ds = '2008-04-09' )and (src1.hr = '12' or src1.hr = '11')

Modified: hive/branches/ptf-windowing/ql/src/test/results/clientpositive/merge4.q.out
URL: http://svn.apache.org/viewvc/hive/branches/ptf-windowing/ql/src/test/results/clientpositive/merge4.q.out?rev=1462670&r1=1462669&r2=1462670&view=diff
==============================================================================
--- hive/branches/ptf-windowing/ql/src/test/results/clientpositive/merge4.q.out (original)
+++ hive/branches/ptf-windowing/ql/src/test/results/clientpositive/merge4.q.out Fri Mar 29 22:58:43 2013
@@ -124,11 +124,13 @@ POSTHOOK: Lineage: nzhang_part PARTITION
 POSTHOOK: Lineage: nzhang_part PARTITION(ds=2010-08-15,hr=12).value SIMPLE [(srcpart)srcpart.FieldSchema(name:value, type:string, comment:default), ]
 PREHOOK: query: select * from nzhang_part
 PREHOOK: type: QUERY
+PREHOOK: Input: default@nzhang_part
 PREHOOK: Input: default@nzhang_part@ds=2010-08-15/hr=11
 PREHOOK: Input: default@nzhang_part@ds=2010-08-15/hr=12
 #### A masked pattern was here ####
 POSTHOOK: query: select * from nzhang_part
 POSTHOOK: type: QUERY
+POSTHOOK: Input: default@nzhang_part
 POSTHOOK: Input: default@nzhang_part@ds=2010-08-15/hr=11
 POSTHOOK: Input: default@nzhang_part@ds=2010-08-15/hr=12
 #### A masked pattern was here ####
@@ -1260,11 +1262,13 @@ POSTHOOK: Lineage: nzhang_part PARTITION
 POSTHOOK: Lineage: nzhang_part PARTITION(ds=2010-08-15,hr=12).value SIMPLE [(srcpart)srcpart.FieldSchema(name:value, type:string, comment:default), ]
 PREHOOK: query: select * from nzhang_part
 PREHOOK: type: QUERY
+PREHOOK: Input: default@nzhang_part
 PREHOOK: Input: default@nzhang_part@ds=2010-08-15/hr=11
 PREHOOK: Input: default@nzhang_part@ds=2010-08-15/hr=12
 #### A masked pattern was here ####
 POSTHOOK: query: select * from nzhang_part
 POSTHOOK: type: QUERY
+POSTHOOK: Input: default@nzhang_part
 POSTHOOK: Input: default@nzhang_part@ds=2010-08-15/hr=11
 POSTHOOK: Input: default@nzhang_part@ds=2010-08-15/hr=12
 #### A masked pattern was here ####
@@ -3014,10 +3018,12 @@ ds=2010-08-15/hr=12
 ds=2010-08-15/hr=file,
 PREHOOK: query: select * from nzhang_part where hr = 'file,'
 PREHOOK: type: QUERY
+PREHOOK: Input: default@nzhang_part
 PREHOOK: Input: default@nzhang_part@ds=2010-08-15/hr=file,
 #### A masked pattern was here ####
 POSTHOOK: query: select * from nzhang_part where hr = 'file,'
 POSTHOOK: type: QUERY
+POSTHOOK: Input: default@nzhang_part
 POSTHOOK: Input: default@nzhang_part@ds=2010-08-15/hr=file,
 #### A masked pattern was here ####
 POSTHOOK: Lineage: nzhang_part PARTITION(ds=2010-08-15,hr=11).key SIMPLE [(srcpart)srcpart.FieldSchema(name:key, type:string, comment:default), ]

Modified: hive/branches/ptf-windowing/ql/src/test/results/clientpositive/mi.q.out
URL: http://svn.apache.org/viewvc/hive/branches/ptf-windowing/ql/src/test/results/clientpositive/mi.q.out?rev=1462670&r1=1462669&r2=1462670&view=diff
==============================================================================
--- hive/branches/ptf-windowing/ql/src/test/results/clientpositive/mi.q.out (original)
+++ hive/branches/ptf-windowing/ql/src/test/results/clientpositive/mi.q.out Fri Mar 29 22:58:43 2013
@@ -60,10 +60,12 @@ POSTHOOK: Lineage: nzhang_t2 PARTITION(d
 ds=2008-04-08/hr=12
 PREHOOK: query: select * from nzhang_t1
 PREHOOK: type: QUERY
+PREHOOK: Input: default@nzhang_t1
 PREHOOK: Input: default@nzhang_t1@ds=2008-04-08/hr=11
 #### A masked pattern was here ####
 POSTHOOK: query: select * from nzhang_t1
 POSTHOOK: type: QUERY
+POSTHOOK: Input: default@nzhang_t1
 POSTHOOK: Input: default@nzhang_t1@ds=2008-04-08/hr=11
 #### A masked pattern was here ####
 POSTHOOK: Lineage: nzhang_t1 PARTITION(ds=2008-04-08,hr=11).key SIMPLE [(srcpart)srcpart.FieldSchema(name:key, type:string, comment:default), ]
@@ -572,10 +574,12 @@ POSTHOOK: Lineage: nzhang_t2 PARTITION(d
 97	val_97	2008-04-08	11
 PREHOOK: query: select * from nzhang_t2
 PREHOOK: type: QUERY
+PREHOOK: Input: default@nzhang_t2
 PREHOOK: Input: default@nzhang_t2@ds=2008-04-08/hr=12
 #### A masked pattern was here ####
 POSTHOOK: query: select * from nzhang_t2
 POSTHOOK: type: QUERY
+POSTHOOK: Input: default@nzhang_t2
 POSTHOOK: Input: default@nzhang_t2@ds=2008-04-08/hr=12
 #### A masked pattern was here ####
 POSTHOOK: Lineage: nzhang_t1 PARTITION(ds=2008-04-08,hr=11).key SIMPLE [(srcpart)srcpart.FieldSchema(name:key, type:string, comment:default), ]