You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hive.apache.org by xu...@apache.org on 2014/10/27 20:57:03 UTC

svn commit: r1634671 [23/46] - in /hive/branches/spark: itests/src/test/resources/ ql/src/test/results/clientpositive/spark/

Added: hive/branches/spark/ql/src/test/results/clientpositive/spark/join_star.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/join_star.q.out?rev=1634671&view=auto
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/spark/join_star.q.out (added)
+++ hive/branches/spark/ql/src/test/results/clientpositive/spark/join_star.q.out Mon Oct 27 19:56:58 2014
@@ -0,0 +1,864 @@
+PREHOOK: query: create table fact(m1 int, m2 int, d1 int, d2 int)
+PREHOOK: type: CREATETABLE
+PREHOOK: Output: database:default
+PREHOOK: Output: default@fact
+POSTHOOK: query: create table fact(m1 int, m2 int, d1 int, d2 int)
+POSTHOOK: type: CREATETABLE
+POSTHOOK: Output: database:default
+POSTHOOK: Output: default@fact
+PREHOOK: query: create table dim1(f1 int, f2 int)
+PREHOOK: type: CREATETABLE
+PREHOOK: Output: database:default
+PREHOOK: Output: default@dim1
+POSTHOOK: query: create table dim1(f1 int, f2 int)
+POSTHOOK: type: CREATETABLE
+POSTHOOK: Output: database:default
+POSTHOOK: Output: default@dim1
+PREHOOK: query: create table dim2(f3 int, f4 int)
+PREHOOK: type: CREATETABLE
+PREHOOK: Output: database:default
+PREHOOK: Output: default@dim2
+POSTHOOK: query: create table dim2(f3 int, f4 int)
+POSTHOOK: type: CREATETABLE
+POSTHOOK: Output: database:default
+POSTHOOK: Output: default@dim2
+PREHOOK: query: create table dim3(f5 int, f6 int)
+PREHOOK: type: CREATETABLE
+PREHOOK: Output: database:default
+PREHOOK: Output: default@dim3
+POSTHOOK: query: create table dim3(f5 int, f6 int)
+POSTHOOK: type: CREATETABLE
+POSTHOOK: Output: database:default
+POSTHOOK: Output: default@dim3
+PREHOOK: query: create table dim4(f7 int, f8 int)
+PREHOOK: type: CREATETABLE
+PREHOOK: Output: database:default
+PREHOOK: Output: default@dim4
+POSTHOOK: query: create table dim4(f7 int, f8 int)
+POSTHOOK: type: CREATETABLE
+POSTHOOK: Output: database:default
+POSTHOOK: Output: default@dim4
+PREHOOK: query: create table dim5(f9 int, f10 int)
+PREHOOK: type: CREATETABLE
+PREHOOK: Output: database:default
+PREHOOK: Output: default@dim5
+POSTHOOK: query: create table dim5(f9 int, f10 int)
+POSTHOOK: type: CREATETABLE
+POSTHOOK: Output: database:default
+POSTHOOK: Output: default@dim5
+PREHOOK: query: create table dim6(f11 int, f12 int)
+PREHOOK: type: CREATETABLE
+PREHOOK: Output: database:default
+PREHOOK: Output: default@dim6
+POSTHOOK: query: create table dim6(f11 int, f12 int)
+POSTHOOK: type: CREATETABLE
+POSTHOOK: Output: database:default
+POSTHOOK: Output: default@dim6
+PREHOOK: query: create table dim7(f13 int, f14 int)
+PREHOOK: type: CREATETABLE
+PREHOOK: Output: database:default
+PREHOOK: Output: default@dim7
+POSTHOOK: query: create table dim7(f13 int, f14 int)
+POSTHOOK: type: CREATETABLE
+POSTHOOK: Output: database:default
+POSTHOOK: Output: default@dim7
+PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/fact-data.txt' INTO TABLE fact
+PREHOOK: type: LOAD
+#### A masked pattern was here ####
+PREHOOK: Output: default@fact
+POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/fact-data.txt' INTO TABLE fact
+POSTHOOK: type: LOAD
+#### A masked pattern was here ####
+POSTHOOK: Output: default@fact
+PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/dim-data.txt' INTO TABLE dim1
+PREHOOK: type: LOAD
+#### A masked pattern was here ####
+PREHOOK: Output: default@dim1
+POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/dim-data.txt' INTO TABLE dim1
+POSTHOOK: type: LOAD
+#### A masked pattern was here ####
+POSTHOOK: Output: default@dim1
+PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/dim-data.txt' INTO TABLE dim2
+PREHOOK: type: LOAD
+#### A masked pattern was here ####
+PREHOOK: Output: default@dim2
+POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/dim-data.txt' INTO TABLE dim2
+POSTHOOK: type: LOAD
+#### A masked pattern was here ####
+POSTHOOK: Output: default@dim2
+PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/dim-data.txt' INTO TABLE dim3
+PREHOOK: type: LOAD
+#### A masked pattern was here ####
+PREHOOK: Output: default@dim3
+POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/dim-data.txt' INTO TABLE dim3
+POSTHOOK: type: LOAD
+#### A masked pattern was here ####
+POSTHOOK: Output: default@dim3
+PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/dim-data.txt' INTO TABLE dim4
+PREHOOK: type: LOAD
+#### A masked pattern was here ####
+PREHOOK: Output: default@dim4
+POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/dim-data.txt' INTO TABLE dim4
+POSTHOOK: type: LOAD
+#### A masked pattern was here ####
+POSTHOOK: Output: default@dim4
+PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/dim-data.txt' INTO TABLE dim5
+PREHOOK: type: LOAD
+#### A masked pattern was here ####
+PREHOOK: Output: default@dim5
+POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/dim-data.txt' INTO TABLE dim5
+POSTHOOK: type: LOAD
+#### A masked pattern was here ####
+POSTHOOK: Output: default@dim5
+PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/dim-data.txt' INTO TABLE dim6
+PREHOOK: type: LOAD
+#### A masked pattern was here ####
+PREHOOK: Output: default@dim6
+POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/dim-data.txt' INTO TABLE dim6
+POSTHOOK: type: LOAD
+#### A masked pattern was here ####
+POSTHOOK: Output: default@dim6
+PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/dim-data.txt' INTO TABLE dim7
+PREHOOK: type: LOAD
+#### A masked pattern was here ####
+PREHOOK: Output: default@dim7
+POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/dim-data.txt' INTO TABLE dim7
+POSTHOOK: type: LOAD
+#### A masked pattern was here ####
+POSTHOOK: Output: default@dim7
+PREHOOK: query: explain select m1, m2, f2 from fact join dim1 on fact.d1=dim1.f1
+PREHOOK: type: QUERY
+POSTHOOK: query: explain select m1, m2, f2 from fact join dim1 on fact.d1=dim1.f1
+POSTHOOK: type: QUERY
+STAGE DEPENDENCIES:
+  Stage-1 is a root stage
+  Stage-0 depends on stages: Stage-1
+
+STAGE PLANS:
+  Stage: Stage-1
+    Spark
+      Edges:
+        Reducer 2 <- Map 1 (GROUP PARTITION-LEVEL SORT, 1), Map 3 (GROUP PARTITION-LEVEL SORT, 1)
+#### A masked pattern was here ####
+      Vertices:
+        Map 1 
+            Map Operator Tree:
+                TableScan
+                  alias: dim1
+                  Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE Column stats: NONE
+                  Filter Operator
+                    predicate: f1 is not null (type: boolean)
+                    Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE
+                    Reduce Output Operator
+                      key expressions: f1 (type: int)
+                      sort order: +
+                      Map-reduce partition columns: f1 (type: int)
+                      Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE
+                      value expressions: f2 (type: int)
+        Map 3 
+            Map Operator Tree:
+                TableScan
+                  alias: fact
+                  Statistics: Num rows: 8 Data size: 98 Basic stats: COMPLETE Column stats: NONE
+                  Filter Operator
+                    predicate: d1 is not null (type: boolean)
+                    Statistics: Num rows: 4 Data size: 49 Basic stats: COMPLETE Column stats: NONE
+                    Reduce Output Operator
+                      key expressions: d1 (type: int)
+                      sort order: +
+                      Map-reduce partition columns: d1 (type: int)
+                      Statistics: Num rows: 4 Data size: 49 Basic stats: COMPLETE Column stats: NONE
+                      value expressions: m1 (type: int), m2 (type: int)
+        Reducer 2 
+            Reduce Operator Tree:
+              Join Operator
+                condition map:
+                     Inner Join 0 to 1
+                condition expressions:
+                  0 {VALUE._col0} {VALUE._col1}
+                  1 {VALUE._col0}
+                outputColumnNames: _col0, _col1, _col8
+                Statistics: Num rows: 4 Data size: 53 Basic stats: COMPLETE Column stats: NONE
+                Select Operator
+                  expressions: _col0 (type: int), _col1 (type: int), _col8 (type: int)
+                  outputColumnNames: _col0, _col1, _col2
+                  Statistics: Num rows: 4 Data size: 53 Basic stats: COMPLETE Column stats: NONE
+                  File Output Operator
+                    compressed: false
+                    Statistics: Num rows: 4 Data size: 53 Basic stats: COMPLETE Column stats: NONE
+                    table:
+                        input format: org.apache.hadoop.mapred.TextInputFormat
+                        output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+                        serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+
+  Stage: Stage-0
+    Fetch Operator
+      limit: -1
+      Processor Tree:
+        ListSink
+
+PREHOOK: query: select m1, m2, f2 from fact join dim1 on fact.d1=dim1.f1
+PREHOOK: type: QUERY
+PREHOOK: Input: default@dim1
+PREHOOK: Input: default@fact
+#### A masked pattern was here ####
+POSTHOOK: query: select m1, m2, f2 from fact join dim1 on fact.d1=dim1.f1
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@dim1
+POSTHOOK: Input: default@fact
+#### A masked pattern was here ####
+1	2	1
+11	12	1
+21	22	1
+31	32	1
+41	42	1
+51	52	3
+61	62	3
+71	72	3
+81	82	3
+91	92	3
+PREHOOK: query: explain select m1, m2, f2, f4 from fact join dim1 on fact.d1=dim1.f1 join dim2 on fact.d2=dim2.f3
+PREHOOK: type: QUERY
+POSTHOOK: query: explain select m1, m2, f2, f4 from fact join dim1 on fact.d1=dim1.f1 join dim2 on fact.d2=dim2.f3
+POSTHOOK: type: QUERY
+STAGE DEPENDENCIES:
+  Stage-1 is a root stage
+  Stage-0 depends on stages: Stage-1
+
+STAGE PLANS:
+  Stage: Stage-1
+    Spark
+      Edges:
+        Reducer 2 <- Map 1 (GROUP PARTITION-LEVEL SORT, 1), Reducer 4 (GROUP PARTITION-LEVEL SORT, 1)
+        Reducer 4 <- Map 3 (GROUP PARTITION-LEVEL SORT, 1), Map 5 (GROUP PARTITION-LEVEL SORT, 1)
+#### A masked pattern was here ####
+      Vertices:
+        Map 1 
+            Map Operator Tree:
+                TableScan
+                  alias: dim2
+                  Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE Column stats: NONE
+                  Filter Operator
+                    predicate: f3 is not null (type: boolean)
+                    Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE
+                    Reduce Output Operator
+                      key expressions: f3 (type: int)
+                      sort order: +
+                      Map-reduce partition columns: f3 (type: int)
+                      Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE
+                      value expressions: f4 (type: int)
+        Map 3 
+            Map Operator Tree:
+                TableScan
+                  alias: dim1
+                  Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE Column stats: NONE
+                  Filter Operator
+                    predicate: f1 is not null (type: boolean)
+                    Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE
+                    Reduce Output Operator
+                      key expressions: f1 (type: int)
+                      sort order: +
+                      Map-reduce partition columns: f1 (type: int)
+                      Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE
+                      value expressions: f2 (type: int)
+        Map 5 
+            Map Operator Tree:
+                TableScan
+                  alias: fact
+                  Statistics: Num rows: 6 Data size: 98 Basic stats: COMPLETE Column stats: NONE
+                  Filter Operator
+                    predicate: (d1 is not null and d2 is not null) (type: boolean)
+                    Statistics: Num rows: 2 Data size: 32 Basic stats: COMPLETE Column stats: NONE
+                    Reduce Output Operator
+                      key expressions: d1 (type: int)
+                      sort order: +
+                      Map-reduce partition columns: d1 (type: int)
+                      Statistics: Num rows: 2 Data size: 32 Basic stats: COMPLETE Column stats: NONE
+                      value expressions: m1 (type: int), m2 (type: int), d2 (type: int)
+        Reducer 2 
+            Reduce Operator Tree:
+              Join Operator
+                condition map:
+                     Inner Join 0 to 1
+                condition expressions:
+                  0 {VALUE._col0} {VALUE._col1} {VALUE._col7}
+                  1 {VALUE._col0}
+                outputColumnNames: _col0, _col1, _col8, _col13
+                Statistics: Num rows: 2 Data size: 38 Basic stats: COMPLETE Column stats: NONE
+                Select Operator
+                  expressions: _col0 (type: int), _col1 (type: int), _col8 (type: int), _col13 (type: int)
+                  outputColumnNames: _col0, _col1, _col2, _col3
+                  Statistics: Num rows: 2 Data size: 38 Basic stats: COMPLETE Column stats: NONE
+                  File Output Operator
+                    compressed: false
+                    Statistics: Num rows: 2 Data size: 38 Basic stats: COMPLETE Column stats: NONE
+                    table:
+                        input format: org.apache.hadoop.mapred.TextInputFormat
+                        output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+                        serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+        Reducer 4 
+            Reduce Operator Tree:
+              Join Operator
+                condition map:
+                     Inner Join 0 to 1
+                condition expressions:
+                  0 {VALUE._col0} {VALUE._col1} {VALUE._col2}
+                  1 {VALUE._col0}
+                outputColumnNames: _col0, _col1, _col3, _col8
+                Statistics: Num rows: 2 Data size: 35 Basic stats: COMPLETE Column stats: NONE
+                Reduce Output Operator
+                  key expressions: _col3 (type: int)
+                  sort order: +
+                  Map-reduce partition columns: _col3 (type: int)
+                  Statistics: Num rows: 2 Data size: 35 Basic stats: COMPLETE Column stats: NONE
+                  value expressions: _col0 (type: int), _col1 (type: int), _col8 (type: int)
+
+  Stage: Stage-0
+    Fetch Operator
+      limit: -1
+      Processor Tree:
+        ListSink
+
+PREHOOK: query: select m1, m2, f2, f4 from fact join dim1 on fact.d1=dim1.f1 join dim2 on fact.d2=dim2.f3
+PREHOOK: type: QUERY
+PREHOOK: Input: default@dim1
+PREHOOK: Input: default@dim2
+PREHOOK: Input: default@fact
+#### A masked pattern was here ####
+POSTHOOK: query: select m1, m2, f2, f4 from fact join dim1 on fact.d1=dim1.f1 join dim2 on fact.d2=dim2.f3
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@dim1
+POSTHOOK: Input: default@dim2
+POSTHOOK: Input: default@fact
+#### A masked pattern was here ####
+1	2	1	2
+11	12	1	2
+21	22	1	2
+31	32	1	2
+41	42	1	2
+51	52	3	4
+61	62	3	4
+71	72	3	4
+81	82	3	4
+91	92	3	4
+PREHOOK: query: explain select m1, m2, f2, f4 from fact join dim1 on fact.d1= dim1.f1 join dim2 on dim1.f2 = dim2.f3
+PREHOOK: type: QUERY
+POSTHOOK: query: explain select m1, m2, f2, f4 from fact join dim1 on fact.d1= dim1.f1 join dim2 on dim1.f2 = dim2.f3
+POSTHOOK: type: QUERY
+STAGE DEPENDENCIES:
+  Stage-1 is a root stage
+  Stage-0 depends on stages: Stage-1
+
+STAGE PLANS:
+  Stage: Stage-1
+    Spark
+      Edges:
+        Reducer 2 <- Map 1 (GROUP PARTITION-LEVEL SORT, 1), Reducer 4 (GROUP PARTITION-LEVEL SORT, 1)
+        Reducer 4 <- Map 3 (GROUP PARTITION-LEVEL SORT, 1), Map 5 (GROUP PARTITION-LEVEL SORT, 1)
+#### A masked pattern was here ####
+      Vertices:
+        Map 1 
+            Map Operator Tree:
+                TableScan
+                  alias: dim2
+                  Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE Column stats: NONE
+                  Filter Operator
+                    predicate: f3 is not null (type: boolean)
+                    Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE
+                    Reduce Output Operator
+                      key expressions: f3 (type: int)
+                      sort order: +
+                      Map-reduce partition columns: f3 (type: int)
+                      Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE
+                      value expressions: f4 (type: int)
+        Map 3 
+            Map Operator Tree:
+                TableScan
+                  alias: dim1
+                  Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE Column stats: NONE
+                  Filter Operator
+                    predicate: (f1 is not null and f2 is not null) (type: boolean)
+                    Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE
+                    Reduce Output Operator
+                      key expressions: f1 (type: int)
+                      sort order: +
+                      Map-reduce partition columns: f1 (type: int)
+                      Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE
+                      value expressions: f2 (type: int)
+        Map 5 
+            Map Operator Tree:
+                TableScan
+                  alias: fact
+                  Statistics: Num rows: 8 Data size: 98 Basic stats: COMPLETE Column stats: NONE
+                  Filter Operator
+                    predicate: d1 is not null (type: boolean)
+                    Statistics: Num rows: 4 Data size: 49 Basic stats: COMPLETE Column stats: NONE
+                    Reduce Output Operator
+                      key expressions: d1 (type: int)
+                      sort order: +
+                      Map-reduce partition columns: d1 (type: int)
+                      Statistics: Num rows: 4 Data size: 49 Basic stats: COMPLETE Column stats: NONE
+                      value expressions: m1 (type: int), m2 (type: int)
+        Reducer 2 
+            Reduce Operator Tree:
+              Join Operator
+                condition map:
+                     Inner Join 0 to 1
+                condition expressions:
+                  0 {VALUE._col0} {VALUE._col1} {KEY.reducesinkkey0}
+                  1 {VALUE._col0}
+                outputColumnNames: _col0, _col1, _col8, _col13
+                Statistics: Num rows: 4 Data size: 58 Basic stats: COMPLETE Column stats: NONE
+                Select Operator
+                  expressions: _col0 (type: int), _col1 (type: int), _col8 (type: int), _col13 (type: int)
+                  outputColumnNames: _col0, _col1, _col2, _col3
+                  Statistics: Num rows: 4 Data size: 58 Basic stats: COMPLETE Column stats: NONE
+                  File Output Operator
+                    compressed: false
+                    Statistics: Num rows: 4 Data size: 58 Basic stats: COMPLETE Column stats: NONE
+                    table:
+                        input format: org.apache.hadoop.mapred.TextInputFormat
+                        output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+                        serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+        Reducer 4 
+            Reduce Operator Tree:
+              Join Operator
+                condition map:
+                     Inner Join 0 to 1
+                condition expressions:
+                  0 {VALUE._col0} {VALUE._col1}
+                  1 {VALUE._col0}
+                outputColumnNames: _col0, _col1, _col8
+                Statistics: Num rows: 4 Data size: 53 Basic stats: COMPLETE Column stats: NONE
+                Reduce Output Operator
+                  key expressions: _col8 (type: int)
+                  sort order: +
+                  Map-reduce partition columns: _col8 (type: int)
+                  Statistics: Num rows: 4 Data size: 53 Basic stats: COMPLETE Column stats: NONE
+                  value expressions: _col0 (type: int), _col1 (type: int)
+
+  Stage: Stage-0
+    Fetch Operator
+      limit: -1
+      Processor Tree:
+        ListSink
+
+PREHOOK: query: select m1, m2, f2, f4 from fact join dim1 on fact.d1= dim1.f1 join dim2 on dim1.f2 = dim2.f3
+PREHOOK: type: QUERY
+PREHOOK: Input: default@dim1
+PREHOOK: Input: default@dim2
+PREHOOK: Input: default@fact
+#### A masked pattern was here ####
+POSTHOOK: query: select m1, m2, f2, f4 from fact join dim1 on fact.d1= dim1.f1 join dim2 on dim1.f2 = dim2.f3
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@dim1
+POSTHOOK: Input: default@dim2
+POSTHOOK: Input: default@fact
+#### A masked pattern was here ####
+1	2	1	1
+11	12	1	1
+21	22	1	1
+31	32	1	1
+41	42	1	1
+51	52	3	3
+61	62	3	3
+71	72	3	3
+81	82	3	3
+91	92	3	3
+PREHOOK: query: explain select m1, m2, f2, f4 from fact Left outer join dim1 on fact.d1= dim1.f1 Left outer join dim2 on dim1.f2 = dim2.f3
+PREHOOK: type: QUERY
+POSTHOOK: query: explain select m1, m2, f2, f4 from fact Left outer join dim1 on fact.d1= dim1.f1 Left outer join dim2 on dim1.f2 = dim2.f3
+POSTHOOK: type: QUERY
+STAGE DEPENDENCIES:
+  Stage-1 is a root stage
+  Stage-0 depends on stages: Stage-1
+
+STAGE PLANS:
+  Stage: Stage-1
+    Spark
+      Edges:
+        Reducer 2 <- Map 1 (GROUP PARTITION-LEVEL SORT, 1), Reducer 4 (GROUP PARTITION-LEVEL SORT, 1)
+        Reducer 4 <- Map 3 (GROUP PARTITION-LEVEL SORT, 1), Map 5 (GROUP PARTITION-LEVEL SORT, 1)
+#### A masked pattern was here ####
+      Vertices:
+        Map 1 
+            Map Operator Tree:
+                TableScan
+                  alias: dim2
+                  Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE Column stats: NONE
+                  Reduce Output Operator
+                    key expressions: f3 (type: int)
+                    sort order: +
+                    Map-reduce partition columns: f3 (type: int)
+                    Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE Column stats: NONE
+                    value expressions: f4 (type: int)
+        Map 3 
+            Map Operator Tree:
+                TableScan
+                  alias: dim1
+                  Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE Column stats: NONE
+                  Reduce Output Operator
+                    key expressions: f1 (type: int)
+                    sort order: +
+                    Map-reduce partition columns: f1 (type: int)
+                    Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE Column stats: NONE
+                    value expressions: f2 (type: int)
+        Map 5 
+            Map Operator Tree:
+                TableScan
+                  alias: fact
+                  Statistics: Num rows: 8 Data size: 98 Basic stats: COMPLETE Column stats: NONE
+                  Reduce Output Operator
+                    key expressions: d1 (type: int)
+                    sort order: +
+                    Map-reduce partition columns: d1 (type: int)
+                    Statistics: Num rows: 8 Data size: 98 Basic stats: COMPLETE Column stats: NONE
+                    value expressions: m1 (type: int), m2 (type: int)
+        Reducer 2 
+            Reduce Operator Tree:
+              Join Operator
+                condition map:
+                     Left Outer Join0 to 1
+                condition expressions:
+                  0 {VALUE._col0} {VALUE._col1} {KEY.reducesinkkey0}
+                  1 {VALUE._col0}
+                outputColumnNames: _col0, _col1, _col8, _col13
+                Statistics: Num rows: 8 Data size: 117 Basic stats: COMPLETE Column stats: NONE
+                Select Operator
+                  expressions: _col0 (type: int), _col1 (type: int), _col8 (type: int), _col13 (type: int)
+                  outputColumnNames: _col0, _col1, _col2, _col3
+                  Statistics: Num rows: 8 Data size: 117 Basic stats: COMPLETE Column stats: NONE
+                  File Output Operator
+                    compressed: false
+                    Statistics: Num rows: 8 Data size: 117 Basic stats: COMPLETE Column stats: NONE
+                    table:
+                        input format: org.apache.hadoop.mapred.TextInputFormat
+                        output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+                        serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+        Reducer 4 
+            Reduce Operator Tree:
+              Join Operator
+                condition map:
+                     Left Outer Join0 to 1
+                condition expressions:
+                  0 {VALUE._col0} {VALUE._col1}
+                  1 {VALUE._col0}
+                outputColumnNames: _col0, _col1, _col8
+                Statistics: Num rows: 8 Data size: 107 Basic stats: COMPLETE Column stats: NONE
+                Reduce Output Operator
+                  key expressions: _col8 (type: int)
+                  sort order: +
+                  Map-reduce partition columns: _col8 (type: int)
+                  Statistics: Num rows: 8 Data size: 107 Basic stats: COMPLETE Column stats: NONE
+                  value expressions: _col0 (type: int), _col1 (type: int)
+
+  Stage: Stage-0
+    Fetch Operator
+      limit: -1
+      Processor Tree:
+        ListSink
+
+PREHOOK: query: select m1, m2, f2, f4 from fact Left outer join dim1 on fact.d1= dim1.f1 Left outer join dim2 on dim1.f2 = dim2.f3
+PREHOOK: type: QUERY
+PREHOOK: Input: default@dim1
+PREHOOK: Input: default@dim2
+PREHOOK: Input: default@fact
+#### A masked pattern was here ####
+POSTHOOK: query: select m1, m2, f2, f4 from fact Left outer join dim1 on fact.d1= dim1.f1 Left outer join dim2 on dim1.f2 = dim2.f3
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@dim1
+POSTHOOK: Input: default@dim2
+POSTHOOK: Input: default@fact
+#### A masked pattern was here ####
+1	2	1	1
+11	12	1	1
+21	22	1	1
+31	32	1	1
+41	42	1	1
+51	52	3	3
+61	62	3	3
+71	72	3	3
+81	82	3	3
+91	92	3	3
+PREHOOK: query: explain Select m1, m2, f2, f4, f6, f8, f10, f12, f14
+ from fact 
+ Left outer join dim1 on  fact.d1= dim1.f1
+ Left outer join dim2 on  dim1.f2 = dim2.f3
+ Left outer Join dim3 on  fact.d2= dim3.f5
+ Left outer Join dim4 on  dim3.f6= dim4.f7
+ Left outer join dim5 on  dim4.f8= dim5.f9
+ Left outer Join dim6 on  dim3.f6= dim6.f11
+ Left outer Join dim7 on  dim6.f12 = dim7.f13
+PREHOOK: type: QUERY
+POSTHOOK: query: explain Select m1, m2, f2, f4, f6, f8, f10, f12, f14
+ from fact 
+ Left outer join dim1 on  fact.d1= dim1.f1
+ Left outer join dim2 on  dim1.f2 = dim2.f3
+ Left outer Join dim3 on  fact.d2= dim3.f5
+ Left outer Join dim4 on  dim3.f6= dim4.f7
+ Left outer join dim5 on  dim4.f8= dim5.f9
+ Left outer Join dim6 on  dim3.f6= dim6.f11
+ Left outer Join dim7 on  dim6.f12 = dim7.f13
+POSTHOOK: type: QUERY
+STAGE DEPENDENCIES:
+  Stage-1 is a root stage
+  Stage-0 depends on stages: Stage-1
+
+STAGE PLANS:
+  Stage: Stage-1
+    Spark
+      Edges:
+        Reducer 2 <- Map 1 (GROUP PARTITION-LEVEL SORT, 1), Reducer 7 (GROUP PARTITION-LEVEL SORT, 1)
+        Reducer 3 <- Map 12 (GROUP PARTITION-LEVEL SORT, 1), Map 14 (GROUP PARTITION-LEVEL SORT, 1), Reducer 2 (GROUP PARTITION-LEVEL SORT, 1)
+        Reducer 4 <- Map 13 (GROUP PARTITION-LEVEL SORT, 1), Reducer 3 (GROUP PARTITION-LEVEL SORT, 1)
+        Reducer 5 <- Map 11 (GROUP PARTITION-LEVEL SORT, 1), Reducer 4 (GROUP PARTITION-LEVEL SORT, 1)
+        Reducer 7 <- Map 6 (GROUP PARTITION-LEVEL SORT, 1), Reducer 9 (GROUP PARTITION-LEVEL SORT, 1)
+        Reducer 9 <- Map 10 (GROUP PARTITION-LEVEL SORT, 1), Map 8 (GROUP PARTITION-LEVEL SORT, 1)
+#### A masked pattern was here ####
+      Vertices:
+        Map 1 
+            Map Operator Tree:
+                TableScan
+                  alias: dim3
+                  Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE Column stats: NONE
+                  Reduce Output Operator
+                    key expressions: f5 (type: int)
+                    sort order: +
+                    Map-reduce partition columns: f5 (type: int)
+                    Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE Column stats: NONE
+                    value expressions: f6 (type: int)
+        Map 10 
+            Map Operator Tree:
+                TableScan
+                  alias: fact
+                  Statistics: Num rows: 6 Data size: 98 Basic stats: COMPLETE Column stats: NONE
+                  Reduce Output Operator
+                    key expressions: d1 (type: int)
+                    sort order: +
+                    Map-reduce partition columns: d1 (type: int)
+                    Statistics: Num rows: 6 Data size: 98 Basic stats: COMPLETE Column stats: NONE
+                    value expressions: m1 (type: int), m2 (type: int), d2 (type: int)
+        Map 11 
+            Map Operator Tree:
+                TableScan
+                  alias: dim7
+                  Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE Column stats: NONE
+                  Reduce Output Operator
+                    key expressions: f13 (type: int)
+                    sort order: +
+                    Map-reduce partition columns: f13 (type: int)
+                    Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE Column stats: NONE
+                    value expressions: f14 (type: int)
+        Map 12 
+            Map Operator Tree:
+                TableScan
+                  alias: dim6
+                  Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE Column stats: NONE
+                  Reduce Output Operator
+                    key expressions: f11 (type: int)
+                    sort order: +
+                    Map-reduce partition columns: f11 (type: int)
+                    Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE Column stats: NONE
+                    value expressions: f12 (type: int)
+        Map 13 
+            Map Operator Tree:
+                TableScan
+                  alias: dim5
+                  Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE Column stats: NONE
+                  Reduce Output Operator
+                    key expressions: f9 (type: int)
+                    sort order: +
+                    Map-reduce partition columns: f9 (type: int)
+                    Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE Column stats: NONE
+                    value expressions: f10 (type: int)
+        Map 14 
+            Map Operator Tree:
+                TableScan
+                  alias: dim4
+                  Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE Column stats: NONE
+                  Reduce Output Operator
+                    key expressions: f7 (type: int)
+                    sort order: +
+                    Map-reduce partition columns: f7 (type: int)
+                    Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE Column stats: NONE
+                    value expressions: f8 (type: int)
+        Map 6 
+            Map Operator Tree:
+                TableScan
+                  alias: dim2
+                  Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE Column stats: NONE
+                  Reduce Output Operator
+                    key expressions: f3 (type: int)
+                    sort order: +
+                    Map-reduce partition columns: f3 (type: int)
+                    Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE Column stats: NONE
+                    value expressions: f4 (type: int)
+        Map 8 
+            Map Operator Tree:
+                TableScan
+                  alias: dim1
+                  Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE Column stats: NONE
+                  Reduce Output Operator
+                    key expressions: f1 (type: int)
+                    sort order: +
+                    Map-reduce partition columns: f1 (type: int)
+                    Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE Column stats: NONE
+                    value expressions: f2 (type: int)
+        Reducer 2 
+            Reduce Operator Tree:
+              Join Operator
+                condition map:
+                     Left Outer Join0 to 1
+                condition expressions:
+                  0 {VALUE._col0} {VALUE._col1} {VALUE._col7} {VALUE._col12}
+                  1 {VALUE._col0}
+                outputColumnNames: _col0, _col1, _col8, _col13, _col18
+                Statistics: Num rows: 6 Data size: 128 Basic stats: COMPLETE Column stats: NONE
+                Reduce Output Operator
+                  key expressions: _col18 (type: int)
+                  sort order: +
+                  Map-reduce partition columns: _col18 (type: int)
+                  Statistics: Num rows: 6 Data size: 128 Basic stats: COMPLETE Column stats: NONE
+                  value expressions: _col0 (type: int), _col1 (type: int), _col8 (type: int), _col13 (type: int)
+        Reducer 3 
+            Reduce Operator Tree:
+              Join Operator
+                condition map:
+                     Left Outer Join0 to 1
+                     Left Outer Join0 to 2
+                condition expressions:
+                  0 {VALUE._col0} {VALUE._col1} {VALUE._col8} {VALUE._col13} {KEY.reducesinkkey0}
+                  1 {VALUE._col0}
+                  2 {VALUE._col0}
+                outputColumnNames: _col0, _col1, _col8, _col13, _col18, _col23, _col28
+                Statistics: Num rows: 13 Data size: 281 Basic stats: COMPLETE Column stats: NONE
+                Reduce Output Operator
+                  key expressions: _col23 (type: int)
+                  sort order: +
+                  Map-reduce partition columns: _col23 (type: int)
+                  Statistics: Num rows: 13 Data size: 281 Basic stats: COMPLETE Column stats: NONE
+                  value expressions: _col0 (type: int), _col1 (type: int), _col8 (type: int), _col13 (type: int), _col18 (type: int), _col28 (type: int)
+        Reducer 4 
+            Reduce Operator Tree:
+              Join Operator
+                condition map:
+                     Left Outer Join0 to 1
+                condition expressions:
+                  0 {VALUE._col0} {VALUE._col1} {VALUE._col8} {VALUE._col13} {VALUE._col18} {KEY.reducesinkkey0} {VALUE._col27}
+                  1 {VALUE._col0}
+                outputColumnNames: _col0, _col1, _col8, _col13, _col18, _col23, _col28, _col33
+                Statistics: Num rows: 14 Data size: 309 Basic stats: COMPLETE Column stats: NONE
+                Reduce Output Operator
+                  key expressions: _col28 (type: int)
+                  sort order: +
+                  Map-reduce partition columns: _col28 (type: int)
+                  Statistics: Num rows: 14 Data size: 309 Basic stats: COMPLETE Column stats: NONE
+                  value expressions: _col0 (type: int), _col1 (type: int), _col8 (type: int), _col13 (type: int), _col18 (type: int), _col23 (type: int), _col33 (type: int)
+        Reducer 5 
+            Reduce Operator Tree:
+              Join Operator
+                condition map:
+                     Left Outer Join0 to 1
+                condition expressions:
+                  0 {VALUE._col0} {VALUE._col1} {VALUE._col8} {VALUE._col13} {VALUE._col18} {VALUE._col23} {KEY.reducesinkkey0} {VALUE._col32}
+                  1 {VALUE._col0}
+                outputColumnNames: _col0, _col1, _col8, _col13, _col18, _col23, _col28, _col33, _col38
+                Statistics: Num rows: 15 Data size: 339 Basic stats: COMPLETE Column stats: NONE
+                Select Operator
+                  expressions: _col0 (type: int), _col1 (type: int), _col8 (type: int), _col13 (type: int), _col18 (type: int), _col23 (type: int), _col33 (type: int), _col28 (type: int), _col38 (type: int)
+                  outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8
+                  Statistics: Num rows: 15 Data size: 339 Basic stats: COMPLETE Column stats: NONE
+                  File Output Operator
+                    compressed: false
+                    Statistics: Num rows: 15 Data size: 339 Basic stats: COMPLETE Column stats: NONE
+                    table:
+                        input format: org.apache.hadoop.mapred.TextInputFormat
+                        output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+                        serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+        Reducer 7 
+            Reduce Operator Tree:
+              Join Operator
+                condition map:
+                     Left Outer Join0 to 1
+                condition expressions:
+                  0 {VALUE._col0} {VALUE._col1} {VALUE._col3} {KEY.reducesinkkey0}
+                  1 {VALUE._col0}
+                outputColumnNames: _col0, _col1, _col3, _col8, _col13
+                Statistics: Num rows: 6 Data size: 117 Basic stats: COMPLETE Column stats: NONE
+                Reduce Output Operator
+                  key expressions: _col3 (type: int)
+                  sort order: +
+                  Map-reduce partition columns: _col3 (type: int)
+                  Statistics: Num rows: 6 Data size: 117 Basic stats: COMPLETE Column stats: NONE
+                  value expressions: _col0 (type: int), _col1 (type: int), _col8 (type: int), _col13 (type: int)
+        Reducer 9 
+            Reduce Operator Tree:
+              Join Operator
+                condition map:
+                     Left Outer Join0 to 1
+                condition expressions:
+                  0 {VALUE._col0} {VALUE._col1} {VALUE._col2}
+                  1 {VALUE._col0}
+                outputColumnNames: _col0, _col1, _col3, _col8
+                Statistics: Num rows: 6 Data size: 107 Basic stats: COMPLETE Column stats: NONE
+                Reduce Output Operator
+                  key expressions: _col8 (type: int)
+                  sort order: +
+                  Map-reduce partition columns: _col8 (type: int)
+                  Statistics: Num rows: 6 Data size: 107 Basic stats: COMPLETE Column stats: NONE
+                  value expressions: _col0 (type: int), _col1 (type: int), _col3 (type: int)
+
+  Stage: Stage-0
+    Fetch Operator
+      limit: -1
+      Processor Tree:
+        ListSink
+
+PREHOOK: query: Select m1, m2, f2, f4, f6, f8, f10, f12, f14
+ from fact 
+ Left outer join dim1 on  fact.d1= dim1.f1
+ Left outer join dim2 on  dim1.f2 = dim2.f3
+ Left outer Join dim3 on  fact.d2= dim3.f5
+ Left outer Join dim4 on  dim3.f6= dim4.f7
+ Left outer join dim5 on  dim4.f8= dim5.f9
+ Left outer Join dim6 on  dim3.f6= dim6.f11
+ Left outer Join dim7 on  dim6.f12 = dim7.f13
+PREHOOK: type: QUERY
+PREHOOK: Input: default@dim1
+PREHOOK: Input: default@dim2
+PREHOOK: Input: default@dim3
+PREHOOK: Input: default@dim4
+PREHOOK: Input: default@dim5
+PREHOOK: Input: default@dim6
+PREHOOK: Input: default@dim7
+PREHOOK: Input: default@fact
+#### A masked pattern was here ####
+POSTHOOK: query: Select m1, m2, f2, f4, f6, f8, f10, f12, f14
+ from fact 
+ Left outer join dim1 on  fact.d1= dim1.f1
+ Left outer join dim2 on  dim1.f2 = dim2.f3
+ Left outer Join dim3 on  fact.d2= dim3.f5
+ Left outer Join dim4 on  dim3.f6= dim4.f7
+ Left outer join dim5 on  dim4.f8= dim5.f9
+ Left outer Join dim6 on  dim3.f6= dim6.f11
+ Left outer Join dim7 on  dim6.f12 = dim7.f13
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@dim1
+POSTHOOK: Input: default@dim2
+POSTHOOK: Input: default@dim3
+POSTHOOK: Input: default@dim4
+POSTHOOK: Input: default@dim5
+POSTHOOK: Input: default@dim6
+POSTHOOK: Input: default@dim7
+POSTHOOK: Input: default@fact
+#### A masked pattern was here ####
+1	2	1	1	2	2	2	2	2
+11	12	1	1	2	2	2	2	2
+21	22	1	1	2	2	2	2	2
+31	32	1	1	2	2	2	2	2
+41	42	1	1	2	2	2	2	2
+51	52	3	3	4	4	4	4	4
+61	62	3	3	4	4	4	4	4
+71	72	3	3	4	4	4	4	4
+81	82	3	3	4	4	4	4	4
+91	92	3	3	4	4	4	4	4

Added: hive/branches/spark/ql/src/test/results/clientpositive/spark/join_thrift.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/join_thrift.q.out?rev=1634671&view=auto
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/spark/join_thrift.q.out (added)
+++ hive/branches/spark/ql/src/test/results/clientpositive/spark/join_thrift.q.out Mon Oct 27 19:56:58 2014
@@ -0,0 +1,119 @@
+PREHOOK: query: DESCRIBE src_thrift
+PREHOOK: type: DESCTABLE
+PREHOOK: Input: default@src_thrift
+POSTHOOK: query: DESCRIBE src_thrift
+POSTHOOK: type: DESCTABLE
+POSTHOOK: Input: default@src_thrift
+aint                	int                 	from deserializer   
+astring             	string              	from deserializer   
+lint                	array<int>          	from deserializer   
+lstring             	array<string>       	from deserializer   
+lintstring          	array<struct<myint:int,mystring:string,underscore_int:int>>	from deserializer   
+mstringstring       	map<string,string>  	from deserializer   
+attributes          	map<string,map<string,map<string,uniontype<int,bigint,string,double,boolean,array<string>,map<string,string>>>>>	from deserializer   
+unionfield1         	uniontype<int,bigint,string,double,boolean,array<string>,map<string,string>>	from deserializer   
+unionfield2         	uniontype<int,bigint,string,double,boolean,array<string>,map<string,string>>	from deserializer   
+unionfield3         	uniontype<int,bigint,string,double,boolean,array<string>,map<string,string>>	from deserializer   
+PREHOOK: query: EXPLAIN
+SELECT s1.aint, s2.lintstring
+FROM src_thrift s1
+JOIN src_thrift s2
+ON s1.aint = s2.aint
+PREHOOK: type: QUERY
+POSTHOOK: query: EXPLAIN
+SELECT s1.aint, s2.lintstring
+FROM src_thrift s1
+JOIN src_thrift s2
+ON s1.aint = s2.aint
+POSTHOOK: type: QUERY
+STAGE DEPENDENCIES:
+  Stage-1 is a root stage
+  Stage-0 depends on stages: Stage-1
+
+STAGE PLANS:
+  Stage: Stage-1
+    Spark
+      Edges:
+        Reducer 2 <- Map 1 (GROUP PARTITION-LEVEL SORT, 1), Map 3 (GROUP PARTITION-LEVEL SORT, 1)
+#### A masked pattern was here ####
+      Vertices:
+        Map 1 
+            Map Operator Tree:
+                TableScan
+                  alias: s2
+                  Statistics: Num rows: 11 Data size: 3070 Basic stats: COMPLETE Column stats: NONE
+                  Filter Operator
+                    predicate: aint is not null (type: boolean)
+                    Statistics: Num rows: 6 Data size: 1674 Basic stats: COMPLETE Column stats: NONE
+                    Reduce Output Operator
+                      key expressions: aint (type: int)
+                      sort order: +
+                      Map-reduce partition columns: aint (type: int)
+                      Statistics: Num rows: 6 Data size: 1674 Basic stats: COMPLETE Column stats: NONE
+                      value expressions: lintstring (type: array<struct<myint:int,mystring:string,underscore_int:int>>)
+        Map 3 
+            Map Operator Tree:
+                TableScan
+                  alias: s1
+                  Statistics: Num rows: 11 Data size: 3070 Basic stats: COMPLETE Column stats: NONE
+                  Filter Operator
+                    predicate: aint is not null (type: boolean)
+                    Statistics: Num rows: 6 Data size: 1674 Basic stats: COMPLETE Column stats: NONE
+                    Reduce Output Operator
+                      key expressions: aint (type: int)
+                      sort order: +
+                      Map-reduce partition columns: aint (type: int)
+                      Statistics: Num rows: 6 Data size: 1674 Basic stats: COMPLETE Column stats: NONE
+        Reducer 2 
+            Reduce Operator Tree:
+              Join Operator
+                condition map:
+                     Inner Join 0 to 1
+                condition expressions:
+                  0 {KEY.reducesinkkey0}
+                  1 {VALUE._col3}
+                outputColumnNames: _col0, _col17
+                Statistics: Num rows: 6 Data size: 1841 Basic stats: COMPLETE Column stats: NONE
+                Select Operator
+                  expressions: _col0 (type: int), _col17 (type: array<struct<myint:int,mystring:string,underscore_int:int>>)
+                  outputColumnNames: _col0, _col1
+                  Statistics: Num rows: 6 Data size: 1841 Basic stats: COMPLETE Column stats: NONE
+                  File Output Operator
+                    compressed: false
+                    Statistics: Num rows: 6 Data size: 1841 Basic stats: COMPLETE Column stats: NONE
+                    table:
+                        input format: org.apache.hadoop.mapred.TextInputFormat
+                        output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+                        serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+
+  Stage: Stage-0
+    Fetch Operator
+      limit: -1
+      Processor Tree:
+        ListSink
+
+PREHOOK: query: SELECT s1.aint, s2.lintstring
+FROM src_thrift s1
+JOIN src_thrift s2
+ON s1.aint = s2.aint
+PREHOOK: type: QUERY
+PREHOOK: Input: default@src_thrift
+#### A masked pattern was here ####
+POSTHOOK: query: SELECT s1.aint, s2.lintstring
+FROM src_thrift s1
+JOIN src_thrift s2
+ON s1.aint = s2.aint
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@src_thrift
+#### A masked pattern was here ####
+-1952710710	[{"myint":25,"mystring":"125","underscore_int":5}]
+-1461153973	[{"myint":49,"mystring":"343","underscore_int":7}]
+-751827638	[{"myint":4,"mystring":"8","underscore_int":2}]
+-734328909	[{"myint":16,"mystring":"64","underscore_int":4}]
+0	NULL
+336964413	[{"myint":81,"mystring":"729","underscore_int":9}]
+465985200	[{"myint":1,"mystring":"1","underscore_int":1}]
+477111222	[{"myint":9,"mystring":"27","underscore_int":3}]
+1244525190	[{"myint":36,"mystring":"216","underscore_int":6}]
+1638581578	[{"myint":64,"mystring":"512","underscore_int":8}]
+1712634731	[{"myint":0,"mystring":"0","underscore_int":0}]

Added: hive/branches/spark/ql/src/test/results/clientpositive/spark/join_vc.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/join_vc.q.out?rev=1634671&view=auto
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/spark/join_vc.q.out (added)
+++ hive/branches/spark/ql/src/test/results/clientpositive/spark/join_vc.q.out Mon Oct 27 19:56:58 2014
@@ -0,0 +1,129 @@
+PREHOOK: query: -- see HIVE-4033 earlier a flag named hasVC was not initialized correctly in MapOperator.java, resulting in NPE for following query. order by and limit in the query is not relevant, problem would be evident even without those. They are there to keep .q.out file small and sorted.
+
+explain select t3.BLOCK__OFFSET__INSIDE__FILE,t3.key,t3.value from src t1 join src t2 on t1.key = t2.key join src t3 on t2.value = t3.value order by t3.BLOCK__OFFSET__INSIDE__FILE,t3.key,t3.value limit 3
+PREHOOK: type: QUERY
+POSTHOOK: query: -- see HIVE-4033 earlier a flag named hasVC was not initialized correctly in MapOperator.java, resulting in NPE for following query. order by and limit in the query is not relevant, problem would be evident even without those. They are there to keep .q.out file small and sorted.
+
+explain select t3.BLOCK__OFFSET__INSIDE__FILE,t3.key,t3.value from src t1 join src t2 on t1.key = t2.key join src t3 on t2.value = t3.value order by t3.BLOCK__OFFSET__INSIDE__FILE,t3.key,t3.value limit 3
+POSTHOOK: type: QUERY
+STAGE DEPENDENCIES:
+  Stage-1 is a root stage
+  Stage-0 depends on stages: Stage-1
+
+STAGE PLANS:
+  Stage: Stage-1
+    Spark
+      Edges:
+        Reducer 2 <- Map 1 (GROUP PARTITION-LEVEL SORT, 1), Reducer 5 (GROUP PARTITION-LEVEL SORT, 1)
+        Reducer 3 <- Reducer 2 (GROUP SORT, 1)
+        Reducer 5 <- Map 4 (GROUP PARTITION-LEVEL SORT, 1), Map 6 (GROUP PARTITION-LEVEL SORT, 1)
+#### A masked pattern was here ####
+      Vertices:
+        Map 1 
+            Map Operator Tree:
+                TableScan
+                  alias: t3
+                  Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
+                  Filter Operator
+                    predicate: value is not null (type: boolean)
+                    Statistics: Num rows: 250 Data size: 2656 Basic stats: COMPLETE Column stats: NONE
+                    Reduce Output Operator
+                      key expressions: value (type: string)
+                      sort order: +
+                      Map-reduce partition columns: value (type: string)
+                      Statistics: Num rows: 250 Data size: 2656 Basic stats: COMPLETE Column stats: NONE
+                      value expressions: key (type: string), BLOCK__OFFSET__INSIDE__FILE (type: bigint)
+        Map 4 
+            Map Operator Tree:
+                TableScan
+                  alias: t2
+                  Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
+                  Filter Operator
+                    predicate: (key is not null and value is not null) (type: boolean)
+                    Statistics: Num rows: 125 Data size: 1328 Basic stats: COMPLETE Column stats: NONE
+                    Reduce Output Operator
+                      key expressions: key (type: string)
+                      sort order: +
+                      Map-reduce partition columns: key (type: string)
+                      Statistics: Num rows: 125 Data size: 1328 Basic stats: COMPLETE Column stats: NONE
+                      value expressions: value (type: string)
+        Map 6 
+            Map Operator Tree:
+                TableScan
+                  alias: t1
+                  Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
+                  Filter Operator
+                    predicate: key is not null (type: boolean)
+                    Statistics: Num rows: 250 Data size: 2656 Basic stats: COMPLETE Column stats: NONE
+                    Reduce Output Operator
+                      key expressions: key (type: string)
+                      sort order: +
+                      Map-reduce partition columns: key (type: string)
+                      Statistics: Num rows: 250 Data size: 2656 Basic stats: COMPLETE Column stats: NONE
+        Reducer 2 
+            Reduce Operator Tree:
+              Join Operator
+                condition map:
+                     Inner Join 0 to 1
+                condition expressions:
+                  0 
+                  1 {VALUE._col0} {KEY.reducesinkkey0} {VALUE._col1}
+                outputColumnNames: _col10, _col11, _col12
+                Statistics: Num rows: 302 Data size: 3213 Basic stats: COMPLETE Column stats: NONE
+                Select Operator
+                  expressions: _col12 (type: bigint), _col10 (type: string), _col11 (type: string)
+                  outputColumnNames: _col0, _col1, _col2
+                  Statistics: Num rows: 302 Data size: 3213 Basic stats: COMPLETE Column stats: NONE
+                  Reduce Output Operator
+                    key expressions: _col0 (type: bigint), _col1 (type: string), _col2 (type: string)
+                    sort order: +++
+                    Statistics: Num rows: 302 Data size: 3213 Basic stats: COMPLETE Column stats: NONE
+        Reducer 3 
+            Reduce Operator Tree:
+              Select Operator
+                expressions: KEY.reducesinkkey0 (type: bigint), KEY.reducesinkkey1 (type: string), KEY.reducesinkkey2 (type: string)
+                outputColumnNames: _col0, _col1, _col2
+                Statistics: Num rows: 302 Data size: 3213 Basic stats: COMPLETE Column stats: NONE
+                Limit
+                  Number of rows: 3
+                  Statistics: Num rows: 3 Data size: 30 Basic stats: COMPLETE Column stats: NONE
+                  File Output Operator
+                    compressed: false
+                    Statistics: Num rows: 3 Data size: 30 Basic stats: COMPLETE Column stats: NONE
+                    table:
+                        input format: org.apache.hadoop.mapred.TextInputFormat
+                        output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+                        serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+        Reducer 5 
+            Reduce Operator Tree:
+              Join Operator
+                condition map:
+                     Inner Join 0 to 1
+                condition expressions:
+                  0 
+                  1 {VALUE._col0}
+                outputColumnNames: _col6
+                Statistics: Num rows: 275 Data size: 2921 Basic stats: COMPLETE Column stats: NONE
+                Reduce Output Operator
+                  key expressions: _col6 (type: string)
+                  sort order: +
+                  Map-reduce partition columns: _col6 (type: string)
+                  Statistics: Num rows: 275 Data size: 2921 Basic stats: COMPLETE Column stats: NONE
+
+  Stage: Stage-0
+    Fetch Operator
+      limit: 3
+      Processor Tree:
+        ListSink
+
+PREHOOK: query: select t3.BLOCK__OFFSET__INSIDE__FILE,t3.key,t3.value from src t1 join src t2 on t1.key = t2.key join src t3 on t2.value = t3.value order by t3.BLOCK__OFFSET__INSIDE__FILE,t3.key,t3.value limit 3
+PREHOOK: type: QUERY
+PREHOOK: Input: default@src
+#### A masked pattern was here ####
+POSTHOOK: query: select t3.BLOCK__OFFSET__INSIDE__FILE,t3.key,t3.value from src t1 join src t2 on t1.key = t2.key join src t3 on t2.value = t3.value order by t3.BLOCK__OFFSET__INSIDE__FILE,t3.key,t3.value limit 3
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@src
+#### A masked pattern was here ####
+0	238	val_238
+0	238	val_238
+0	238	val_238

Added: hive/branches/spark/ql/src/test/results/clientpositive/spark/join_view.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/join_view.q.out?rev=1634671&view=auto
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/spark/join_view.q.out (added)
+++ hive/branches/spark/ql/src/test/results/clientpositive/spark/join_view.q.out Mon Oct 27 19:56:58 2014
@@ -0,0 +1,109 @@
+PREHOOK: query: drop table invites
+PREHOOK: type: DROPTABLE
+POSTHOOK: query: drop table invites
+POSTHOOK: type: DROPTABLE
+PREHOOK: query: drop table invites2
+PREHOOK: type: DROPTABLE
+POSTHOOK: query: drop table invites2
+POSTHOOK: type: DROPTABLE
+PREHOOK: query: create table invites (foo int, bar string) partitioned by (ds string)
+PREHOOK: type: CREATETABLE
+PREHOOK: Output: database:default
+PREHOOK: Output: default@invites
+POSTHOOK: query: create table invites (foo int, bar string) partitioned by (ds string)
+POSTHOOK: type: CREATETABLE
+POSTHOOK: Output: database:default
+POSTHOOK: Output: default@invites
+PREHOOK: query: create table invites2 (foo int, bar string) partitioned by (ds string)
+PREHOOK: type: CREATETABLE
+PREHOOK: Output: database:default
+PREHOOK: Output: default@invites2
+POSTHOOK: query: create table invites2 (foo int, bar string) partitioned by (ds string)
+POSTHOOK: type: CREATETABLE
+POSTHOOK: Output: database:default
+POSTHOOK: Output: default@invites2
+PREHOOK: query: -- test join views: see HIVE-1989
+
+create view v as select invites.bar, invites2.foo, invites2.ds from invites join invites2 on invites.ds=invites2.ds
+PREHOOK: type: CREATEVIEW
+PREHOOK: Input: default@invites
+PREHOOK: Input: default@invites2
+PREHOOK: Output: database:default
+PREHOOK: Output: default@v
+POSTHOOK: query: -- test join views: see HIVE-1989
+
+create view v as select invites.bar, invites2.foo, invites2.ds from invites join invites2 on invites.ds=invites2.ds
+POSTHOOK: type: CREATEVIEW
+POSTHOOK: Input: default@invites
+POSTHOOK: Input: default@invites2
+POSTHOOK: Output: database:default
+POSTHOOK: Output: default@v
+PREHOOK: query: explain select * from v where ds='2011-09-01'
+PREHOOK: type: QUERY
+POSTHOOK: query: explain select * from v where ds='2011-09-01'
+POSTHOOK: type: QUERY
+STAGE DEPENDENCIES:
+  Stage-1 is a root stage
+  Stage-0 depends on stages: Stage-1
+
+STAGE PLANS:
+  Stage: Stage-1
+    Spark
+      Edges:
+        Reducer 2 <- Map 1 (GROUP PARTITION-LEVEL SORT, 1), Map 3 (GROUP PARTITION-LEVEL SORT, 1)
+#### A masked pattern was here ####
+      Vertices:
+        Map 1 
+        Map 3 
+        Reducer 2 
+            Reduce Operator Tree:
+              Join Operator
+                condition map:
+                     Inner Join 0 to 1
+                condition expressions:
+                  0 {VALUE._col1}
+                  1 {VALUE._col0}
+                outputColumnNames: _col1, _col6
+                Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
+                Select Operator
+                  expressions: _col1 (type: string), _col6 (type: int), '2011-09-01' (type: string)
+                  outputColumnNames: _col0, _col1, _col2
+                  Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
+                  File Output Operator
+                    compressed: false
+                    Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
+                    table:
+                        input format: org.apache.hadoop.mapred.TextInputFormat
+                        output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+                        serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+
+  Stage: Stage-0
+    Fetch Operator
+      limit: -1
+      Processor Tree:
+        ListSink
+
+PREHOOK: query: drop view v
+PREHOOK: type: DROPVIEW
+PREHOOK: Input: default@v
+PREHOOK: Output: default@v
+POSTHOOK: query: drop view v
+POSTHOOK: type: DROPVIEW
+POSTHOOK: Input: default@v
+POSTHOOK: Output: default@v
+PREHOOK: query: drop table invites
+PREHOOK: type: DROPTABLE
+PREHOOK: Input: default@invites
+PREHOOK: Output: default@invites
+POSTHOOK: query: drop table invites
+POSTHOOK: type: DROPTABLE
+POSTHOOK: Input: default@invites
+POSTHOOK: Output: default@invites
+PREHOOK: query: drop table invites2
+PREHOOK: type: DROPTABLE
+PREHOOK: Input: default@invites2
+PREHOOK: Output: default@invites2
+POSTHOOK: query: drop table invites2
+POSTHOOK: type: DROPTABLE
+POSTHOOK: Input: default@invites2
+POSTHOOK: Output: default@invites2

Added: hive/branches/spark/ql/src/test/results/clientpositive/spark/leftsemijoin.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/leftsemijoin.q.out?rev=1634671&view=auto
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/spark/leftsemijoin.q.out (added)
+++ hive/branches/spark/ql/src/test/results/clientpositive/spark/leftsemijoin.q.out Mon Oct 27 19:56:58 2014
@@ -0,0 +1,114 @@
+PREHOOK: query: -- SORT_QUERY_RESULTS
+
+drop table sales
+PREHOOK: type: DROPTABLE
+POSTHOOK: query: -- SORT_QUERY_RESULTS
+
+drop table sales
+POSTHOOK: type: DROPTABLE
+PREHOOK: query: drop table things
+PREHOOK: type: DROPTABLE
+POSTHOOK: query: drop table things
+POSTHOOK: type: DROPTABLE
+PREHOOK: query: CREATE TABLE sales (name STRING, id INT)
+ROW FORMAT DELIMITED FIELDS TERMINATED BY '\t'
+PREHOOK: type: CREATETABLE
+PREHOOK: Output: database:default
+PREHOOK: Output: default@sales
+POSTHOOK: query: CREATE TABLE sales (name STRING, id INT)
+ROW FORMAT DELIMITED FIELDS TERMINATED BY '\t'
+POSTHOOK: type: CREATETABLE
+POSTHOOK: Output: database:default
+POSTHOOK: Output: default@sales
+PREHOOK: query: CREATE TABLE things (id INT, name STRING) partitioned by (ds string)
+ROW FORMAT DELIMITED FIELDS TERMINATED BY '\t'
+PREHOOK: type: CREATETABLE
+PREHOOK: Output: database:default
+PREHOOK: Output: default@things
+POSTHOOK: query: CREATE TABLE things (id INT, name STRING) partitioned by (ds string)
+ROW FORMAT DELIMITED FIELDS TERMINATED BY '\t'
+POSTHOOK: type: CREATETABLE
+POSTHOOK: Output: database:default
+POSTHOOK: Output: default@things
+PREHOOK: query: load data local inpath '../../data/files/sales.txt' INTO TABLE sales
+PREHOOK: type: LOAD
+#### A masked pattern was here ####
+PREHOOK: Output: default@sales
+POSTHOOK: query: load data local inpath '../../data/files/sales.txt' INTO TABLE sales
+POSTHOOK: type: LOAD
+#### A masked pattern was here ####
+POSTHOOK: Output: default@sales
+PREHOOK: query: load data local inpath '../../data/files/things.txt' INTO TABLE things partition(ds='2011-10-23')
+PREHOOK: type: LOAD
+#### A masked pattern was here ####
+PREHOOK: Output: default@things
+POSTHOOK: query: load data local inpath '../../data/files/things.txt' INTO TABLE things partition(ds='2011-10-23')
+POSTHOOK: type: LOAD
+#### A masked pattern was here ####
+POSTHOOK: Output: default@things
+POSTHOOK: Output: default@things@ds=2011-10-23
+PREHOOK: query: load data local inpath '../../data/files/things2.txt' INTO TABLE things partition(ds='2011-10-24')
+PREHOOK: type: LOAD
+#### A masked pattern was here ####
+PREHOOK: Output: default@things
+POSTHOOK: query: load data local inpath '../../data/files/things2.txt' INTO TABLE things partition(ds='2011-10-24')
+POSTHOOK: type: LOAD
+#### A masked pattern was here ####
+POSTHOOK: Output: default@things
+POSTHOOK: Output: default@things@ds=2011-10-24
+PREHOOK: query: SELECT name,id FROM sales
+PREHOOK: type: QUERY
+PREHOOK: Input: default@sales
+#### A masked pattern was here ####
+POSTHOOK: query: SELECT name,id FROM sales
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@sales
+#### A masked pattern was here ####
+Hank	2
+Joe	2
+PREHOOK: query: SELECT id,name FROM things
+PREHOOK: type: QUERY
+PREHOOK: Input: default@things
+PREHOOK: Input: default@things@ds=2011-10-23
+PREHOOK: Input: default@things@ds=2011-10-24
+#### A masked pattern was here ####
+POSTHOOK: query: SELECT id,name FROM things
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@things
+POSTHOOK: Input: default@things@ds=2011-10-23
+POSTHOOK: Input: default@things@ds=2011-10-24
+#### A masked pattern was here ####
+2	Tie
+2	Tie
+PREHOOK: query: SELECT name,id FROM sales LEFT SEMI JOIN things ON (sales.id = things.id)
+PREHOOK: type: QUERY
+PREHOOK: Input: default@sales
+PREHOOK: Input: default@things
+PREHOOK: Input: default@things@ds=2011-10-23
+PREHOOK: Input: default@things@ds=2011-10-24
+#### A masked pattern was here ####
+POSTHOOK: query: SELECT name,id FROM sales LEFT SEMI JOIN things ON (sales.id = things.id)
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@sales
+POSTHOOK: Input: default@things
+POSTHOOK: Input: default@things@ds=2011-10-23
+POSTHOOK: Input: default@things@ds=2011-10-24
+#### A masked pattern was here ####
+Hank	2
+Joe	2
+PREHOOK: query: drop table sales
+PREHOOK: type: DROPTABLE
+PREHOOK: Input: default@sales
+PREHOOK: Output: default@sales
+POSTHOOK: query: drop table sales
+POSTHOOK: type: DROPTABLE
+POSTHOOK: Input: default@sales
+POSTHOOK: Output: default@sales
+PREHOOK: query: drop table things
+PREHOOK: type: DROPTABLE
+PREHOOK: Input: default@things
+PREHOOK: Output: default@things
+POSTHOOK: query: drop table things
+POSTHOOK: type: DROPTABLE
+POSTHOOK: Input: default@things
+POSTHOOK: Output: default@things

Added: hive/branches/spark/ql/src/test/results/clientpositive/spark/leftsemijoin_mr.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/leftsemijoin_mr.q.out?rev=1634671&view=auto
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/spark/leftsemijoin_mr.q.out (added)
+++ hive/branches/spark/ql/src/test/results/clientpositive/spark/leftsemijoin_mr.q.out Mon Oct 27 19:56:58 2014
@@ -0,0 +1,98 @@
+PREHOOK: query: CREATE TABLE T1(key INT)
+PREHOOK: type: CREATETABLE
+PREHOOK: Output: database:default
+PREHOOK: Output: default@T1
+POSTHOOK: query: CREATE TABLE T1(key INT)
+POSTHOOK: type: CREATETABLE
+POSTHOOK: Output: database:default
+POSTHOOK: Output: default@T1
+PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/leftsemijoin_mr_t1.txt' INTO TABLE T1
+PREHOOK: type: LOAD
+#### A masked pattern was here ####
+PREHOOK: Output: default@t1
+POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/leftsemijoin_mr_t1.txt' INTO TABLE T1
+POSTHOOK: type: LOAD
+#### A masked pattern was here ####
+POSTHOOK: Output: default@t1
+PREHOOK: query: CREATE TABLE T2(key INT)
+PREHOOK: type: CREATETABLE
+PREHOOK: Output: database:default
+PREHOOK: Output: default@T2
+POSTHOOK: query: CREATE TABLE T2(key INT)
+POSTHOOK: type: CREATETABLE
+POSTHOOK: Output: database:default
+POSTHOOK: Output: default@T2
+PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/leftsemijoin_mr_t2.txt' INTO TABLE T2
+PREHOOK: type: LOAD
+#### A masked pattern was here ####
+PREHOOK: Output: default@t2
+POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/leftsemijoin_mr_t2.txt' INTO TABLE T2
+POSTHOOK: type: LOAD
+#### A masked pattern was here ####
+POSTHOOK: Output: default@t2
+PREHOOK: query: -- Run this query using TestMinimrCliDriver
+
+SELECT * FROM T1
+PREHOOK: type: QUERY
+PREHOOK: Input: default@t1
+#### A masked pattern was here ####
+POSTHOOK: query: -- Run this query using TestMinimrCliDriver
+
+SELECT * FROM T1
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@t1
+#### A masked pattern was here ####
+1
+1
+PREHOOK: query: SELECT * FROM T2
+PREHOOK: type: QUERY
+PREHOOK: Input: default@t2
+#### A masked pattern was here ####
+POSTHOOK: query: SELECT * FROM T2
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@t2
+#### A masked pattern was here ####
+1
+1
+1
+1
+1
+1
+1
+1
+1
+1
+1
+1
+1
+1
+1
+1
+1
+1
+1
+1
+PREHOOK: query: SELECT T1.key FROM T1 LEFT SEMI JOIN (SELECT key FROM T2 SORT BY key) tmp ON (T1.key=tmp.key)
+PREHOOK: type: QUERY
+PREHOOK: Input: default@t1
+PREHOOK: Input: default@t2
+#### A masked pattern was here ####
+POSTHOOK: query: SELECT T1.key FROM T1 LEFT SEMI JOIN (SELECT key FROM T2 SORT BY key) tmp ON (T1.key=tmp.key)
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@t1
+POSTHOOK: Input: default@t2
+#### A masked pattern was here ####
+1
+1
+PREHOOK: query: SELECT T1.key FROM T1 LEFT SEMI JOIN (SELECT key FROM T2 SORT BY key) tmp ON (T1.key=tmp.key)
+PREHOOK: type: QUERY
+PREHOOK: Input: default@t1
+PREHOOK: Input: default@t2
+#### A masked pattern was here ####
+POSTHOOK: query: SELECT T1.key FROM T1 LEFT SEMI JOIN (SELECT key FROM T2 SORT BY key) tmp ON (T1.key=tmp.key)
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@t1
+POSTHOOK: Input: default@t2
+#### A masked pattern was here ####
+1
+1

Added: hive/branches/spark/ql/src/test/results/clientpositive/spark/louter_join_ppr.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/louter_join_ppr.q.out?rev=1634671&view=auto
==============================================================================
Files hive/branches/spark/ql/src/test/results/clientpositive/spark/louter_join_ppr.q.out (added) and hive/branches/spark/ql/src/test/results/clientpositive/spark/louter_join_ppr.q.out Mon Oct 27 19:56:58 2014 differ

Added: hive/branches/spark/ql/src/test/results/clientpositive/spark/mapjoin_addjar.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/mapjoin_addjar.q.out?rev=1634671&view=auto
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/spark/mapjoin_addjar.q.out (added)
+++ hive/branches/spark/ql/src/test/results/clientpositive/spark/mapjoin_addjar.q.out Mon Oct 27 19:56:58 2014
@@ -0,0 +1,37 @@
+PREHOOK: query: CREATE TABLE t1 (a string, b string)
+ROW FORMAT SERDE 'org.apache.hive.hcatalog.data.JsonSerDe'
+PREHOOK: type: CREATETABLE
+PREHOOK: Output: database:default
+PREHOOK: Output: default@t1
+POSTHOOK: query: CREATE TABLE t1 (a string, b string)
+ROW FORMAT SERDE 'org.apache.hive.hcatalog.data.JsonSerDe'
+POSTHOOK: type: CREATETABLE
+POSTHOOK: Output: database:default
+POSTHOOK: Output: default@t1
+PREHOOK: query: LOAD DATA LOCAL INPATH "../../data/files/sample.json" INTO TABLE t1
+PREHOOK: type: LOAD
+#### A masked pattern was here ####
+PREHOOK: Output: default@t1
+POSTHOOK: query: LOAD DATA LOCAL INPATH "../../data/files/sample.json" INTO TABLE t1
+POSTHOOK: type: LOAD
+#### A masked pattern was here ####
+POSTHOOK: Output: default@t1
+PREHOOK: query: select * from src join t1 on src.key =t1.a
+PREHOOK: type: QUERY
+PREHOOK: Input: default@src
+PREHOOK: Input: default@t1
+#### A masked pattern was here ####
+POSTHOOK: query: select * from src join t1 on src.key =t1.a
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@src
+POSTHOOK: Input: default@t1
+#### A masked pattern was here ####
+2	val_2	2	blah
+PREHOOK: query: drop table t1
+PREHOOK: type: DROPTABLE
+PREHOOK: Input: default@t1
+PREHOOK: Output: default@t1
+POSTHOOK: query: drop table t1
+POSTHOOK: type: DROPTABLE
+POSTHOOK: Input: default@t1
+POSTHOOK: Output: default@t1

Added: hive/branches/spark/ql/src/test/results/clientpositive/spark/mapjoin_decimal.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/mapjoin_decimal.q.out?rev=1634671&view=auto
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/spark/mapjoin_decimal.q.out (added)
+++ hive/branches/spark/ql/src/test/results/clientpositive/spark/mapjoin_decimal.q.out Mon Oct 27 19:56:58 2014
@@ -0,0 +1,177 @@
+PREHOOK: query: CREATE TABLE over1k(t tinyint,
+           si smallint,
+           i int,
+           b bigint,
+           f float,
+           d double,
+           bo boolean,
+           s string,
+           ts timestamp,
+           dec decimal(4,2),
+           bin binary)
+ROW FORMAT DELIMITED FIELDS TERMINATED BY '|'
+STORED AS TEXTFILE
+PREHOOK: type: CREATETABLE
+PREHOOK: Output: database:default
+PREHOOK: Output: default@over1k
+POSTHOOK: query: CREATE TABLE over1k(t tinyint,
+           si smallint,
+           i int,
+           b bigint,
+           f float,
+           d double,
+           bo boolean,
+           s string,
+           ts timestamp,
+           dec decimal(4,2),
+           bin binary)
+ROW FORMAT DELIMITED FIELDS TERMINATED BY '|'
+STORED AS TEXTFILE
+POSTHOOK: type: CREATETABLE
+POSTHOOK: Output: database:default
+POSTHOOK: Output: default@over1k
+PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/over1k' OVERWRITE INTO TABLE over1k
+PREHOOK: type: LOAD
+#### A masked pattern was here ####
+PREHOOK: Output: default@over1k
+POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/over1k' OVERWRITE INTO TABLE over1k
+POSTHOOK: type: LOAD
+#### A masked pattern was here ####
+POSTHOOK: Output: default@over1k
+PREHOOK: query: CREATE TABLE t1(dec decimal(4,2)) STORED AS ORC
+PREHOOK: type: CREATETABLE
+PREHOOK: Output: database:default
+PREHOOK: Output: default@t1
+POSTHOOK: query: CREATE TABLE t1(dec decimal(4,2)) STORED AS ORC
+POSTHOOK: type: CREATETABLE
+POSTHOOK: Output: database:default
+POSTHOOK: Output: default@t1
+PREHOOK: query: INSERT INTO TABLE t1 select dec from over1k
+PREHOOK: type: QUERY
+PREHOOK: Input: default@over1k
+PREHOOK: Output: default@t1
+[Error 30017]: Skipping stats aggregation by error org.apache.hadoop.hive.ql.metadata.HiveException: [Error 30015]: Stats aggregator of type counter cannot be connected to
+POSTHOOK: query: INSERT INTO TABLE t1 select dec from over1k
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@over1k
+POSTHOOK: Output: default@t1
+POSTHOOK: Lineage: t1.dec SIMPLE [(over1k)over1k.FieldSchema(name:dec, type:decimal(4,2), comment:null), ]
+PREHOOK: query: CREATE TABLE t2(dec decimal(4,0)) STORED AS ORC
+PREHOOK: type: CREATETABLE
+PREHOOK: Output: database:default
+PREHOOK: Output: default@t2
+POSTHOOK: query: CREATE TABLE t2(dec decimal(4,0)) STORED AS ORC
+POSTHOOK: type: CREATETABLE
+POSTHOOK: Output: database:default
+POSTHOOK: Output: default@t2
+PREHOOK: query: INSERT INTO TABLE t2 select dec from over1k
+PREHOOK: type: QUERY
+PREHOOK: Input: default@over1k
+PREHOOK: Output: default@t2
+[Error 30017]: Skipping stats aggregation by error org.apache.hadoop.hive.ql.metadata.HiveException: [Error 30015]: Stats aggregator of type counter cannot be connected to
+POSTHOOK: query: INSERT INTO TABLE t2 select dec from over1k
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@over1k
+POSTHOOK: Output: default@t2
+POSTHOOK: Lineage: t2.dec EXPRESSION [(over1k)over1k.FieldSchema(name:dec, type:decimal(4,2), comment:null), ]
+PREHOOK: query: explain
+select t1.dec, t2.dec from t1 join t2 on (t1.dec=t2.dec)
+PREHOOK: type: QUERY
+POSTHOOK: query: explain
+select t1.dec, t2.dec from t1 join t2 on (t1.dec=t2.dec)
+POSTHOOK: type: QUERY
+STAGE DEPENDENCIES:
+  Stage-1 is a root stage
+  Stage-0 depends on stages: Stage-1
+
+STAGE PLANS:
+  Stage: Stage-1
+    Spark
+      Edges:
+        Reducer 2 <- Map 1 (GROUP PARTITION-LEVEL SORT, 1), Map 3 (GROUP PARTITION-LEVEL SORT, 1)
+#### A masked pattern was here ####
+      Vertices:
+        Map 1 
+            Map Operator Tree:
+                TableScan
+                  alias: t2
+                  Statistics: Num rows: 12 Data size: 1359 Basic stats: COMPLETE Column stats: NONE
+                  Filter Operator
+                    predicate: dec is not null (type: boolean)
+                    Statistics: Num rows: 6 Data size: 679 Basic stats: COMPLETE Column stats: NONE
+                    Reduce Output Operator
+                      key expressions: dec (type: decimal(4,0))
+                      sort order: +
+                      Map-reduce partition columns: dec (type: decimal(4,0))
+                      Statistics: Num rows: 6 Data size: 679 Basic stats: COMPLETE Column stats: NONE
+        Map 3 
+            Map Operator Tree:
+                TableScan
+                  alias: t1
+                  Statistics: Num rows: 21 Data size: 2422 Basic stats: COMPLETE Column stats: NONE
+                  Filter Operator
+                    predicate: dec is not null (type: boolean)
+                    Statistics: Num rows: 11 Data size: 1268 Basic stats: COMPLETE Column stats: NONE
+                    Reduce Output Operator
+                      key expressions: dec (type: decimal(4,2))
+                      sort order: +
+                      Map-reduce partition columns: dec (type: decimal(4,2))
+                      Statistics: Num rows: 11 Data size: 1268 Basic stats: COMPLETE Column stats: NONE
+        Reducer 2 
+            Reduce Operator Tree:
+              Join Operator
+                condition map:
+                     Inner Join 0 to 1
+                condition expressions:
+                  0 {KEY.reducesinkkey0}
+                  1 {KEY.reducesinkkey0}
+                outputColumnNames: _col0, _col4
+                Statistics: Num rows: 12 Data size: 1394 Basic stats: COMPLETE Column stats: NONE
+                Select Operator
+                  expressions: _col0 (type: decimal(4,2)), _col4 (type: decimal(4,0))
+                  outputColumnNames: _col0, _col1
+                  Statistics: Num rows: 12 Data size: 1394 Basic stats: COMPLETE Column stats: NONE
+                  File Output Operator
+                    compressed: false
+                    Statistics: Num rows: 12 Data size: 1394 Basic stats: COMPLETE Column stats: NONE
+                    table:
+                        input format: org.apache.hadoop.mapred.TextInputFormat
+                        output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+                        serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+
+  Stage: Stage-0
+    Fetch Operator
+      limit: -1
+      Processor Tree:
+        ListSink
+
+PREHOOK: query: select t1.dec, t2.dec from t1 join t2 on (t1.dec=t2.dec)
+PREHOOK: type: QUERY
+PREHOOK: Input: default@t1
+PREHOOK: Input: default@t2
+#### A masked pattern was here ####
+POSTHOOK: query: select t1.dec, t2.dec from t1 join t2 on (t1.dec=t2.dec)
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@t1
+POSTHOOK: Input: default@t2
+#### A masked pattern was here ####
+PREHOOK: query: select t1.dec, t2.dec from t1 join t2 on (t1.dec=t2.dec)
+PREHOOK: type: QUERY
+PREHOOK: Input: default@t1
+PREHOOK: Input: default@t2
+#### A masked pattern was here ####
+POSTHOOK: query: select t1.dec, t2.dec from t1 join t2 on (t1.dec=t2.dec)
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@t1
+POSTHOOK: Input: default@t2
+#### A masked pattern was here ####
+PREHOOK: query: select t1.dec, t2.dec from t1 join t2 on (t1.dec=t2.dec)
+PREHOOK: type: QUERY
+PREHOOK: Input: default@t1
+PREHOOK: Input: default@t2
+#### A masked pattern was here ####
+POSTHOOK: query: select t1.dec, t2.dec from t1 join t2 on (t1.dec=t2.dec)
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@t1
+POSTHOOK: Input: default@t2
+#### A masked pattern was here ####