You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hive.apache.org by xu...@apache.org on 2014/12/02 20:57:14 UTC
svn commit: r1642997 [1/31] - in /hive/branches/spark:
itests/src/test/resources/
ql/src/java/org/apache/hadoop/hive/ql/optimizer/spark/
ql/src/test/results/clientpositive/spark/
Author: xuefu
Date: Tue Dec 2 19:57:10 2014
New Revision: 1642997
URL: http://svn.apache.org/r1642997
Log:
HIVE-8970: Enable map join optimization only when hive.auto.convert.join is true [Spark Branch] (Chao via Xuefu)
Modified:
hive/branches/spark/itests/src/test/resources/testconfiguration.properties
hive/branches/spark/ql/src/java/org/apache/hadoop/hive/ql/optimizer/spark/SparkMapJoinOptimizer.java
hive/branches/spark/ql/src/test/results/clientpositive/spark/annotate_stats_join.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/auto_join_reordering_values.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/auto_smb_mapjoin_14.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/bucket_map_join_1.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/bucket_map_join_2.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin1.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin10.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin11.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin12.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin13.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin2.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin3.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin4.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin5.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin7.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin8.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin9.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin_negative.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin_negative2.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin_negative3.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/column_access_stats.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/cross_join.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/filter_join_breaktask.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/groupby_position.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/groupby_sort_1_23.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/groupby_sort_skew_1_23.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/index_auto_self_join.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/innerjoin.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/join0.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/join1.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/join10.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/join11.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/join12.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/join13.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/join14.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/join15.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/join16.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/join17.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/join19.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/join2.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/join20.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/join21.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/join22.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/join23.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/join25.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/join26.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/join27.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/join3.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/join30.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/join36.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/join37.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/join39.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/join4.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/join40.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/join41.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/join5.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/join8.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/join9.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/join_alt_syntax.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/join_cond_pushdown_1.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/join_cond_pushdown_2.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/join_cond_pushdown_3.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/join_cond_pushdown_4.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/join_cond_pushdown_unqual1.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/join_cond_pushdown_unqual2.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/join_cond_pushdown_unqual3.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/join_cond_pushdown_unqual4.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/join_filters_overlap.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/join_hive_626.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/join_map_ppr.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/join_merge_multi_expressions.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/join_merging.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/join_rc.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/join_reorder.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/join_reorder2.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/join_reorder3.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/join_thrift.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/join_vc.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/join_view.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/limit_pushdown.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/louter_join_ppr.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/mapjoin1.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/mapjoin_distinct.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/mapjoin_filter_on_outerjoin.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/mapjoin_test_outer.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/mergejoins.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/mergejoins_mixed.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/optimize_nullscan.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/parallel_join0.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/parallel_join1.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/pcr.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/ppd_gby_join.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/ppd_join.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/ppd_join2.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/ppd_join3.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/ppd_join5.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/ppd_join_filter.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/ppd_outer_join1.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/ppd_outer_join2.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/ppd_outer_join4.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/ppd_outer_join5.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/router_join_ppr.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/sample8.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/semijoin.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/skewjoin.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/skewjoin_noskew.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/skewjoin_union_remove_1.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/skewjoin_union_remove_2.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/skewjoinopt1.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/skewjoinopt10.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/skewjoinopt11.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/skewjoinopt12.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/skewjoinopt13.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/skewjoinopt14.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/skewjoinopt15.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/skewjoinopt16.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/skewjoinopt17.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/skewjoinopt18.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/skewjoinopt19.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/skewjoinopt2.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/skewjoinopt20.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/skewjoinopt3.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/skewjoinopt4.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/skewjoinopt5.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/skewjoinopt6.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/skewjoinopt7.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/skewjoinopt8.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/skewjoinopt9.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/smb_mapjoin9.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/smb_mapjoin_1.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/smb_mapjoin_10.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/smb_mapjoin_13.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/smb_mapjoin_14.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/smb_mapjoin_15.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/smb_mapjoin_16.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/smb_mapjoin_17.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/smb_mapjoin_2.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/smb_mapjoin_25.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/smb_mapjoin_3.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/smb_mapjoin_4.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/smb_mapjoin_5.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/smb_mapjoin_6.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/sort_merge_join_desc_1.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/sort_merge_join_desc_2.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/sort_merge_join_desc_3.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/sort_merge_join_desc_4.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/sort_merge_join_desc_5.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/sort_merge_join_desc_6.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/sort_merge_join_desc_7.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/sort_merge_join_desc_8.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/subquery_exists.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/subquery_multiinsert.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/temp_table_join1.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/tez_join_tests.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/tez_joins_explain.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/vectorized_bucketmapjoin1.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/vectorized_ptf.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/vectorized_shufflejoin.q.out
hive/branches/spark/ql/src/test/results/clientpositive/spark/windowing.q.out
Modified: hive/branches/spark/itests/src/test/resources/testconfiguration.properties
URL: http://svn.apache.org/viewvc/hive/branches/spark/itests/src/test/resources/testconfiguration.properties?rev=1642997&r1=1642996&r2=1642997&view=diff
==============================================================================
--- hive/branches/spark/itests/src/test/resources/testconfiguration.properties (original)
+++ hive/branches/spark/itests/src/test/resources/testconfiguration.properties Tue Dec 2 19:57:10 2014
@@ -642,7 +642,6 @@ spark.query.files=add_part_multiple.q, \
join35.q, \
join36.q, \
join37.q, \
- join38.q, \
join39.q, \
join40.q, \
join41.q, \
@@ -662,12 +661,10 @@ spark.query.files=add_part_multiple.q, \
join_filters.q, \
join_filters_overlap.q, \
join_hive_626.q, \
- join_literals.q, \
join_map_ppr.q, \
join_merge_multi_expressions.q, \
join_merging.q, \
join_nulls.q, \
- join_nullsafe.q, \
join_rc.q, \
join_reorder.q, \
join_reorder2.q, \
@@ -739,10 +736,8 @@ spark.query.files=add_part_multiple.q, \
ppd_join.q, \
ppd_join2.q, \
ppd_join3.q, \
- ppd_join4.q, \
ppd_join5.q, \
ppd_join_filter.q, \
- ppd_multi_insert.q, \
ppd_outer_join1.q, \
ppd_outer_join2.q, \
ppd_outer_join3.q, \
@@ -853,7 +848,6 @@ spark.query.files=add_part_multiple.q, \
stats9.q, \
statsfs.q, \
subquery_exists.q, \
- subquery_in.q, \
subquery_multiinsert.q, \
table_access_keys_stats.q, \
temp_table.q, \
Modified: hive/branches/spark/ql/src/java/org/apache/hadoop/hive/ql/optimizer/spark/SparkMapJoinOptimizer.java
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/java/org/apache/hadoop/hive/ql/optimizer/spark/SparkMapJoinOptimizer.java?rev=1642997&r1=1642996&r2=1642997&view=diff
==============================================================================
--- hive/branches/spark/ql/src/java/org/apache/hadoop/hive/ql/optimizer/spark/SparkMapJoinOptimizer.java (original)
+++ hive/branches/spark/ql/src/java/org/apache/hadoop/hive/ql/optimizer/spark/SparkMapJoinOptimizer.java Tue Dec 2 19:57:10 2014
@@ -73,15 +73,15 @@ public class SparkMapJoinOptimizer imple
ParseContext parseContext = context.getParseContext();
JoinOperator joinOp = (JoinOperator) nd;
- /*
- if (!conf.getBoolVar(HiveConf.ConfVars.HIVECONVERTJOIN)
- && !(conf.getBoolVar(HiveConf.ConfVars.HIVE_AUTO_SORTMERGE_JOIN))) {
+
+ if (!conf.getBoolVar(HiveConf.ConfVars.HIVECONVERTJOIN)) {
+ // && !(conf.getBoolVar(HiveConf.ConfVars.HIVE_AUTO_SORTMERGE_JOIN))) {
// we are just converting to a common merge join operator. The shuffle
// join in map-reduce case.
- int pos = 0; // it doesn't matter which position we use in this case.
- convertJoinSMBJoin(joinOp, context, pos, 0, false, false);
+ // int pos = 0; // it doesn't matter which position we use in this case.
+ // convertJoinSMBJoin(joinOp, context, pos, 0, false, false);
return null;
- }*/
+ }
// if we have traits, and table info is present in the traits, we know the
// exact number of buckets. Else choose the largest number of estimated
Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/annotate_stats_join.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/annotate_stats_join.q.out?rev=1642997&r1=1642996&r2=1642997&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/spark/annotate_stats_join.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/spark/annotate_stats_join.q.out Tue Dec 2 19:57:10 2014
@@ -155,35 +155,14 @@ POSTHOOK: query: -- number of rows
explain select * from emp e join dept d on (e.deptid = d.deptid)
POSTHOOK: type: QUERY
STAGE DEPENDENCIES:
- Stage-2 is a root stage
- Stage-1 depends on stages: Stage-2
+ Stage-1 is a root stage
Stage-0 depends on stages: Stage-1
STAGE PLANS:
- Stage: Stage-2
- Spark
-#### A masked pattern was here ####
- Vertices:
- Map 2
- Map Operator Tree:
- TableScan
- alias: d
- Statistics: Num rows: 6 Data size: 62 Basic stats: COMPLETE Column stats: COMPLETE
- Filter Operator
- predicate: deptid is not null (type: boolean)
- Statistics: Num rows: 6 Data size: 570 Basic stats: COMPLETE Column stats: COMPLETE
- Spark HashTable Sink Operator
- condition expressions:
- 0 {lastname} {deptid} {locid}
- 1 {deptname}
- keys:
- 0 deptid (type: int)
- 1 deptid (type: int)
- Local Work:
- Map Reduce Local Work
-
Stage: Stage-1
Spark
+ Edges:
+ Reducer 2 <- Map 1 (PARTITION-LEVEL SORT, 3), Map 3 (PARTITION-LEVEL SORT, 3)
#### A masked pattern was here ####
Vertices:
Map 1
@@ -194,32 +173,47 @@ STAGE PLANS:
Filter Operator
predicate: deptid is not null (type: boolean)
Statistics: Num rows: 48 Data size: 4752 Basic stats: COMPLETE Column stats: COMPLETE
- Map Join Operator
- condition map:
- Inner Join 0 to 1
- condition expressions:
- 0 {lastname} {deptid} {locid}
- 1 {deptid} {deptname}
- keys:
- 0 deptid (type: int)
- 1 deptid (type: int)
- outputColumnNames: _col0, _col1, _col2, _col6, _col7
- input vertices:
- 1 Map 2
- Statistics: Num rows: 41 Data size: 7954 Basic stats: COMPLETE Column stats: COMPLETE
- Select Operator
- expressions: _col0 (type: string), _col1 (type: int), _col2 (type: int), _col6 (type: int), _col7 (type: string)
- outputColumnNames: _col0, _col1, _col2, _col3, _col4
- Statistics: Num rows: 41 Data size: 7954 Basic stats: COMPLETE Column stats: COMPLETE
- File Output Operator
- compressed: false
- Statistics: Num rows: 41 Data size: 7954 Basic stats: COMPLETE Column stats: COMPLETE
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- Local Work:
- Map Reduce Local Work
+ Reduce Output Operator
+ key expressions: deptid (type: int)
+ sort order: +
+ Map-reduce partition columns: deptid (type: int)
+ Statistics: Num rows: 48 Data size: 4752 Basic stats: COMPLETE Column stats: COMPLETE
+ value expressions: lastname (type: string), locid (type: int)
+ Map 3
+ Map Operator Tree:
+ TableScan
+ alias: d
+ Statistics: Num rows: 6 Data size: 62 Basic stats: COMPLETE Column stats: COMPLETE
+ Filter Operator
+ predicate: deptid is not null (type: boolean)
+ Statistics: Num rows: 6 Data size: 570 Basic stats: COMPLETE Column stats: COMPLETE
+ Reduce Output Operator
+ key expressions: deptid (type: int)
+ sort order: +
+ Map-reduce partition columns: deptid (type: int)
+ Statistics: Num rows: 6 Data size: 570 Basic stats: COMPLETE Column stats: COMPLETE
+ value expressions: deptname (type: string)
+ Reducer 2
+ Reduce Operator Tree:
+ Join Operator
+ condition map:
+ Inner Join 0 to 1
+ condition expressions:
+ 0 {VALUE._col0} {KEY.reducesinkkey0} {VALUE._col1}
+ 1 {KEY.reducesinkkey0} {VALUE._col0}
+ outputColumnNames: _col0, _col1, _col2, _col6, _col7
+ Statistics: Num rows: 41 Data size: 7954 Basic stats: COMPLETE Column stats: COMPLETE
+ Select Operator
+ expressions: _col0 (type: string), _col1 (type: int), _col2 (type: int), _col6 (type: int), _col7 (type: string)
+ outputColumnNames: _col0, _col1, _col2, _col3, _col4
+ Statistics: Num rows: 41 Data size: 7954 Basic stats: COMPLETE Column stats: COMPLETE
+ File Output Operator
+ compressed: false
+ Statistics: Num rows: 41 Data size: 7954 Basic stats: COMPLETE Column stats: COMPLETE
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
Stage: Stage-0
Fetch Operator
@@ -236,35 +230,14 @@ POSTHOOK: query: -- 2 relations, 2 attri
explain select * from emp,dept where emp.deptid = dept.deptid and emp.lastname = dept.deptname
POSTHOOK: type: QUERY
STAGE DEPENDENCIES:
- Stage-2 is a root stage
- Stage-1 depends on stages: Stage-2
+ Stage-1 is a root stage
Stage-0 depends on stages: Stage-1
STAGE PLANS:
- Stage: Stage-2
- Spark
-#### A masked pattern was here ####
- Vertices:
- Map 2
- Map Operator Tree:
- TableScan
- alias: dept
- Statistics: Num rows: 6 Data size: 62 Basic stats: COMPLETE Column stats: COMPLETE
- Filter Operator
- predicate: (deptid is not null and deptname is not null) (type: boolean)
- Statistics: Num rows: 6 Data size: 570 Basic stats: COMPLETE Column stats: COMPLETE
- Spark HashTable Sink Operator
- condition expressions:
- 0 {lastname} {deptid} {locid}
- 1
- keys:
- 0 deptid (type: int), lastname (type: string)
- 1 deptid (type: int), deptname (type: string)
- Local Work:
- Map Reduce Local Work
-
Stage: Stage-1
Spark
+ Edges:
+ Reducer 2 <- Map 1 (PARTITION-LEVEL SORT, 3), Map 3 (PARTITION-LEVEL SORT, 3)
#### A masked pattern was here ####
Vertices:
Map 1
@@ -275,35 +248,49 @@ STAGE PLANS:
Filter Operator
predicate: (deptid is not null and lastname is not null) (type: boolean)
Statistics: Num rows: 48 Data size: 4752 Basic stats: COMPLETE Column stats: COMPLETE
- Map Join Operator
- condition map:
- Inner Join 0 to 1
- condition expressions:
- 0 {lastname} {deptid} {locid}
- 1 {deptid} {deptname}
- keys:
- 0 deptid (type: int), lastname (type: string)
- 1 deptid (type: int), deptname (type: string)
- outputColumnNames: _col0, _col1, _col2, _col6, _col7
- input vertices:
- 1 Map 2
- Statistics: Num rows: 6 Data size: 1164 Basic stats: COMPLETE Column stats: COMPLETE
- Filter Operator
- predicate: ((_col1 = _col6) and (_col0 = _col7)) (type: boolean)
- Statistics: Num rows: 1 Data size: 194 Basic stats: COMPLETE Column stats: COMPLETE
- Select Operator
- expressions: _col0 (type: string), _col1 (type: int), _col2 (type: int), _col6 (type: int), _col7 (type: string)
- outputColumnNames: _col0, _col1, _col2, _col3, _col4
- Statistics: Num rows: 1 Data size: 194 Basic stats: COMPLETE Column stats: COMPLETE
- File Output Operator
- compressed: false
- Statistics: Num rows: 1 Data size: 194 Basic stats: COMPLETE Column stats: COMPLETE
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- Local Work:
- Map Reduce Local Work
+ Reduce Output Operator
+ key expressions: deptid (type: int), lastname (type: string)
+ sort order: ++
+ Map-reduce partition columns: deptid (type: int), lastname (type: string)
+ Statistics: Num rows: 48 Data size: 4752 Basic stats: COMPLETE Column stats: COMPLETE
+ value expressions: locid (type: int)
+ Map 3
+ Map Operator Tree:
+ TableScan
+ alias: dept
+ Statistics: Num rows: 6 Data size: 62 Basic stats: COMPLETE Column stats: COMPLETE
+ Filter Operator
+ predicate: (deptid is not null and deptname is not null) (type: boolean)
+ Statistics: Num rows: 6 Data size: 570 Basic stats: COMPLETE Column stats: COMPLETE
+ Reduce Output Operator
+ key expressions: deptid (type: int), deptname (type: string)
+ sort order: ++
+ Map-reduce partition columns: deptid (type: int), deptname (type: string)
+ Statistics: Num rows: 6 Data size: 570 Basic stats: COMPLETE Column stats: COMPLETE
+ Reducer 2
+ Reduce Operator Tree:
+ Join Operator
+ condition map:
+ Inner Join 0 to 1
+ condition expressions:
+ 0 {KEY.reducesinkkey1} {KEY.reducesinkkey0} {VALUE._col0}
+ 1 {KEY.reducesinkkey0} {KEY.reducesinkkey1}
+ outputColumnNames: _col0, _col1, _col2, _col6, _col7
+ Statistics: Num rows: 6 Data size: 1164 Basic stats: COMPLETE Column stats: COMPLETE
+ Filter Operator
+ predicate: ((_col1 = _col6) and (_col0 = _col7)) (type: boolean)
+ Statistics: Num rows: 1 Data size: 194 Basic stats: COMPLETE Column stats: COMPLETE
+ Select Operator
+ expressions: _col0 (type: string), _col1 (type: int), _col2 (type: int), _col6 (type: int), _col7 (type: string)
+ outputColumnNames: _col0, _col1, _col2, _col3, _col4
+ Statistics: Num rows: 1 Data size: 194 Basic stats: COMPLETE Column stats: COMPLETE
+ File Output Operator
+ compressed: false
+ Statistics: Num rows: 1 Data size: 194 Basic stats: COMPLETE Column stats: COMPLETE
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
Stage: Stage-0
Fetch Operator
@@ -316,35 +303,14 @@ PREHOOK: type: QUERY
POSTHOOK: query: explain select * from emp e join dept d on (e.deptid = d.deptid and e.lastname = d.deptname)
POSTHOOK: type: QUERY
STAGE DEPENDENCIES:
- Stage-2 is a root stage
- Stage-1 depends on stages: Stage-2
+ Stage-1 is a root stage
Stage-0 depends on stages: Stage-1
STAGE PLANS:
- Stage: Stage-2
- Spark
-#### A masked pattern was here ####
- Vertices:
- Map 2
- Map Operator Tree:
- TableScan
- alias: d
- Statistics: Num rows: 6 Data size: 62 Basic stats: COMPLETE Column stats: COMPLETE
- Filter Operator
- predicate: (deptid is not null and deptname is not null) (type: boolean)
- Statistics: Num rows: 6 Data size: 570 Basic stats: COMPLETE Column stats: COMPLETE
- Spark HashTable Sink Operator
- condition expressions:
- 0 {lastname} {deptid} {locid}
- 1
- keys:
- 0 deptid (type: int), lastname (type: string)
- 1 deptid (type: int), deptname (type: string)
- Local Work:
- Map Reduce Local Work
-
Stage: Stage-1
Spark
+ Edges:
+ Reducer 2 <- Map 1 (PARTITION-LEVEL SORT, 3), Map 3 (PARTITION-LEVEL SORT, 3)
#### A masked pattern was here ####
Vertices:
Map 1
@@ -355,32 +321,46 @@ STAGE PLANS:
Filter Operator
predicate: (deptid is not null and lastname is not null) (type: boolean)
Statistics: Num rows: 48 Data size: 4752 Basic stats: COMPLETE Column stats: COMPLETE
- Map Join Operator
- condition map:
- Inner Join 0 to 1
- condition expressions:
- 0 {lastname} {deptid} {locid}
- 1 {deptid} {deptname}
- keys:
- 0 deptid (type: int), lastname (type: string)
- 1 deptid (type: int), deptname (type: string)
- outputColumnNames: _col0, _col1, _col2, _col6, _col7
- input vertices:
- 1 Map 2
- Statistics: Num rows: 6 Data size: 1164 Basic stats: COMPLETE Column stats: COMPLETE
- Select Operator
- expressions: _col0 (type: string), _col1 (type: int), _col2 (type: int), _col6 (type: int), _col7 (type: string)
- outputColumnNames: _col0, _col1, _col2, _col3, _col4
- Statistics: Num rows: 6 Data size: 1164 Basic stats: COMPLETE Column stats: COMPLETE
- File Output Operator
- compressed: false
- Statistics: Num rows: 6 Data size: 1164 Basic stats: COMPLETE Column stats: COMPLETE
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- Local Work:
- Map Reduce Local Work
+ Reduce Output Operator
+ key expressions: deptid (type: int), lastname (type: string)
+ sort order: ++
+ Map-reduce partition columns: deptid (type: int), lastname (type: string)
+ Statistics: Num rows: 48 Data size: 4752 Basic stats: COMPLETE Column stats: COMPLETE
+ value expressions: locid (type: int)
+ Map 3
+ Map Operator Tree:
+ TableScan
+ alias: d
+ Statistics: Num rows: 6 Data size: 62 Basic stats: COMPLETE Column stats: COMPLETE
+ Filter Operator
+ predicate: (deptid is not null and deptname is not null) (type: boolean)
+ Statistics: Num rows: 6 Data size: 570 Basic stats: COMPLETE Column stats: COMPLETE
+ Reduce Output Operator
+ key expressions: deptid (type: int), deptname (type: string)
+ sort order: ++
+ Map-reduce partition columns: deptid (type: int), deptname (type: string)
+ Statistics: Num rows: 6 Data size: 570 Basic stats: COMPLETE Column stats: COMPLETE
+ Reducer 2
+ Reduce Operator Tree:
+ Join Operator
+ condition map:
+ Inner Join 0 to 1
+ condition expressions:
+ 0 {KEY.reducesinkkey1} {KEY.reducesinkkey0} {VALUE._col0}
+ 1 {KEY.reducesinkkey0} {KEY.reducesinkkey1}
+ outputColumnNames: _col0, _col1, _col2, _col6, _col7
+ Statistics: Num rows: 6 Data size: 1164 Basic stats: COMPLETE Column stats: COMPLETE
+ Select Operator
+ expressions: _col0 (type: string), _col1 (type: int), _col2 (type: int), _col6 (type: int), _col7 (type: string)
+ outputColumnNames: _col0, _col1, _col2, _col3, _col4
+ Statistics: Num rows: 6 Data size: 1164 Basic stats: COMPLETE Column stats: COMPLETE
+ File Output Operator
+ compressed: false
+ Statistics: Num rows: 6 Data size: 1164 Basic stats: COMPLETE Column stats: COMPLETE
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
Stage: Stage-0
Fetch Operator
@@ -397,35 +377,14 @@ POSTHOOK: query: -- 2 relations, 3 attri
explain select * from emp,dept where emp.deptid = dept.deptid and emp.lastname = dept.deptname and dept.deptname = emp.lastname
POSTHOOK: type: QUERY
STAGE DEPENDENCIES:
- Stage-2 is a root stage
- Stage-1 depends on stages: Stage-2
+ Stage-1 is a root stage
Stage-0 depends on stages: Stage-1
STAGE PLANS:
- Stage: Stage-2
- Spark
-#### A masked pattern was here ####
- Vertices:
- Map 2
- Map Operator Tree:
- TableScan
- alias: dept
- Statistics: Num rows: 6 Data size: 62 Basic stats: COMPLETE Column stats: COMPLETE
- Filter Operator
- predicate: (deptid is not null and deptname is not null) (type: boolean)
- Statistics: Num rows: 6 Data size: 570 Basic stats: COMPLETE Column stats: COMPLETE
- Spark HashTable Sink Operator
- condition expressions:
- 0 {lastname} {deptid} {locid}
- 1
- keys:
- 0 deptid (type: int), lastname (type: string), lastname (type: string)
- 1 deptid (type: int), deptname (type: string), deptname (type: string)
- Local Work:
- Map Reduce Local Work
-
Stage: Stage-1
Spark
+ Edges:
+ Reducer 2 <- Map 1 (PARTITION-LEVEL SORT, 3), Map 3 (PARTITION-LEVEL SORT, 3)
#### A masked pattern was here ####
Vertices:
Map 1
@@ -436,35 +395,49 @@ STAGE PLANS:
Filter Operator
predicate: (deptid is not null and lastname is not null) (type: boolean)
Statistics: Num rows: 48 Data size: 4752 Basic stats: COMPLETE Column stats: COMPLETE
- Map Join Operator
- condition map:
- Inner Join 0 to 1
- condition expressions:
- 0 {lastname} {deptid} {locid}
- 1 {deptid} {deptname}
- keys:
- 0 deptid (type: int), lastname (type: string), lastname (type: string)
- 1 deptid (type: int), deptname (type: string), deptname (type: string)
- outputColumnNames: _col0, _col1, _col2, _col6, _col7
- input vertices:
- 1 Map 2
- Statistics: Num rows: 11 Data size: 2134 Basic stats: COMPLETE Column stats: COMPLETE
- Filter Operator
- predicate: (((_col1 = _col6) and (_col0 = _col7)) and (_col7 = _col0)) (type: boolean)
- Statistics: Num rows: 1 Data size: 194 Basic stats: COMPLETE Column stats: COMPLETE
- Select Operator
- expressions: _col0 (type: string), _col1 (type: int), _col2 (type: int), _col6 (type: int), _col7 (type: string)
- outputColumnNames: _col0, _col1, _col2, _col3, _col4
- Statistics: Num rows: 1 Data size: 194 Basic stats: COMPLETE Column stats: COMPLETE
- File Output Operator
- compressed: false
- Statistics: Num rows: 1 Data size: 194 Basic stats: COMPLETE Column stats: COMPLETE
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- Local Work:
- Map Reduce Local Work
+ Reduce Output Operator
+ key expressions: deptid (type: int), lastname (type: string), lastname (type: string)
+ sort order: +++
+ Map-reduce partition columns: deptid (type: int), lastname (type: string), lastname (type: string)
+ Statistics: Num rows: 48 Data size: 4752 Basic stats: COMPLETE Column stats: COMPLETE
+ value expressions: locid (type: int)
+ Map 3
+ Map Operator Tree:
+ TableScan
+ alias: dept
+ Statistics: Num rows: 6 Data size: 62 Basic stats: COMPLETE Column stats: COMPLETE
+ Filter Operator
+ predicate: (deptid is not null and deptname is not null) (type: boolean)
+ Statistics: Num rows: 6 Data size: 570 Basic stats: COMPLETE Column stats: COMPLETE
+ Reduce Output Operator
+ key expressions: deptid (type: int), deptname (type: string), deptname (type: string)
+ sort order: +++
+ Map-reduce partition columns: deptid (type: int), deptname (type: string), deptname (type: string)
+ Statistics: Num rows: 6 Data size: 570 Basic stats: COMPLETE Column stats: COMPLETE
+ Reducer 2
+ Reduce Operator Tree:
+ Join Operator
+ condition map:
+ Inner Join 0 to 1
+ condition expressions:
+ 0 {KEY.reducesinkkey1} {KEY.reducesinkkey0} {VALUE._col0}
+ 1 {KEY.reducesinkkey0} {KEY.reducesinkkey1}
+ outputColumnNames: _col0, _col1, _col2, _col6, _col7
+ Statistics: Num rows: 11 Data size: 2134 Basic stats: COMPLETE Column stats: COMPLETE
+ Filter Operator
+ predicate: (((_col1 = _col6) and (_col0 = _col7)) and (_col7 = _col0)) (type: boolean)
+ Statistics: Num rows: 1 Data size: 194 Basic stats: COMPLETE Column stats: COMPLETE
+ Select Operator
+ expressions: _col0 (type: string), _col1 (type: int), _col2 (type: int), _col6 (type: int), _col7 (type: string)
+ outputColumnNames: _col0, _col1, _col2, _col3, _col4
+ Statistics: Num rows: 1 Data size: 194 Basic stats: COMPLETE Column stats: COMPLETE
+ File Output Operator
+ compressed: false
+ Statistics: Num rows: 1 Data size: 194 Basic stats: COMPLETE Column stats: COMPLETE
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
Stage: Stage-0
Fetch Operator
@@ -481,96 +454,81 @@ POSTHOOK: query: -- 3 relations, 1 attri
explain select * from emp e join dept d on (e.deptid = d.deptid) join emp e1 on (e.deptid = e1.deptid)
POSTHOOK: type: QUERY
STAGE DEPENDENCIES:
- Stage-2 is a root stage
- Stage-1 depends on stages: Stage-2
+ Stage-1 is a root stage
Stage-0 depends on stages: Stage-1
STAGE PLANS:
- Stage: Stage-2
+ Stage: Stage-1
Spark
+ Edges:
+ Reducer 2 <- Map 1 (PARTITION-LEVEL SORT, 3), Map 3 (PARTITION-LEVEL SORT, 3), Map 4 (PARTITION-LEVEL SORT, 3)
#### A masked pattern was here ####
Vertices:
- Map 2
+ Map 1
Map Operator Tree:
TableScan
- alias: d
- Statistics: Num rows: 6 Data size: 62 Basic stats: COMPLETE Column stats: COMPLETE
+ alias: e
+ Statistics: Num rows: 48 Data size: 552 Basic stats: COMPLETE Column stats: COMPLETE
Filter Operator
predicate: deptid is not null (type: boolean)
- Statistics: Num rows: 6 Data size: 570 Basic stats: COMPLETE Column stats: COMPLETE
- Spark HashTable Sink Operator
- condition expressions:
- 0 {lastname} {deptid} {locid}
- 1 {deptname}
- 2 {lastname} {deptid} {locid}
- keys:
- 0 deptid (type: int)
- 1 deptid (type: int)
- 2 deptid (type: int)
- Local Work:
- Map Reduce Local Work
+ Statistics: Num rows: 48 Data size: 4752 Basic stats: COMPLETE Column stats: COMPLETE
+ Reduce Output Operator
+ key expressions: deptid (type: int)
+ sort order: +
+ Map-reduce partition columns: deptid (type: int)
+ Statistics: Num rows: 48 Data size: 4752 Basic stats: COMPLETE Column stats: COMPLETE
+ value expressions: lastname (type: string), locid (type: int)
Map 3
Map Operator Tree:
TableScan
- alias: e1
- Statistics: Num rows: 48 Data size: 552 Basic stats: COMPLETE Column stats: COMPLETE
+ alias: d
+ Statistics: Num rows: 6 Data size: 62 Basic stats: COMPLETE Column stats: COMPLETE
Filter Operator
predicate: deptid is not null (type: boolean)
- Statistics: Num rows: 48 Data size: 4752 Basic stats: COMPLETE Column stats: COMPLETE
- Spark HashTable Sink Operator
- condition expressions:
- 0 {lastname} {deptid} {locid}
- 1 {deptid} {deptname}
- 2 {lastname} {locid}
- keys:
- 0 deptid (type: int)
- 1 deptid (type: int)
- 2 deptid (type: int)
- Local Work:
- Map Reduce Local Work
-
- Stage: Stage-1
- Spark
-#### A masked pattern was here ####
- Vertices:
- Map 1
+ Statistics: Num rows: 6 Data size: 570 Basic stats: COMPLETE Column stats: COMPLETE
+ Reduce Output Operator
+ key expressions: deptid (type: int)
+ sort order: +
+ Map-reduce partition columns: deptid (type: int)
+ Statistics: Num rows: 6 Data size: 570 Basic stats: COMPLETE Column stats: COMPLETE
+ value expressions: deptname (type: string)
+ Map 4
Map Operator Tree:
TableScan
- alias: e
+ alias: e1
Statistics: Num rows: 48 Data size: 552 Basic stats: COMPLETE Column stats: COMPLETE
Filter Operator
predicate: deptid is not null (type: boolean)
Statistics: Num rows: 48 Data size: 4752 Basic stats: COMPLETE Column stats: COMPLETE
- Map Join Operator
- condition map:
- Inner Join 0 to 1
- Inner Join 0 to 2
- condition expressions:
- 0 {lastname} {deptid} {locid}
- 1 {deptid} {deptname}
- 2 {lastname} {deptid} {locid}
- keys:
- 0 deptid (type: int)
- 1 deptid (type: int)
- 2 deptid (type: int)
- outputColumnNames: _col0, _col1, _col2, _col6, _col7, _col11, _col12, _col13
- input vertices:
- 1 Map 2
- 2 Map 3
- Statistics: Num rows: 658 Data size: 192794 Basic stats: COMPLETE Column stats: COMPLETE
- Select Operator
- expressions: _col0 (type: string), _col1 (type: int), _col2 (type: int), _col6 (type: int), _col7 (type: string), _col11 (type: string), _col12 (type: int), _col13 (type: int)
- outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7
- Statistics: Num rows: 658 Data size: 192794 Basic stats: COMPLETE Column stats: COMPLETE
- File Output Operator
- compressed: false
- Statistics: Num rows: 658 Data size: 192794 Basic stats: COMPLETE Column stats: COMPLETE
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- Local Work:
- Map Reduce Local Work
+ Reduce Output Operator
+ key expressions: deptid (type: int)
+ sort order: +
+ Map-reduce partition columns: deptid (type: int)
+ Statistics: Num rows: 48 Data size: 4752 Basic stats: COMPLETE Column stats: COMPLETE
+ value expressions: lastname (type: string), locid (type: int)
+ Reducer 2
+ Reduce Operator Tree:
+ Join Operator
+ condition map:
+ Inner Join 0 to 1
+ Inner Join 0 to 2
+ condition expressions:
+ 0 {VALUE._col0} {KEY.reducesinkkey0} {VALUE._col1}
+ 1 {KEY.reducesinkkey0} {VALUE._col0}
+ 2 {VALUE._col0} {KEY.reducesinkkey0} {VALUE._col1}
+ outputColumnNames: _col0, _col1, _col2, _col6, _col7, _col11, _col12, _col13
+ Statistics: Num rows: 658 Data size: 192794 Basic stats: COMPLETE Column stats: COMPLETE
+ Select Operator
+ expressions: _col0 (type: string), _col1 (type: int), _col2 (type: int), _col6 (type: int), _col7 (type: string), _col11 (type: string), _col12 (type: int), _col13 (type: int)
+ outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7
+ Statistics: Num rows: 658 Data size: 192794 Basic stats: COMPLETE Column stats: COMPLETE
+ File Output Operator
+ compressed: false
+ Statistics: Num rows: 658 Data size: 192794 Basic stats: COMPLETE Column stats: COMPLETE
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
Stage: Stage-0
Fetch Operator
@@ -585,16 +543,31 @@ POSTHOOK: query: -- Expected output rows
explain select * from emp e join dept d on (e.deptid = d.deptid) join loc l on (e.deptid = l.locid)
POSTHOOK: type: QUERY
STAGE DEPENDENCIES:
- Stage-2 is a root stage
- Stage-1 depends on stages: Stage-2
+ Stage-1 is a root stage
Stage-0 depends on stages: Stage-1
STAGE PLANS:
- Stage: Stage-2
+ Stage: Stage-1
Spark
+ Edges:
+ Reducer 2 <- Map 1 (PARTITION-LEVEL SORT, 3), Map 3 (PARTITION-LEVEL SORT, 3), Map 4 (PARTITION-LEVEL SORT, 3)
#### A masked pattern was here ####
Vertices:
- Map 2
+ Map 1
+ Map Operator Tree:
+ TableScan
+ alias: e
+ Statistics: Num rows: 48 Data size: 552 Basic stats: COMPLETE Column stats: COMPLETE
+ Filter Operator
+ predicate: deptid is not null (type: boolean)
+ Statistics: Num rows: 48 Data size: 4752 Basic stats: COMPLETE Column stats: COMPLETE
+ Reduce Output Operator
+ key expressions: deptid (type: int)
+ sort order: +
+ Map-reduce partition columns: deptid (type: int)
+ Statistics: Num rows: 48 Data size: 4752 Basic stats: COMPLETE Column stats: COMPLETE
+ value expressions: lastname (type: string), locid (type: int)
+ Map 3
Map Operator Tree:
TableScan
alias: d
@@ -602,18 +575,13 @@ STAGE PLANS:
Filter Operator
predicate: deptid is not null (type: boolean)
Statistics: Num rows: 6 Data size: 570 Basic stats: COMPLETE Column stats: COMPLETE
- Spark HashTable Sink Operator
- condition expressions:
- 0 {lastname} {deptid} {locid}
- 1 {deptname}
- 2 {state} {locid} {zip} {year}
- keys:
- 0 deptid (type: int)
- 1 deptid (type: int)
- 2 locid (type: int)
- Local Work:
- Map Reduce Local Work
- Map 3
+ Reduce Output Operator
+ key expressions: deptid (type: int)
+ sort order: +
+ Map-reduce partition columns: deptid (type: int)
+ Statistics: Num rows: 6 Data size: 570 Basic stats: COMPLETE Column stats: COMPLETE
+ value expressions: deptname (type: string)
+ Map 4
Map Operator Tree:
TableScan
alias: l
@@ -621,60 +589,35 @@ STAGE PLANS:
Filter Operator
predicate: locid is not null (type: boolean)
Statistics: Num rows: 8 Data size: 804 Basic stats: COMPLETE Column stats: COMPLETE
- Spark HashTable Sink Operator
- condition expressions:
- 0 {lastname} {deptid} {locid}
- 1 {deptid} {deptname}
- 2 {state} {zip} {year}
- keys:
- 0 deptid (type: int)
- 1 deptid (type: int)
- 2 locid (type: int)
- Local Work:
- Map Reduce Local Work
-
- Stage: Stage-1
- Spark
-#### A masked pattern was here ####
- Vertices:
- Map 1
- Map Operator Tree:
- TableScan
- alias: e
- Statistics: Num rows: 48 Data size: 552 Basic stats: COMPLETE Column stats: COMPLETE
- Filter Operator
- predicate: deptid is not null (type: boolean)
- Statistics: Num rows: 48 Data size: 4752 Basic stats: COMPLETE Column stats: COMPLETE
- Map Join Operator
- condition map:
- Inner Join 0 to 1
- Inner Join 0 to 2
- condition expressions:
- 0 {lastname} {deptid} {locid}
- 1 {deptid} {deptname}
- 2 {state} {locid} {zip} {year}
- keys:
- 0 deptid (type: int)
- 1 deptid (type: int)
- 2 locid (type: int)
- outputColumnNames: _col0, _col1, _col2, _col6, _col7, _col11, _col12, _col13, _col14
- input vertices:
- 1 Map 2
- 2 Map 3
- Statistics: Num rows: 47 Data size: 13912 Basic stats: COMPLETE Column stats: COMPLETE
- Select Operator
- expressions: _col0 (type: string), _col1 (type: int), _col2 (type: int), _col6 (type: int), _col7 (type: string), _col11 (type: string), _col12 (type: int), _col13 (type: bigint), _col14 (type: int)
- outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8
- Statistics: Num rows: 47 Data size: 13912 Basic stats: COMPLETE Column stats: COMPLETE
- File Output Operator
- compressed: false
- Statistics: Num rows: 47 Data size: 13912 Basic stats: COMPLETE Column stats: COMPLETE
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- Local Work:
- Map Reduce Local Work
+ Reduce Output Operator
+ key expressions: locid (type: int)
+ sort order: +
+ Map-reduce partition columns: locid (type: int)
+ Statistics: Num rows: 8 Data size: 804 Basic stats: COMPLETE Column stats: COMPLETE
+ value expressions: state (type: string), zip (type: bigint), year (type: int)
+ Reducer 2
+ Reduce Operator Tree:
+ Join Operator
+ condition map:
+ Inner Join 0 to 1
+ Inner Join 0 to 2
+ condition expressions:
+ 0 {VALUE._col0} {KEY.reducesinkkey0} {VALUE._col1}
+ 1 {KEY.reducesinkkey0} {VALUE._col0}
+ 2 {VALUE._col0} {KEY.reducesinkkey0} {VALUE._col1} {VALUE._col2}
+ outputColumnNames: _col0, _col1, _col2, _col6, _col7, _col11, _col12, _col13, _col14
+ Statistics: Num rows: 47 Data size: 13912 Basic stats: COMPLETE Column stats: COMPLETE
+ Select Operator
+ expressions: _col0 (type: string), _col1 (type: int), _col2 (type: int), _col6 (type: int), _col7 (type: string), _col11 (type: string), _col12 (type: int), _col13 (type: bigint), _col14 (type: int)
+ outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8
+ Statistics: Num rows: 47 Data size: 13912 Basic stats: COMPLETE Column stats: COMPLETE
+ File Output Operator
+ compressed: false
+ Statistics: Num rows: 47 Data size: 13912 Basic stats: COMPLETE Column stats: COMPLETE
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
Stage: Stage-0
Fetch Operator
@@ -691,16 +634,31 @@ POSTHOOK: query: -- 3 relations and 2 at
explain select * from emp e join dept d on (e.deptid = d.deptid and e.lastname = d.deptname) join loc l on (e.deptid = l.locid and e.lastname = l.state)
POSTHOOK: type: QUERY
STAGE DEPENDENCIES:
- Stage-2 is a root stage
- Stage-1 depends on stages: Stage-2
+ Stage-1 is a root stage
Stage-0 depends on stages: Stage-1
STAGE PLANS:
- Stage: Stage-2
+ Stage: Stage-1
Spark
+ Edges:
+ Reducer 2 <- Map 1 (PARTITION-LEVEL SORT, 3), Map 3 (PARTITION-LEVEL SORT, 3), Map 4 (PARTITION-LEVEL SORT, 3)
#### A masked pattern was here ####
Vertices:
- Map 2
+ Map 1
+ Map Operator Tree:
+ TableScan
+ alias: e
+ Statistics: Num rows: 48 Data size: 552 Basic stats: COMPLETE Column stats: COMPLETE
+ Filter Operator
+ predicate: (deptid is not null and lastname is not null) (type: boolean)
+ Statistics: Num rows: 48 Data size: 4752 Basic stats: COMPLETE Column stats: COMPLETE
+ Reduce Output Operator
+ key expressions: deptid (type: int), lastname (type: string)
+ sort order: ++
+ Map-reduce partition columns: deptid (type: int), lastname (type: string)
+ Statistics: Num rows: 48 Data size: 4752 Basic stats: COMPLETE Column stats: COMPLETE
+ value expressions: locid (type: int)
+ Map 3
Map Operator Tree:
TableScan
alias: d
@@ -708,18 +666,12 @@ STAGE PLANS:
Filter Operator
predicate: (deptid is not null and deptname is not null) (type: boolean)
Statistics: Num rows: 6 Data size: 570 Basic stats: COMPLETE Column stats: COMPLETE
- Spark HashTable Sink Operator
- condition expressions:
- 0 {lastname} {deptid} {locid}
- 1
- 2 {state} {locid} {zip} {year}
- keys:
- 0 deptid (type: int), lastname (type: string)
- 1 deptid (type: int), deptname (type: string)
- 2 locid (type: int), state (type: string)
- Local Work:
- Map Reduce Local Work
- Map 3
+ Reduce Output Operator
+ key expressions: deptid (type: int), deptname (type: string)
+ sort order: ++
+ Map-reduce partition columns: deptid (type: int), deptname (type: string)
+ Statistics: Num rows: 6 Data size: 570 Basic stats: COMPLETE Column stats: COMPLETE
+ Map 4
Map Operator Tree:
TableScan
alias: l
@@ -727,60 +679,35 @@ STAGE PLANS:
Filter Operator
predicate: (locid is not null and state is not null) (type: boolean)
Statistics: Num rows: 8 Data size: 804 Basic stats: COMPLETE Column stats: COMPLETE
- Spark HashTable Sink Operator
- condition expressions:
- 0 {lastname} {deptid} {locid}
- 1 {deptid} {deptname}
- 2 {zip} {year}
- keys:
- 0 deptid (type: int), lastname (type: string)
- 1 deptid (type: int), deptname (type: string)
- 2 locid (type: int), state (type: string)
- Local Work:
- Map Reduce Local Work
-
- Stage: Stage-1
- Spark
-#### A masked pattern was here ####
- Vertices:
- Map 1
- Map Operator Tree:
- TableScan
- alias: e
- Statistics: Num rows: 48 Data size: 552 Basic stats: COMPLETE Column stats: COMPLETE
- Filter Operator
- predicate: (deptid is not null and lastname is not null) (type: boolean)
- Statistics: Num rows: 48 Data size: 4752 Basic stats: COMPLETE Column stats: COMPLETE
- Map Join Operator
- condition map:
- Inner Join 0 to 1
- Inner Join 0 to 2
- condition expressions:
- 0 {lastname} {deptid} {locid}
- 1 {deptid} {deptname}
- 2 {state} {locid} {zip} {year}
- keys:
- 0 deptid (type: int), lastname (type: string)
- 1 deptid (type: int), deptname (type: string)
- 2 locid (type: int), state (type: string)
- outputColumnNames: _col0, _col1, _col2, _col6, _col7, _col11, _col12, _col13, _col14
- input vertices:
- 1 Map 2
- 2 Map 3
- Statistics: Num rows: 1 Data size: 296 Basic stats: COMPLETE Column stats: COMPLETE
- Select Operator
- expressions: _col0 (type: string), _col1 (type: int), _col2 (type: int), _col6 (type: int), _col7 (type: string), _col11 (type: string), _col12 (type: int), _col13 (type: bigint), _col14 (type: int)
- outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8
- Statistics: Num rows: 1 Data size: 296 Basic stats: COMPLETE Column stats: COMPLETE
- File Output Operator
- compressed: false
- Statistics: Num rows: 1 Data size: 296 Basic stats: COMPLETE Column stats: COMPLETE
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- Local Work:
- Map Reduce Local Work
+ Reduce Output Operator
+ key expressions: locid (type: int), state (type: string)
+ sort order: ++
+ Map-reduce partition columns: locid (type: int), state (type: string)
+ Statistics: Num rows: 8 Data size: 804 Basic stats: COMPLETE Column stats: COMPLETE
+ value expressions: zip (type: bigint), year (type: int)
+ Reducer 2
+ Reduce Operator Tree:
+ Join Operator
+ condition map:
+ Inner Join 0 to 1
+ Inner Join 0 to 2
+ condition expressions:
+ 0 {KEY.reducesinkkey1} {KEY.reducesinkkey0} {VALUE._col0}
+ 1 {KEY.reducesinkkey0} {KEY.reducesinkkey1}
+ 2 {KEY.reducesinkkey1} {KEY.reducesinkkey0} {VALUE._col0} {VALUE._col1}
+ outputColumnNames: _col0, _col1, _col2, _col6, _col7, _col11, _col12, _col13, _col14
+ Statistics: Num rows: 1 Data size: 296 Basic stats: COMPLETE Column stats: COMPLETE
+ Select Operator
+ expressions: _col0 (type: string), _col1 (type: int), _col2 (type: int), _col6 (type: int), _col7 (type: string), _col11 (type: string), _col12 (type: int), _col13 (type: bigint), _col14 (type: int)
+ outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8
+ Statistics: Num rows: 1 Data size: 296 Basic stats: COMPLETE Column stats: COMPLETE
+ File Output Operator
+ compressed: false
+ Statistics: Num rows: 1 Data size: 296 Basic stats: COMPLETE Column stats: COMPLETE
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
Stage: Stage-0
Fetch Operator
Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/auto_join_reordering_values.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/auto_join_reordering_values.q.out?rev=1642997&r1=1642996&r2=1642997&view=diff
==============================================================================
Files hive/branches/spark/ql/src/test/results/clientpositive/spark/auto_join_reordering_values.q.out (original) and hive/branches/spark/ql/src/test/results/clientpositive/spark/auto_join_reordering_values.q.out Tue Dec 2 19:57:10 2014 differ
Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/auto_smb_mapjoin_14.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/auto_smb_mapjoin_14.q.out?rev=1642997&r1=1642996&r2=1642997&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/spark/auto_smb_mapjoin_14.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/spark/auto_smb_mapjoin_14.q.out Tue Dec 2 19:57:10 2014
@@ -929,16 +929,18 @@ select count(*) from
on subq1.key = subq2.key
POSTHOOK: type: QUERY
STAGE DEPENDENCIES:
- Stage-2 is a root stage
- Stage-1 depends on stages: Stage-2
+ Stage-1 is a root stage
Stage-0 depends on stages: Stage-1
STAGE PLANS:
- Stage: Stage-2
+ Stage: Stage-1
Spark
+ Edges:
+ Reducer 2 <- Map 1 (PARTITION-LEVEL SORT, 3), Map 4 (PARTITION-LEVEL SORT, 3)
+ Reducer 3 <- Reducer 2 (GROUP, 1)
#### A masked pattern was here ####
Vertices:
- Map 3
+ Map 1
Map Operator Tree:
TableScan
alias: a
@@ -950,23 +952,12 @@ STAGE PLANS:
Filter Operator
predicate: _col0 is not null (type: boolean)
Statistics: Num rows: 5 Data size: 35 Basic stats: COMPLETE Column stats: NONE
- Spark HashTable Sink Operator
- condition expressions:
- 0
- 1
- keys:
- 0 _col0 (type: int)
- 1 _col0 (type: int)
- Local Work:
- Map Reduce Local Work
-
- Stage: Stage-1
- Spark
- Edges:
- Reducer 2 <- Map 1 (GROUP, 1)
-#### A masked pattern was here ####
- Vertices:
- Map 1
+ Reduce Output Operator
+ key expressions: _col0 (type: int)
+ sort order: +
+ Map-reduce partition columns: _col0 (type: int)
+ Statistics: Num rows: 5 Data size: 35 Basic stats: COMPLETE Column stats: NONE
+ Map 4
Map Operator Tree:
TableScan
alias: a
@@ -978,31 +969,31 @@ STAGE PLANS:
Filter Operator
predicate: _col0 is not null (type: boolean)
Statistics: Num rows: 5 Data size: 35 Basic stats: COMPLETE Column stats: NONE
- Map Join Operator
- condition map:
- Inner Join 0 to 1
- condition expressions:
- 0
- 1
- keys:
- 0 _col0 (type: int)
- 1 _col0 (type: int)
- input vertices:
- 1 Map 3
- Statistics: Num rows: 5 Data size: 38 Basic stats: COMPLETE Column stats: NONE
- Group By Operator
- aggregations: count()
- mode: hash
- outputColumnNames: _col0
- Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE
- Reduce Output Operator
- sort order:
- Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE
- value expressions: _col0 (type: bigint)
- Local Work:
- Map Reduce Local Work
+ Reduce Output Operator
+ key expressions: _col0 (type: int)
+ sort order: +
+ Map-reduce partition columns: _col0 (type: int)
+ Statistics: Num rows: 5 Data size: 35 Basic stats: COMPLETE Column stats: NONE
Reducer 2
Reduce Operator Tree:
+ Join Operator
+ condition map:
+ Inner Join 0 to 1
+ condition expressions:
+ 0
+ 1
+ Statistics: Num rows: 5 Data size: 38 Basic stats: COMPLETE Column stats: NONE
+ Group By Operator
+ aggregations: count()
+ mode: hash
+ outputColumnNames: _col0
+ Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE
+ Reduce Output Operator
+ sort order:
+ Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE
+ value expressions: _col0 (type: bigint)
+ Reducer 3
+ Reduce Operator Tree:
Group By Operator
aggregations: count(VALUE._col0)
mode: mergepartial
Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/bucket_map_join_1.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/bucket_map_join_1.q.out?rev=1642997&r1=1642996&r2=1642997&view=diff
==============================================================================
Files hive/branches/spark/ql/src/test/results/clientpositive/spark/bucket_map_join_1.q.out (original) and hive/branches/spark/ql/src/test/results/clientpositive/spark/bucket_map_join_1.q.out Tue Dec 2 19:57:10 2014 differ
Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/bucket_map_join_2.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/bucket_map_join_2.q.out?rev=1642997&r1=1642996&r2=1642997&view=diff
==============================================================================
Files hive/branches/spark/ql/src/test/results/clientpositive/spark/bucket_map_join_2.q.out (original) and hive/branches/spark/ql/src/test/results/clientpositive/spark/bucket_map_join_2.q.out Tue Dec 2 19:57:10 2014 differ
Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin1.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin1.q.out?rev=1642997&r1=1642996&r2=1642997&view=diff
==============================================================================
Files hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin1.q.out (original) and hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin1.q.out Tue Dec 2 19:57:10 2014 differ
Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin10.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin10.q.out?rev=1642997&r1=1642996&r2=1642997&view=diff
==============================================================================
Files hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin10.q.out (original) and hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin10.q.out Tue Dec 2 19:57:10 2014 differ
Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin11.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin11.q.out?rev=1642997&r1=1642996&r2=1642997&view=diff
==============================================================================
Files hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin11.q.out (original) and hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin11.q.out Tue Dec 2 19:57:10 2014 differ
Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin12.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin12.q.out?rev=1642997&r1=1642996&r2=1642997&view=diff
==============================================================================
Files hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin12.q.out (original) and hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin12.q.out Tue Dec 2 19:57:10 2014 differ
Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin13.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin13.q.out?rev=1642997&r1=1642996&r2=1642997&view=diff
==============================================================================
Files hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin13.q.out (original) and hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin13.q.out Tue Dec 2 19:57:10 2014 differ
Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin2.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin2.q.out?rev=1642997&r1=1642996&r2=1642997&view=diff
==============================================================================
Files hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin2.q.out (original) and hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin2.q.out Tue Dec 2 19:57:10 2014 differ
Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin3.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin3.q.out?rev=1642997&r1=1642996&r2=1642997&view=diff
==============================================================================
Files hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin3.q.out (original) and hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin3.q.out Tue Dec 2 19:57:10 2014 differ
Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin4.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin4.q.out?rev=1642997&r1=1642996&r2=1642997&view=diff
==============================================================================
Files hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin4.q.out (original) and hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin4.q.out Tue Dec 2 19:57:10 2014 differ
Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin5.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin5.q.out?rev=1642997&r1=1642996&r2=1642997&view=diff
==============================================================================
Files hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin5.q.out (original) and hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin5.q.out Tue Dec 2 19:57:10 2014 differ
Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin7.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin7.q.out?rev=1642997&r1=1642996&r2=1642997&view=diff
==============================================================================
Files hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin7.q.out (original) and hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin7.q.out Tue Dec 2 19:57:10 2014 differ
Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin8.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin8.q.out?rev=1642997&r1=1642996&r2=1642997&view=diff
==============================================================================
Files hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin8.q.out (original) and hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin8.q.out Tue Dec 2 19:57:10 2014 differ
Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin9.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin9.q.out?rev=1642997&r1=1642996&r2=1642997&view=diff
==============================================================================
Files hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin9.q.out (original) and hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin9.q.out Tue Dec 2 19:57:10 2014 differ
Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin_negative.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin_negative.q.out?rev=1642997&r1=1642996&r2=1642997&view=diff
==============================================================================
Files hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin_negative.q.out (original) and hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin_negative.q.out Tue Dec 2 19:57:10 2014 differ
Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin_negative2.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin_negative2.q.out?rev=1642997&r1=1642996&r2=1642997&view=diff
==============================================================================
Files hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin_negative2.q.out (original) and hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin_negative2.q.out Tue Dec 2 19:57:10 2014 differ
Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin_negative3.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin_negative3.q.out?rev=1642997&r1=1642996&r2=1642997&view=diff
==============================================================================
Files hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin_negative3.q.out (original) and hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin_negative3.q.out Tue Dec 2 19:57:10 2014 differ