You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hive.apache.org by br...@apache.org on 2014/11/29 04:44:28 UTC
svn commit: r1642395 [15/22] - in /hive/branches/spark/ql/src:
java/org/apache/hadoop/hive/ql/exec/spark/
java/org/apache/hadoop/hive/ql/exec/spark/session/
test/results/clientpositive/ test/results/clientpositive/spark/
Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/semijoin.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/semijoin.q.out?rev=1642395&r1=1642394&r2=1642395&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/spark/semijoin.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/spark/semijoin.q.out Sat Nov 29 03:44:22 2014
@@ -128,7 +128,7 @@ STAGE PLANS:
Spark
#### A masked pattern was here ####
Vertices:
- Map 1
+ Map 3
Map Operator Tree:
TableScan
alias: b
@@ -158,10 +158,10 @@ STAGE PLANS:
Stage: Stage-1
Spark
Edges:
- Reducer 3 <- Map 2 (SORT, 3)
+ Reducer 2 <- Map 1 (SORT, 3)
#### A masked pattern was here ####
Vertices:
- Map 2
+ Map 1
Map Operator Tree:
TableScan
alias: a
@@ -180,19 +180,15 @@ STAGE PLANS:
1 _col0 (type: int)
outputColumnNames: _col0, _col1
input vertices:
- 1 Map 1
+ 1 Map 3
Statistics: Num rows: 6 Data size: 47 Basic stats: COMPLETE Column stats: NONE
- Select Operator
- expressions: _col0 (type: int), _col1 (type: string)
- outputColumnNames: _col0, _col1
+ Reduce Output Operator
+ key expressions: _col0 (type: int), _col1 (type: string)
+ sort order: ++
Statistics: Num rows: 6 Data size: 47 Basic stats: COMPLETE Column stats: NONE
- Reduce Output Operator
- key expressions: _col0 (type: int), _col1 (type: string)
- sort order: ++
- Statistics: Num rows: 6 Data size: 47 Basic stats: COMPLETE Column stats: NONE
Local Work:
Map Reduce Local Work
- Reducer 3
+ Reducer 2
Reduce Operator Tree:
Select Operator
expressions: KEY.reducesinkkey0 (type: int), KEY.reducesinkkey1 (type: string)
@@ -242,7 +238,7 @@ STAGE PLANS:
Spark
#### A masked pattern was here ####
Vertices:
- Map 1
+ Map 3
Map Operator Tree:
TableScan
alias: b
@@ -272,10 +268,10 @@ STAGE PLANS:
Stage: Stage-1
Spark
Edges:
- Reducer 3 <- Map 2 (SORT, 3)
+ Reducer 2 <- Map 1 (SORT, 3)
#### A masked pattern was here ####
Vertices:
- Map 2
+ Map 1
Map Operator Tree:
TableScan
alias: a
@@ -294,19 +290,15 @@ STAGE PLANS:
1 _col0 (type: int)
outputColumnNames: _col0, _col1
input vertices:
- 1 Map 1
+ 1 Map 3
Statistics: Num rows: 6 Data size: 49 Basic stats: COMPLETE Column stats: NONE
- Select Operator
- expressions: _col0 (type: int), _col1 (type: string)
- outputColumnNames: _col0, _col1
+ Reduce Output Operator
+ key expressions: _col0 (type: int), _col1 (type: string)
+ sort order: ++
Statistics: Num rows: 6 Data size: 49 Basic stats: COMPLETE Column stats: NONE
- Reduce Output Operator
- key expressions: _col0 (type: int), _col1 (type: string)
- sort order: ++
- Statistics: Num rows: 6 Data size: 49 Basic stats: COMPLETE Column stats: NONE
Local Work:
Map Reduce Local Work
- Reducer 3
+ Reducer 2
Reduce Operator Tree:
Select Operator
expressions: KEY.reducesinkkey0 (type: int), KEY.reducesinkkey1 (type: string)
@@ -358,7 +350,7 @@ STAGE PLANS:
Spark
#### A masked pattern was here ####
Vertices:
- Map 1
+ Map 3
Map Operator Tree:
TableScan
alias: b
@@ -388,10 +380,10 @@ STAGE PLANS:
Stage: Stage-1
Spark
Edges:
- Reducer 3 <- Map 2 (SORT, 3)
+ Reducer 2 <- Map 1 (SORT, 3)
#### A masked pattern was here ####
Vertices:
- Map 2
+ Map 1
Map Operator Tree:
TableScan
alias: a
@@ -410,19 +402,15 @@ STAGE PLANS:
1 _col0 (type: int)
outputColumnNames: _col0, _col1
input vertices:
- 1 Map 1
+ 1 Map 3
Statistics: Num rows: 6 Data size: 47 Basic stats: COMPLETE Column stats: NONE
- Select Operator
- expressions: _col0 (type: int), _col1 (type: string)
- outputColumnNames: _col0, _col1
+ Reduce Output Operator
+ key expressions: _col0 (type: int), _col1 (type: string)
+ sort order: ++
Statistics: Num rows: 6 Data size: 47 Basic stats: COMPLETE Column stats: NONE
- Reduce Output Operator
- key expressions: _col0 (type: int), _col1 (type: string)
- sort order: ++
- Statistics: Num rows: 6 Data size: 47 Basic stats: COMPLETE Column stats: NONE
Local Work:
Map Reduce Local Work
- Reducer 3
+ Reducer 2
Reduce Operator Tree:
Select Operator
expressions: KEY.reducesinkkey0 (type: int), KEY.reducesinkkey1 (type: string)
@@ -466,7 +454,7 @@ STAGE PLANS:
Spark
#### A masked pattern was here ####
Vertices:
- Map 1
+ Map 3
Map Operator Tree:
TableScan
alias: b
@@ -474,32 +462,28 @@ STAGE PLANS:
Filter Operator
predicate: (key < 15) (type: boolean)
Statistics: Num rows: 7 Data size: 51 Basic stats: COMPLETE Column stats: NONE
- Select Operator
- expressions: key (type: int)
- outputColumnNames: key
+ Group By Operator
+ keys: key (type: int), key (type: int)
+ mode: hash
+ outputColumnNames: _col0, _col1
Statistics: Num rows: 7 Data size: 51 Basic stats: COMPLETE Column stats: NONE
- Group By Operator
- keys: key (type: int), key (type: int)
- mode: hash
- outputColumnNames: _col0, _col1
- Statistics: Num rows: 7 Data size: 51 Basic stats: COMPLETE Column stats: NONE
- Spark HashTable Sink Operator
- condition expressions:
- 0 {value}
- 1
- keys:
- 0 key (type: int)
- 1 _col1 (type: int)
+ Spark HashTable Sink Operator
+ condition expressions:
+ 0 {value}
+ 1
+ keys:
+ 0 key (type: int)
+ 1 _col1 (type: int)
Local Work:
Map Reduce Local Work
Stage: Stage-1
Spark
Edges:
- Reducer 3 <- Map 2 (SORT, 3)
+ Reducer 2 <- Map 1 (SORT, 3)
#### A masked pattern was here ####
Vertices:
- Map 2
+ Map 1
Map Operator Tree:
TableScan
alias: a
@@ -518,7 +502,7 @@ STAGE PLANS:
1 _col1 (type: int)
outputColumnNames: _col1
input vertices:
- 1 Map 1
+ 1 Map 3
Statistics: Num rows: 7 Data size: 56 Basic stats: COMPLETE Column stats: NONE
Select Operator
expressions: _col1 (type: string)
@@ -530,7 +514,7 @@ STAGE PLANS:
Statistics: Num rows: 7 Data size: 56 Basic stats: COMPLETE Column stats: NONE
Local Work:
Map Reduce Local Work
- Reducer 3
+ Reducer 2
Reduce Operator Tree:
Select Operator
expressions: KEY.reducesinkkey0 (type: string)
@@ -585,7 +569,7 @@ STAGE PLANS:
Spark
#### A masked pattern was here ####
Vertices:
- Map 1
+ Map 3
Map Operator Tree:
TableScan
alias: b
@@ -615,10 +599,10 @@ STAGE PLANS:
Stage: Stage-1
Spark
Edges:
- Reducer 3 <- Map 2 (SORT, 3)
+ Reducer 2 <- Map 1 (SORT, 3)
#### A masked pattern was here ####
Vertices:
- Map 2
+ Map 1
Map Operator Tree:
TableScan
alias: a
@@ -637,19 +621,15 @@ STAGE PLANS:
1 _col0 (type: int)
outputColumnNames: _col0, _col1
input vertices:
- 1 Map 1
+ 1 Map 3
Statistics: Num rows: 6 Data size: 47 Basic stats: COMPLETE Column stats: NONE
- Select Operator
- expressions: _col0 (type: int), _col1 (type: string)
- outputColumnNames: _col0, _col1
+ Reduce Output Operator
+ key expressions: _col0 (type: int), _col1 (type: string)
+ sort order: ++
Statistics: Num rows: 6 Data size: 47 Basic stats: COMPLETE Column stats: NONE
- Reduce Output Operator
- key expressions: _col0 (type: int), _col1 (type: string)
- sort order: ++
- Statistics: Num rows: 6 Data size: 47 Basic stats: COMPLETE Column stats: NONE
Local Work:
Map Reduce Local Work
- Reducer 3
+ Reducer 2
Reduce Operator Tree:
Select Operator
expressions: KEY.reducesinkkey0 (type: int), KEY.reducesinkkey1 (type: string)
@@ -696,7 +676,7 @@ STAGE PLANS:
Spark
#### A masked pattern was here ####
Vertices:
- Map 3
+ Map 1
Map Operator Tree:
TableScan
alias: t3
@@ -726,10 +706,10 @@ STAGE PLANS:
Stage: Stage-1
Spark
Edges:
- Reducer 2 <- Map 1 (SORT, 3)
+ Reducer 3 <- Map 2 (SORT, 3)
#### A masked pattern was here ####
Vertices:
- Map 1
+ Map 2
Map Operator Tree:
TableScan
alias: a
@@ -748,7 +728,7 @@ STAGE PLANS:
1 _col0 (type: int)
outputColumnNames: _col1
input vertices:
- 1 Map 3
+ 1 Map 1
Statistics: Num rows: 7 Data size: 56 Basic stats: COMPLETE Column stats: NONE
Select Operator
expressions: _col1 (type: string)
@@ -760,7 +740,7 @@ STAGE PLANS:
Statistics: Num rows: 7 Data size: 56 Basic stats: COMPLETE Column stats: NONE
Local Work:
Map Reduce Local Work
- Reducer 2
+ Reducer 3
Reduce Operator Tree:
Select Operator
expressions: KEY.reducesinkkey0 (type: string)
@@ -969,14 +949,10 @@ STAGE PLANS:
input vertices:
1 Map 1
Statistics: Num rows: 6 Data size: 49 Basic stats: COMPLETE Column stats: NONE
- Select Operator
- expressions: _col0 (type: int), _col1 (type: string)
- outputColumnNames: _col0, _col1
+ Reduce Output Operator
+ key expressions: _col0 (type: int), _col1 (type: string)
+ sort order: ++
Statistics: Num rows: 6 Data size: 49 Basic stats: COMPLETE Column stats: NONE
- Reduce Output Operator
- key expressions: _col0 (type: int), _col1 (type: string)
- sort order: ++
- Statistics: Num rows: 6 Data size: 49 Basic stats: COMPLETE Column stats: NONE
Local Work:
Map Reduce Local Work
Reducer 3
@@ -1028,7 +1004,7 @@ STAGE PLANS:
Spark
#### A masked pattern was here ####
Vertices:
- Map 1
+ Map 3
Map Operator Tree:
TableScan
alias: b
@@ -1058,10 +1034,10 @@ STAGE PLANS:
Stage: Stage-1
Spark
Edges:
- Reducer 3 <- Map 2 (SORT, 3)
+ Reducer 2 <- Map 1 (SORT, 3)
#### A masked pattern was here ####
Vertices:
- Map 2
+ Map 1
Map Operator Tree:
TableScan
alias: a
@@ -1080,19 +1056,15 @@ STAGE PLANS:
1 _col0 (type: int)
outputColumnNames: _col0
input vertices:
- 1 Map 1
+ 1 Map 3
Statistics: Num rows: 12 Data size: 89 Basic stats: COMPLETE Column stats: NONE
- Select Operator
- expressions: _col0 (type: int)
- outputColumnNames: _col0
+ Reduce Output Operator
+ key expressions: _col0 (type: int)
+ sort order: +
Statistics: Num rows: 12 Data size: 89 Basic stats: COMPLETE Column stats: NONE
- Reduce Output Operator
- key expressions: _col0 (type: int)
- sort order: +
- Statistics: Num rows: 12 Data size: 89 Basic stats: COMPLETE Column stats: NONE
Local Work:
Map Reduce Local Work
- Reducer 3
+ Reducer 2
Reduce Operator Tree:
Select Operator
expressions: KEY.reducesinkkey0 (type: int)
@@ -1155,7 +1127,7 @@ STAGE PLANS:
Spark
#### A masked pattern was here ####
Vertices:
- Map 1
+ Map 3
Map Operator Tree:
TableScan
alias: b
@@ -1185,10 +1157,10 @@ STAGE PLANS:
Stage: Stage-1
Spark
Edges:
- Reducer 3 <- Map 2 (SORT, 3)
+ Reducer 2 <- Map 1 (SORT, 3)
#### A masked pattern was here ####
Vertices:
- Map 2
+ Map 1
Map Operator Tree:
TableScan
alias: a
@@ -1207,19 +1179,15 @@ STAGE PLANS:
1 (2 * _col0) (type: int)
outputColumnNames: _col0, _col1
input vertices:
- 1 Map 1
+ 1 Map 3
Statistics: Num rows: 6 Data size: 47 Basic stats: COMPLETE Column stats: NONE
- Select Operator
- expressions: _col0 (type: int), _col1 (type: string)
- outputColumnNames: _col0, _col1
+ Reduce Output Operator
+ key expressions: _col0 (type: int), _col1 (type: string)
+ sort order: ++
Statistics: Num rows: 6 Data size: 47 Basic stats: COMPLETE Column stats: NONE
- Reduce Output Operator
- key expressions: _col0 (type: int), _col1 (type: string)
- sort order: ++
- Statistics: Num rows: 6 Data size: 47 Basic stats: COMPLETE Column stats: NONE
Local Work:
Map Reduce Local Work
- Reducer 3
+ Reducer 2
Reduce Operator Tree:
Select Operator
expressions: KEY.reducesinkkey0 (type: int), KEY.reducesinkkey1 (type: string)
@@ -1267,7 +1235,26 @@ STAGE PLANS:
Spark
#### A masked pattern was here ####
Vertices:
- Map 3
+ Map 1
+ Map Operator Tree:
+ TableScan
+ alias: a
+ Statistics: Num rows: 11 Data size: 79 Basic stats: COMPLETE Column stats: NONE
+ Filter Operator
+ predicate: key is not null (type: boolean)
+ Statistics: Num rows: 6 Data size: 43 Basic stats: COMPLETE Column stats: NONE
+ Spark HashTable Sink Operator
+ condition expressions:
+ 0 {value}
+ 1 {key} {value}
+ 2
+ keys:
+ 0 key (type: int)
+ 1 key (type: int)
+ 2 _col0 (type: int)
+ Local Work:
+ Map Reduce Local Work
+ Map 4
Map Operator Tree:
TableScan
alias: c
@@ -1295,33 +1282,14 @@ STAGE PLANS:
2 _col0 (type: int)
Local Work:
Map Reduce Local Work
- Map 4
- Map Operator Tree:
- TableScan
- alias: a
- Statistics: Num rows: 11 Data size: 79 Basic stats: COMPLETE Column stats: NONE
- Filter Operator
- predicate: key is not null (type: boolean)
- Statistics: Num rows: 6 Data size: 43 Basic stats: COMPLETE Column stats: NONE
- Spark HashTable Sink Operator
- condition expressions:
- 0 {value}
- 1 {key} {value}
- 2
- keys:
- 0 key (type: int)
- 1 key (type: int)
- 2 _col0 (type: int)
- Local Work:
- Map Reduce Local Work
Stage: Stage-1
Spark
Edges:
- Reducer 2 <- Map 1 (SORT, 3)
+ Reducer 3 <- Map 2 (SORT, 3)
#### A masked pattern was here ####
Vertices:
- Map 1
+ Map 2
Map Operator Tree:
TableScan
alias: b
@@ -1343,8 +1311,8 @@ STAGE PLANS:
2 _col0 (type: int)
outputColumnNames: _col0, _col1, _col5, _col6
input vertices:
- 0 Map 4
- 2 Map 3
+ 0 Map 1
+ 2 Map 4
Statistics: Num rows: 24 Data size: 178 Basic stats: COMPLETE Column stats: NONE
Select Operator
expressions: _col0 (type: int), _col1 (type: string), _col5 (type: int), _col6 (type: string)
@@ -1357,7 +1325,7 @@ STAGE PLANS:
value expressions: _col2 (type: int), _col3 (type: string)
Local Work:
Map Reduce Local Work
- Reducer 2
+ Reducer 3
Reduce Operator Tree:
Select Operator
expressions: KEY.reducesinkkey0 (type: int), KEY.reducesinkkey1 (type: string), VALUE._col0 (type: int), VALUE._col1 (type: string)
@@ -1417,7 +1385,7 @@ STAGE PLANS:
Spark
#### A masked pattern was here ####
Vertices:
- Map 1
+ Map 3
Map Operator Tree:
TableScan
alias: b
@@ -1447,10 +1415,10 @@ STAGE PLANS:
Stage: Stage-1
Spark
Edges:
- Reducer 3 <- Map 2 (SORT, 3)
+ Reducer 2 <- Map 1 (SORT, 3)
#### A masked pattern was here ####
Vertices:
- Map 2
+ Map 1
Map Operator Tree:
TableScan
alias: a
@@ -1469,19 +1437,15 @@ STAGE PLANS:
1 _col0 (type: int), _col1 (type: string)
outputColumnNames: _col0, _col1
input vertices:
- 1 Map 1
+ 1 Map 3
Statistics: Num rows: 6 Data size: 48 Basic stats: COMPLETE Column stats: NONE
- Select Operator
- expressions: _col0 (type: int), _col1 (type: string)
- outputColumnNames: _col0, _col1
+ Reduce Output Operator
+ key expressions: _col0 (type: int), _col1 (type: string)
+ sort order: ++
Statistics: Num rows: 6 Data size: 48 Basic stats: COMPLETE Column stats: NONE
- Reduce Output Operator
- key expressions: _col0 (type: int), _col1 (type: string)
- sort order: ++
- Statistics: Num rows: 6 Data size: 48 Basic stats: COMPLETE Column stats: NONE
Local Work:
Map Reduce Local Work
- Reducer 3
+ Reducer 2
Reduce Operator Tree:
Select Operator
expressions: KEY.reducesinkkey0 (type: int), KEY.reducesinkkey1 (type: string)
@@ -1539,7 +1503,7 @@ STAGE PLANS:
Spark
#### A masked pattern was here ####
Vertices:
- Map 1
+ Map 3
Map Operator Tree:
TableScan
alias: b
@@ -1567,7 +1531,7 @@ STAGE PLANS:
2 _col0 (type: int)
Local Work:
Map Reduce Local Work
- Map 2
+ Map 4
Map Operator Tree:
TableScan
alias: c
@@ -1599,10 +1563,10 @@ STAGE PLANS:
Stage: Stage-1
Spark
Edges:
- Reducer 4 <- Map 3 (SORT, 3)
+ Reducer 2 <- Map 1 (SORT, 3)
#### A masked pattern was here ####
Vertices:
- Map 3
+ Map 1
Map Operator Tree:
TableScan
alias: a
@@ -1624,20 +1588,16 @@ STAGE PLANS:
2 _col0 (type: int)
outputColumnNames: _col0
input vertices:
- 1 Map 1
- 2 Map 2
+ 1 Map 3
+ 2 Map 4
Statistics: Num rows: 24 Data size: 178 Basic stats: COMPLETE Column stats: NONE
- Select Operator
- expressions: _col0 (type: int)
- outputColumnNames: _col0
+ Reduce Output Operator
+ key expressions: _col0 (type: int)
+ sort order: +
Statistics: Num rows: 24 Data size: 178 Basic stats: COMPLETE Column stats: NONE
- Reduce Output Operator
- key expressions: _col0 (type: int)
- sort order: +
- Statistics: Num rows: 24 Data size: 178 Basic stats: COMPLETE Column stats: NONE
Local Work:
Map Reduce Local Work
- Reducer 4
+ Reducer 2
Reduce Operator Tree:
Select Operator
expressions: KEY.reducesinkkey0 (type: int)
@@ -1697,7 +1657,7 @@ STAGE PLANS:
Spark
#### A masked pattern was here ####
Vertices:
- Map 1
+ Map 3
Map Operator Tree:
TableScan
alias: b
@@ -1713,7 +1673,7 @@ STAGE PLANS:
2 _col0 (type: int)
Local Work:
Map Reduce Local Work
- Map 2
+ Map 4
Map Operator Tree:
TableScan
alias: c
@@ -1742,10 +1702,10 @@ STAGE PLANS:
Stage: Stage-1
Spark
Edges:
- Reducer 4 <- Map 3 (SORT, 3)
+ Reducer 2 <- Map 1 (SORT, 3)
#### A masked pattern was here ####
Vertices:
- Map 3
+ Map 1
Map Operator Tree:
TableScan
alias: a
@@ -1764,20 +1724,16 @@ STAGE PLANS:
2 _col0 (type: int)
outputColumnNames: _col0
input vertices:
- 1 Map 1
- 2 Map 2
+ 1 Map 3
+ 2 Map 4
Statistics: Num rows: 48 Data size: 358 Basic stats: COMPLETE Column stats: NONE
- Select Operator
- expressions: _col0 (type: int)
- outputColumnNames: _col0
+ Reduce Output Operator
+ key expressions: _col0 (type: int)
+ sort order: +
Statistics: Num rows: 48 Data size: 358 Basic stats: COMPLETE Column stats: NONE
- Reduce Output Operator
- key expressions: _col0 (type: int)
- sort order: +
- Statistics: Num rows: 48 Data size: 358 Basic stats: COMPLETE Column stats: NONE
Local Work:
Map Reduce Local Work
- Reducer 4
+ Reducer 2
Reduce Operator Tree:
Select Operator
expressions: KEY.reducesinkkey0 (type: int)
@@ -1849,7 +1805,23 @@ STAGE PLANS:
Spark
#### A masked pattern was here ####
Vertices:
- Map 3
+ Map 1
+ Map Operator Tree:
+ TableScan
+ alias: a
+ Statistics: Num rows: 11 Data size: 79 Basic stats: COMPLETE Column stats: NONE
+ Spark HashTable Sink Operator
+ condition expressions:
+ 0
+ 1
+ 2
+ keys:
+ 0 key (type: int)
+ 1 key (type: int)
+ 2 _col0 (type: int)
+ Local Work:
+ Map Reduce Local Work
+ Map 4
Map Operator Tree:
TableScan
alias: c
@@ -1874,30 +1846,14 @@ STAGE PLANS:
2 _col0 (type: int)
Local Work:
Map Reduce Local Work
- Map 4
- Map Operator Tree:
- TableScan
- alias: a
- Statistics: Num rows: 11 Data size: 79 Basic stats: COMPLETE Column stats: NONE
- Spark HashTable Sink Operator
- condition expressions:
- 0
- 1
- 2
- keys:
- 0 key (type: int)
- 1 key (type: int)
- 2 _col0 (type: int)
- Local Work:
- Map Reduce Local Work
Stage: Stage-1
Spark
Edges:
- Reducer 2 <- Map 1 (SORT, 3)
+ Reducer 3 <- Map 2 (SORT, 3)
#### A masked pattern was here ####
Vertices:
- Map 1
+ Map 2
Map Operator Tree:
TableScan
alias: b
@@ -1916,20 +1872,16 @@ STAGE PLANS:
2 _col0 (type: int)
outputColumnNames: _col0
input vertices:
- 0 Map 4
- 2 Map 3
+ 0 Map 1
+ 2 Map 4
Statistics: Num rows: 48 Data size: 358 Basic stats: COMPLETE Column stats: NONE
- Select Operator
- expressions: _col0 (type: int)
- outputColumnNames: _col0
+ Reduce Output Operator
+ key expressions: _col0 (type: int)
+ sort order: +
Statistics: Num rows: 48 Data size: 358 Basic stats: COMPLETE Column stats: NONE
- Reduce Output Operator
- key expressions: _col0 (type: int)
- sort order: +
- Statistics: Num rows: 48 Data size: 358 Basic stats: COMPLETE Column stats: NONE
Local Work:
Map Reduce Local Work
- Reducer 2
+ Reducer 3
Reduce Operator Tree:
Select Operator
expressions: KEY.reducesinkkey0 (type: int)
@@ -2009,6 +1961,16 @@ STAGE PLANS:
Map 1
Map Operator Tree:
TableScan
+ alias: a
+ Statistics: Num rows: 11 Data size: 79 Basic stats: COMPLETE Column stats: NONE
+ Reduce Output Operator
+ key expressions: key (type: int)
+ sort order: +
+ Map-reduce partition columns: key (type: int)
+ Statistics: Num rows: 11 Data size: 79 Basic stats: COMPLETE Column stats: NONE
+ Map 4
+ Map Operator Tree:
+ TableScan
alias: b
Statistics: Num rows: 22 Data size: 163 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
@@ -2016,7 +1978,7 @@ STAGE PLANS:
sort order: +
Map-reduce partition columns: key (type: int)
Statistics: Num rows: 22 Data size: 163 Basic stats: COMPLETE Column stats: NONE
- Map 4
+ Map 5
Map Operator Tree:
TableScan
alias: c
@@ -2035,16 +1997,6 @@ STAGE PLANS:
sort order: +
Map-reduce partition columns: _col0 (type: int)
Statistics: Num rows: 11 Data size: 84 Basic stats: COMPLETE Column stats: NONE
- Map 5
- Map Operator Tree:
- TableScan
- alias: a
- Statistics: Num rows: 11 Data size: 79 Basic stats: COMPLETE Column stats: NONE
- Reduce Output Operator
- key expressions: key (type: int)
- sort order: +
- Map-reduce partition columns: key (type: int)
- Statistics: Num rows: 11 Data size: 79 Basic stats: COMPLETE Column stats: NONE
Reducer 2
Reduce Operator Tree:
Join Operator
@@ -2057,14 +2009,10 @@ STAGE PLANS:
2
outputColumnNames: _col0
Statistics: Num rows: 48 Data size: 358 Basic stats: COMPLETE Column stats: NONE
- Select Operator
- expressions: _col0 (type: int)
- outputColumnNames: _col0
+ Reduce Output Operator
+ key expressions: _col0 (type: int)
+ sort order: +
Statistics: Num rows: 48 Data size: 358 Basic stats: COMPLETE Column stats: NONE
- Reduce Output Operator
- key expressions: _col0 (type: int)
- sort order: +
- Statistics: Num rows: 48 Data size: 358 Basic stats: COMPLETE Column stats: NONE
Reducer 3
Reduce Operator Tree:
Select Operator
@@ -2140,7 +2088,7 @@ STAGE PLANS:
Spark
#### A masked pattern was here ####
Vertices:
- Map 1
+ Map 3
Map Operator Tree:
TableScan
alias: b
@@ -2165,7 +2113,7 @@ STAGE PLANS:
2 key (type: int)
Local Work:
Map Reduce Local Work
- Map 2
+ Map 4
Map Operator Tree:
TableScan
alias: c
@@ -2185,10 +2133,10 @@ STAGE PLANS:
Stage: Stage-1
Spark
Edges:
- Reducer 4 <- Map 3 (SORT, 3)
+ Reducer 2 <- Map 1 (SORT, 3)
#### A masked pattern was here ####
Vertices:
- Map 3
+ Map 1
Map Operator Tree:
TableScan
alias: a
@@ -2207,20 +2155,16 @@ STAGE PLANS:
2 key (type: int)
outputColumnNames: _col0
input vertices:
- 1 Map 1
- 2 Map 2
+ 1 Map 3
+ 2 Map 4
Statistics: Num rows: 48 Data size: 358 Basic stats: COMPLETE Column stats: NONE
- Select Operator
- expressions: _col0 (type: int)
- outputColumnNames: _col0
+ Reduce Output Operator
+ key expressions: _col0 (type: int)
+ sort order: +
Statistics: Num rows: 48 Data size: 358 Basic stats: COMPLETE Column stats: NONE
- Reduce Output Operator
- key expressions: _col0 (type: int)
- sort order: +
- Statistics: Num rows: 48 Data size: 358 Basic stats: COMPLETE Column stats: NONE
Local Work:
Map Reduce Local Work
- Reducer 4
+ Reducer 2
Reduce Operator Tree:
Select Operator
expressions: KEY.reducesinkkey0 (type: int)
@@ -2298,6 +2242,22 @@ STAGE PLANS:
Map 1
Map Operator Tree:
TableScan
+ alias: a
+ Statistics: Num rows: 22 Data size: 163 Basic stats: COMPLETE Column stats: NONE
+ Spark HashTable Sink Operator
+ condition expressions:
+ 0
+ 1
+ 2
+ keys:
+ 0 key (type: int)
+ 1 _col0 (type: int)
+ 2 key (type: int)
+ Local Work:
+ Map Reduce Local Work
+ Map 2
+ Map Operator Tree:
+ TableScan
alias: b
Statistics: Num rows: 11 Data size: 84 Basic stats: COMPLETE Column stats: NONE
Select Operator
@@ -2320,30 +2280,14 @@ STAGE PLANS:
2 key (type: int)
Local Work:
Map Reduce Local Work
- Map 4
- Map Operator Tree:
- TableScan
- alias: a
- Statistics: Num rows: 22 Data size: 163 Basic stats: COMPLETE Column stats: NONE
- Spark HashTable Sink Operator
- condition expressions:
- 0
- 1
- 2
- keys:
- 0 key (type: int)
- 1 _col0 (type: int)
- 2 key (type: int)
- Local Work:
- Map Reduce Local Work
Stage: Stage-1
Spark
Edges:
- Reducer 3 <- Map 2 (SORT, 3)
+ Reducer 4 <- Map 3 (SORT, 3)
#### A masked pattern was here ####
Vertices:
- Map 2
+ Map 3
Map Operator Tree:
TableScan
alias: c
@@ -2362,20 +2306,16 @@ STAGE PLANS:
2 key (type: int)
outputColumnNames: _col0
input vertices:
- 0 Map 4
- 1 Map 1
+ 0 Map 1
+ 1 Map 2
Statistics: Num rows: 48 Data size: 358 Basic stats: COMPLETE Column stats: NONE
- Select Operator
- expressions: _col0 (type: int)
- outputColumnNames: _col0
+ Reduce Output Operator
+ key expressions: _col0 (type: int)
+ sort order: +
Statistics: Num rows: 48 Data size: 358 Basic stats: COMPLETE Column stats: NONE
- Reduce Output Operator
- key expressions: _col0 (type: int)
- sort order: +
- Statistics: Num rows: 48 Data size: 358 Basic stats: COMPLETE Column stats: NONE
Local Work:
Map Reduce Local Work
- Reducer 3
+ Reducer 4
Reduce Operator Tree:
Select Operator
expressions: KEY.reducesinkkey0 (type: int)
@@ -2457,6 +2397,16 @@ STAGE PLANS:
Map 1
Map Operator Tree:
TableScan
+ alias: a
+ Statistics: Num rows: 22 Data size: 163 Basic stats: COMPLETE Column stats: NONE
+ Reduce Output Operator
+ key expressions: key (type: int)
+ sort order: +
+ Map-reduce partition columns: key (type: int)
+ Statistics: Num rows: 22 Data size: 163 Basic stats: COMPLETE Column stats: NONE
+ Map 4
+ Map Operator Tree:
+ TableScan
alias: b
Statistics: Num rows: 11 Data size: 79 Basic stats: COMPLETE Column stats: NONE
Select Operator
@@ -2473,7 +2423,7 @@ STAGE PLANS:
sort order: +
Map-reduce partition columns: _col0 (type: int)
Statistics: Num rows: 11 Data size: 79 Basic stats: COMPLETE Column stats: NONE
- Map 4
+ Map 5
Map Operator Tree:
TableScan
alias: c
@@ -2483,16 +2433,6 @@ STAGE PLANS:
sort order: +
Map-reduce partition columns: key (type: int)
Statistics: Num rows: 11 Data size: 84 Basic stats: COMPLETE Column stats: NONE
- Map 5
- Map Operator Tree:
- TableScan
- alias: a
- Statistics: Num rows: 22 Data size: 163 Basic stats: COMPLETE Column stats: NONE
- Reduce Output Operator
- key expressions: key (type: int)
- sort order: +
- Map-reduce partition columns: key (type: int)
- Statistics: Num rows: 22 Data size: 163 Basic stats: COMPLETE Column stats: NONE
Reducer 2
Reduce Operator Tree:
Join Operator
@@ -2505,14 +2445,10 @@ STAGE PLANS:
2
outputColumnNames: _col0
Statistics: Num rows: 48 Data size: 358 Basic stats: COMPLETE Column stats: NONE
- Select Operator
- expressions: _col0 (type: int)
- outputColumnNames: _col0
+ Reduce Output Operator
+ key expressions: _col0 (type: int)
+ sort order: +
Statistics: Num rows: 48 Data size: 358 Basic stats: COMPLETE Column stats: NONE
- Reduce Output Operator
- key expressions: _col0 (type: int)
- sort order: +
- Statistics: Num rows: 48 Data size: 358 Basic stats: COMPLETE Column stats: NONE
Reducer 3
Reduce Operator Tree:
Select Operator
@@ -2601,7 +2537,7 @@ STAGE PLANS:
Spark
#### A masked pattern was here ####
Vertices:
- Map 1
+ Map 3
Map Operator Tree:
TableScan
alias: b
@@ -2627,7 +2563,7 @@ STAGE PLANS:
1 _col0 (type: int)
Local Work:
Map Reduce Local Work
- Map 2
+ Map 4
Map Operator Tree:
TableScan
alias: c
@@ -2645,10 +2581,10 @@ STAGE PLANS:
Stage: Stage-1
Spark
Edges:
- Reducer 4 <- Map 3 (SORT, 3)
+ Reducer 2 <- Map 1 (SORT, 3)
#### A masked pattern was here ####
Vertices:
- Map 3
+ Map 1
Map Operator Tree:
TableScan
alias: a
@@ -2667,7 +2603,7 @@ STAGE PLANS:
1 _col0 (type: int)
outputColumnNames: _col0, _col1
input vertices:
- 1 Map 1
+ 1 Map 3
Statistics: Num rows: 12 Data size: 89 Basic stats: COMPLETE Column stats: NONE
Map Join Operator
condition map:
@@ -2680,19 +2616,15 @@ STAGE PLANS:
1 value (type: string)
outputColumnNames: _col0
input vertices:
- 1 Map 2
+ 1 Map 4
Statistics: Num rows: 13 Data size: 97 Basic stats: COMPLETE Column stats: NONE
- Select Operator
- expressions: _col0 (type: int)
- outputColumnNames: _col0
+ Reduce Output Operator
+ key expressions: _col0 (type: int)
+ sort order: +
Statistics: Num rows: 13 Data size: 97 Basic stats: COMPLETE Column stats: NONE
- Reduce Output Operator
- key expressions: _col0 (type: int)
- sort order: +
- Statistics: Num rows: 13 Data size: 97 Basic stats: COMPLETE Column stats: NONE
Local Work:
Map Reduce Local Work
- Reducer 4
+ Reducer 2
Reduce Operator Tree:
Select Operator
expressions: KEY.reducesinkkey0 (type: int)
@@ -2773,7 +2705,7 @@ STAGE PLANS:
Spark
#### A masked pattern was here ####
Vertices:
- Map 1
+ Map 2
Map Operator Tree:
TableScan
alias: b
@@ -2804,7 +2736,7 @@ STAGE PLANS:
Spark
#### A masked pattern was here ####
Vertices:
- Map 2
+ Map 1
Map Operator Tree:
TableScan
alias: a
@@ -2823,7 +2755,7 @@ STAGE PLANS:
1 _col0 (type: string)
outputColumnNames: _col0
input vertices:
- 1 Map 1
+ 1 Map 2
Statistics: Num rows: 6 Data size: 49 Basic stats: COMPLETE Column stats: NONE
Select Operator
expressions: _col0 (type: int)
Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/skewjoin.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/skewjoin.q.out?rev=1642395&r1=1642394&r2=1642395&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/spark/skewjoin.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/spark/skewjoin.q.out Sat Nov 29 03:44:22 2014
@@ -89,7 +89,7 @@ STAGE PLANS:
Spark
#### A masked pattern was here ####
Vertices:
- Map 1
+ Map 2
Map Operator Tree:
TableScan
alias: src2
@@ -111,7 +111,7 @@ STAGE PLANS:
Spark
#### A masked pattern was here ####
Vertices:
- Map 2
+ Map 1
Map Operator Tree:
TableScan
alias: src1
@@ -130,7 +130,7 @@ STAGE PLANS:
1 key (type: string)
outputColumnNames: _col0, _col6
input vertices:
- 1 Map 1
+ 1 Map 2
Statistics: Num rows: 275 Data size: 2921 Basic stats: COMPLETE Column stats: NONE
Select Operator
expressions: UDFToInteger(_col0) (type: int), _col6 (type: string)
@@ -203,10 +203,10 @@ STAGE PLANS:
Spark
#### A masked pattern was here ####
Vertices:
- Map 1
+ Map 2
Map Operator Tree:
TableScan
- alias: d
+ alias: b
Statistics: Num rows: 0 Data size: 30 Basic stats: PARTIAL Column stats: NONE
Filter Operator
predicate: key is not null (type: boolean)
@@ -214,9 +214,9 @@ STAGE PLANS:
Spark HashTable Sink Operator
condition expressions:
0 {key} {val}
- 1 {key} {val}
+ 1 {val}
2 {key} {val}
- 3 {val}
+ 3 {key} {val}
keys:
0 key (type: string)
1 key (type: string)
@@ -224,19 +224,19 @@ STAGE PLANS:
3 key (type: string)
Local Work:
Map Reduce Local Work
- Map 2
+ Map 3
Map Operator Tree:
TableScan
- alias: b
- Statistics: Num rows: 0 Data size: 30 Basic stats: PARTIAL Column stats: NONE
+ alias: c
+ Statistics: Num rows: 0 Data size: 20 Basic stats: PARTIAL Column stats: NONE
Filter Operator
predicate: key is not null (type: boolean)
Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
Spark HashTable Sink Operator
condition expressions:
0 {key} {val}
- 1 {val}
- 2 {key} {val}
+ 1 {key} {val}
+ 2 {val}
3 {key} {val}
keys:
0 key (type: string)
@@ -245,11 +245,11 @@ STAGE PLANS:
3 key (type: string)
Local Work:
Map Reduce Local Work
- Map 3
+ Map 4
Map Operator Tree:
TableScan
- alias: c
- Statistics: Num rows: 0 Data size: 20 Basic stats: PARTIAL Column stats: NONE
+ alias: d
+ Statistics: Num rows: 0 Data size: 30 Basic stats: PARTIAL Column stats: NONE
Filter Operator
predicate: key is not null (type: boolean)
Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
@@ -257,8 +257,8 @@ STAGE PLANS:
condition expressions:
0 {key} {val}
1 {key} {val}
- 2 {val}
- 3 {key} {val}
+ 2 {key} {val}
+ 3 {val}
keys:
0 key (type: string)
1 key (type: string)
@@ -271,7 +271,7 @@ STAGE PLANS:
Spark
#### A masked pattern was here ####
Vertices:
- Map 4
+ Map 1
Map Operator Tree:
TableScan
alias: a
@@ -298,7 +298,7 @@ STAGE PLANS:
input vertices:
1 Map 2
2 Map 3
- 3 Map 1
+ 3 Map 4
Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
Select Operator
expressions: _col0 (type: string), _col1 (type: string), _col5 (type: string), _col6 (type: string), _col10 (type: string), _col11 (type: string), _col15 (type: string), _col16 (type: string)
@@ -363,10 +363,10 @@ STAGE PLANS:
Spark
#### A masked pattern was here ####
Vertices:
- Map 1
+ Map 2
Map Operator Tree:
TableScan
- alias: d
+ alias: b
Statistics: Num rows: 0 Data size: 30 Basic stats: PARTIAL Column stats: NONE
Filter Operator
predicate: key is not null (type: boolean)
@@ -374,9 +374,9 @@ STAGE PLANS:
Spark HashTable Sink Operator
condition expressions:
0 {key} {val}
- 1 {key} {val}
+ 1 {val}
2 {key} {val}
- 3 {val}
+ 3 {key} {val}
keys:
0 key (type: string)
1 key (type: string)
@@ -384,19 +384,19 @@ STAGE PLANS:
3 key (type: string)
Local Work:
Map Reduce Local Work
- Map 2
+ Map 3
Map Operator Tree:
TableScan
- alias: b
- Statistics: Num rows: 0 Data size: 30 Basic stats: PARTIAL Column stats: NONE
+ alias: c
+ Statistics: Num rows: 0 Data size: 20 Basic stats: PARTIAL Column stats: NONE
Filter Operator
predicate: key is not null (type: boolean)
Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
Spark HashTable Sink Operator
condition expressions:
0 {key} {val}
- 1 {val}
- 2 {key} {val}
+ 1 {key} {val}
+ 2 {val}
3 {key} {val}
keys:
0 key (type: string)
@@ -405,11 +405,11 @@ STAGE PLANS:
3 key (type: string)
Local Work:
Map Reduce Local Work
- Map 3
+ Map 4
Map Operator Tree:
TableScan
- alias: c
- Statistics: Num rows: 0 Data size: 20 Basic stats: PARTIAL Column stats: NONE
+ alias: d
+ Statistics: Num rows: 0 Data size: 30 Basic stats: PARTIAL Column stats: NONE
Filter Operator
predicate: key is not null (type: boolean)
Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
@@ -417,8 +417,8 @@ STAGE PLANS:
condition expressions:
0 {key} {val}
1 {key} {val}
- 2 {val}
- 3 {key} {val}
+ 2 {key} {val}
+ 3 {val}
keys:
0 key (type: string)
1 key (type: string)
@@ -431,7 +431,7 @@ STAGE PLANS:
Spark
#### A masked pattern was here ####
Vertices:
- Map 4
+ Map 1
Map Operator Tree:
TableScan
alias: a
@@ -458,7 +458,7 @@ STAGE PLANS:
input vertices:
1 Map 2
2 Map 3
- 3 Map 1
+ 3 Map 4
Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
Select Operator
expressions: _col0 (type: string), _col1 (type: string), _col5 (type: string), _col6 (type: string), _col10 (type: string), _col11 (type: string), _col15 (type: string), _col16 (type: string)
@@ -515,7 +515,7 @@ STAGE PLANS:
Spark
#### A masked pattern was here ####
Vertices:
- Map 3
+ Map 1
Map Operator Tree:
TableScan
alias: a
@@ -536,10 +536,10 @@ STAGE PLANS:
Stage: Stage-1
Spark
Edges:
- Reducer 2 <- Map 1 (GROUP, 1)
+ Reducer 3 <- Map 2 (GROUP, 1)
#### A masked pattern was here ####
Vertices:
- Map 1
+ Map 2
Map Operator Tree:
TableScan
alias: c
@@ -558,24 +558,20 @@ STAGE PLANS:
1 (key + 1) (type: double)
outputColumnNames: _col0, _col1, _col5
input vertices:
- 0 Map 3
+ 0 Map 1
Statistics: Num rows: 275 Data size: 2921 Basic stats: COMPLETE Column stats: NONE
- Select Operator
- expressions: _col0 (type: string), _col1 (type: string), _col5 (type: string)
- outputColumnNames: _col0, _col1, _col5
- Statistics: Num rows: 275 Data size: 2921 Basic stats: COMPLETE Column stats: NONE
- Group By Operator
- aggregations: sum(hash(_col0)), sum(hash(_col1)), sum(hash(_col5))
- mode: hash
- outputColumnNames: _col0, _col1, _col2
+ Group By Operator
+ aggregations: sum(hash(_col0)), sum(hash(_col1)), sum(hash(_col5))
+ mode: hash
+ outputColumnNames: _col0, _col1, _col2
+ Statistics: Num rows: 1 Data size: 24 Basic stats: COMPLETE Column stats: NONE
+ Reduce Output Operator
+ sort order:
Statistics: Num rows: 1 Data size: 24 Basic stats: COMPLETE Column stats: NONE
- Reduce Output Operator
- sort order:
- Statistics: Num rows: 1 Data size: 24 Basic stats: COMPLETE Column stats: NONE
- value expressions: _col0 (type: bigint), _col1 (type: bigint), _col2 (type: bigint)
+ value expressions: _col0 (type: bigint), _col1 (type: bigint), _col2 (type: bigint)
Local Work:
Map Reduce Local Work
- Reducer 2
+ Reducer 3
Reduce Operator Tree:
Group By Operator
aggregations: sum(VALUE._col0), sum(VALUE._col1), sum(VALUE._col2)
@@ -688,19 +684,15 @@ STAGE PLANS:
input vertices:
1 Map 3
Statistics: Num rows: 275 Data size: 2921 Basic stats: COMPLETE Column stats: NONE
- Select Operator
- expressions: _col2 (type: string), _col3 (type: string)
- outputColumnNames: _col2, _col3
- Statistics: Num rows: 275 Data size: 2921 Basic stats: COMPLETE Column stats: NONE
- Group By Operator
- aggregations: sum(hash(_col2)), sum(hash(_col3))
- mode: hash
- outputColumnNames: _col0, _col1
+ Group By Operator
+ aggregations: sum(hash(_col2)), sum(hash(_col3))
+ mode: hash
+ outputColumnNames: _col0, _col1
+ Statistics: Num rows: 1 Data size: 16 Basic stats: COMPLETE Column stats: NONE
+ Reduce Output Operator
+ sort order:
Statistics: Num rows: 1 Data size: 16 Basic stats: COMPLETE Column stats: NONE
- Reduce Output Operator
- sort order:
- Statistics: Num rows: 1 Data size: 16 Basic stats: COMPLETE Column stats: NONE
- value expressions: _col0 (type: bigint), _col1 (type: bigint)
+ value expressions: _col0 (type: bigint), _col1 (type: bigint)
Local Work:
Map Reduce Local Work
Reducer 2
@@ -824,19 +816,15 @@ STAGE PLANS:
input vertices:
1 Map 3
Statistics: Num rows: 137 Data size: 1460 Basic stats: COMPLETE Column stats: NONE
- Select Operator
- expressions: _col2 (type: string), _col3 (type: string)
- outputColumnNames: _col2, _col3
- Statistics: Num rows: 137 Data size: 1460 Basic stats: COMPLETE Column stats: NONE
- Group By Operator
- aggregations: sum(hash(_col2)), sum(hash(_col3))
- mode: hash
- outputColumnNames: _col0, _col1
+ Group By Operator
+ aggregations: sum(hash(_col2)), sum(hash(_col3))
+ mode: hash
+ outputColumnNames: _col0, _col1
+ Statistics: Num rows: 1 Data size: 16 Basic stats: COMPLETE Column stats: NONE
+ Reduce Output Operator
+ sort order:
Statistics: Num rows: 1 Data size: 16 Basic stats: COMPLETE Column stats: NONE
- Reduce Output Operator
- sort order:
- Statistics: Num rows: 1 Data size: 16 Basic stats: COMPLETE Column stats: NONE
- value expressions: _col0 (type: bigint), _col1 (type: bigint)
+ value expressions: _col0 (type: bigint), _col1 (type: bigint)
Local Work:
Map Reduce Local Work
Reducer 2
@@ -915,7 +903,7 @@ STAGE PLANS:
Spark
#### A masked pattern was here ####
Vertices:
- Map 1
+ Map 3
Map Operator Tree:
TableScan
alias: src
@@ -965,10 +953,10 @@ STAGE PLANS:
Stage: Stage-1
Spark
Edges:
- Reducer 3 <- Map 2 (GROUP, 1)
+ Reducer 2 <- Map 1 (GROUP, 1)
#### A masked pattern was here ####
Vertices:
- Map 2
+ Map 1
Map Operator Tree:
TableScan
alias: src
@@ -994,25 +982,21 @@ STAGE PLANS:
2 _col0 (type: string)
outputColumnNames: _col0, _col3
input vertices:
- 1 Map 1
+ 1 Map 3
2 Map 4
Statistics: Num rows: 121 Data size: 1284 Basic stats: COMPLETE Column stats: NONE
- Select Operator
- expressions: _col0 (type: string), _col3 (type: string)
- outputColumnNames: _col0, _col3
- Statistics: Num rows: 121 Data size: 1284 Basic stats: COMPLETE Column stats: NONE
- Group By Operator
- aggregations: sum(hash(_col0)), sum(hash(_col3))
- mode: hash
- outputColumnNames: _col0, _col1
+ Group By Operator
+ aggregations: sum(hash(_col0)), sum(hash(_col3))
+ mode: hash
+ outputColumnNames: _col0, _col1
+ Statistics: Num rows: 1 Data size: 16 Basic stats: COMPLETE Column stats: NONE
+ Reduce Output Operator
+ sort order:
Statistics: Num rows: 1 Data size: 16 Basic stats: COMPLETE Column stats: NONE
- Reduce Output Operator
- sort order:
- Statistics: Num rows: 1 Data size: 16 Basic stats: COMPLETE Column stats: NONE
- value expressions: _col0 (type: bigint), _col1 (type: bigint)
+ value expressions: _col0 (type: bigint), _col1 (type: bigint)
Local Work:
Map Reduce Local Work
- Reducer 3
+ Reducer 2
Reduce Operator Tree:
Group By Operator
aggregations: sum(VALUE._col0), sum(VALUE._col1)
@@ -1078,7 +1062,7 @@ STAGE PLANS:
Spark
#### A masked pattern was here ####
Vertices:
- Map 1
+ Map 3
Map Operator Tree:
TableScan
alias: v
@@ -1096,10 +1080,10 @@ STAGE PLANS:
Stage: Stage-1
Spark
Edges:
- Reducer 3 <- Map 2 (GROUP, 1)
+ Reducer 2 <- Map 1 (GROUP, 1)
#### A masked pattern was here ####
Vertices:
- Map 2
+ Map 1
Map Operator Tree:
TableScan
alias: k
@@ -1115,24 +1099,20 @@ STAGE PLANS:
1 UDFToDouble(key) (type: double)
outputColumnNames: _col0, _col6
input vertices:
- 1 Map 1
+ 1 Map 3
Statistics: Num rows: 0 Data size: 33 Basic stats: PARTIAL Column stats: NONE
- Select Operator
- expressions: _col0 (type: string), _col6 (type: string)
- outputColumnNames: _col0, _col6
- Statistics: Num rows: 0 Data size: 33 Basic stats: PARTIAL Column stats: NONE
- Group By Operator
- aggregations: sum(hash(_col0)), sum(hash(_col6))
- mode: hash
- outputColumnNames: _col0, _col1
+ Group By Operator
+ aggregations: sum(hash(_col0)), sum(hash(_col6))
+ mode: hash
+ outputColumnNames: _col0, _col1
+ Statistics: Num rows: 1 Data size: 16 Basic stats: COMPLETE Column stats: NONE
+ Reduce Output Operator
+ sort order:
Statistics: Num rows: 1 Data size: 16 Basic stats: COMPLETE Column stats: NONE
- Reduce Output Operator
- sort order:
- Statistics: Num rows: 1 Data size: 16 Basic stats: COMPLETE Column stats: NONE
- value expressions: _col0 (type: bigint), _col1 (type: bigint)
+ value expressions: _col0 (type: bigint), _col1 (type: bigint)
Local Work:
Map Reduce Local Work
- Reducer 3
+ Reducer 2
Reduce Operator Tree:
Group By Operator
aggregations: sum(VALUE._col0), sum(VALUE._col1)
Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/skewjoin_noskew.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/skewjoin_noskew.q.out?rev=1642395&r1=1642394&r2=1642395&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/spark/skewjoin_noskew.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/spark/skewjoin_noskew.q.out Sat Nov 29 03:44:22 2014
@@ -16,7 +16,7 @@ STAGE PLANS:
Spark
#### A masked pattern was here ####
Vertices:
- Map 1
+ Map 3
Map Operator Tree:
TableScan
alias: b
@@ -37,10 +37,10 @@ STAGE PLANS:
Stage: Stage-1
Spark
Edges:
- Reducer 3 <- Map 2 (SORT, 1)
+ Reducer 2 <- Map 1 (SORT, 1)
#### A masked pattern was here ####
Vertices:
- Map 2
+ Map 1
Map Operator Tree:
TableScan
alias: a
@@ -59,20 +59,16 @@ STAGE PLANS:
1 key (type: string)
outputColumnNames: _col0, _col1
input vertices:
- 1 Map 1
+ 1 Map 3
Statistics: Num rows: 275 Data size: 2921 Basic stats: COMPLETE Column stats: NONE
- Select Operator
- expressions: _col0 (type: string), _col1 (type: string)
- outputColumnNames: _col0, _col1
+ Reduce Output Operator
+ key expressions: _col0 (type: string)
+ sort order: +
Statistics: Num rows: 275 Data size: 2921 Basic stats: COMPLETE Column stats: NONE
- Reduce Output Operator
- key expressions: _col0 (type: string)
- sort order: +
- Statistics: Num rows: 275 Data size: 2921 Basic stats: COMPLETE Column stats: NONE
- value expressions: _col1 (type: string)
+ value expressions: _col1 (type: string)
Local Work:
Map Reduce Local Work
- Reducer 3
+ Reducer 2
Reduce Operator Tree:
Select Operator
expressions: KEY.reducesinkkey0 (type: string), VALUE._col0 (type: string)
Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/skewjoin_union_remove_1.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/skewjoin_union_remove_1.q.out?rev=1642395&r1=1642394&r2=1642395&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/spark/skewjoin_union_remove_1.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/spark/skewjoin_union_remove_1.q.out Sat Nov 29 03:44:22 2014
@@ -77,7 +77,7 @@ STAGE PLANS:
alias: b
Statistics: Num rows: 0 Data size: 30 Basic stats: PARTIAL Column stats: NONE
Filter Operator
- predicate: (key is not null and (not ((key = '2') or (key = '3')))) (type: boolean)
+ predicate: (key is not null and ((key = '2') or (key = '3'))) (type: boolean)
Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
Spark HashTable Sink Operator
condition expressions:
@@ -100,7 +100,7 @@ STAGE PLANS:
TableScan
alias: a
Filter Operator
- predicate: (key is not null and (not ((key = '2') or (key = '3')))) (type: boolean)
+ predicate: (key is not null and ((key = '2') or (key = '3'))) (type: boolean)
Map Join Operator
condition map:
Inner Join 0 to 1
@@ -116,14 +116,12 @@ STAGE PLANS:
Select Operator
expressions: _col0 (type: string), _col1 (type: string), _col5 (type: string), _col6 (type: string)
outputColumnNames: _col0, _col1, _col2, _col3
- Select Operator
- SELECT * : (no compute)
- File Output Operator
- compressed: false
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ File Output Operator
+ compressed: false
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
Local Work:
Map Reduce Local Work
Map 5
@@ -131,7 +129,7 @@ STAGE PLANS:
TableScan
alias: a
Filter Operator
- predicate: (key is not null and ((key = '2') or (key = '3'))) (type: boolean)
+ predicate: (key is not null and (not ((key = '2') or (key = '3')))) (type: boolean)
Map Join Operator
condition map:
Inner Join 0 to 1
@@ -147,14 +145,12 @@ STAGE PLANS:
Select Operator
expressions: _col0 (type: string), _col1 (type: string), _col5 (type: string), _col6 (type: string)
outputColumnNames: _col0, _col1, _col2, _col3
- Select Operator
- SELECT * : (no compute)
- File Output Operator
- compressed: false
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ File Output Operator
+ compressed: false
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
Local Work:
Map Reduce Local Work
Union 2
@@ -170,7 +166,7 @@ STAGE PLANS:
alias: b
Statistics: Num rows: 0 Data size: 30 Basic stats: PARTIAL Column stats: NONE
Filter Operator
- predicate: (key is not null and ((key = '2') or (key = '3'))) (type: boolean)
+ predicate: (key is not null and (not ((key = '2') or (key = '3')))) (type: boolean)
Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
Spark HashTable Sink Operator
condition expressions:
@@ -233,7 +229,7 @@ STAGE PLANS:
alias: a
Statistics: Num rows: 0 Data size: 30 Basic stats: PARTIAL Column stats: NONE
Filter Operator
- predicate: (not ((key = '2') or (key = '3'))) (type: boolean)
+ predicate: ((key = '2') or (key = '3')) (type: boolean)
Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
Spark HashTable Sink Operator
condition expressions:
@@ -256,7 +252,7 @@ STAGE PLANS:
TableScan
alias: b
Filter Operator
- predicate: (not ((key = '2') or (key = '3'))) (type: boolean)
+ predicate: ((key = '2') or (key = '3')) (type: boolean)
Map Join Operator
condition map:
Right Outer Join0 to 1
@@ -272,14 +268,12 @@ STAGE PLANS:
Select Operator
expressions: _col0 (type: string), _col1 (type: string), _col5 (type: string), _col6 (type: string)
outputColumnNames: _col0, _col1, _col2, _col3
- Select Operator
- SELECT * : (no compute)
- File Output Operator
- compressed: false
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ File Output Operator
+ compressed: false
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
Local Work:
Map Reduce Local Work
Map 4
@@ -287,7 +281,7 @@ STAGE PLANS:
TableScan
alias: b
Filter Operator
- predicate: ((key = '2') or (key = '3')) (type: boolean)
+ predicate: (not ((key = '2') or (key = '3'))) (type: boolean)
Map Join Operator
condition map:
Right Outer Join0 to 1
@@ -303,14 +297,12 @@ STAGE PLANS:
Select Operator
expressions: _col0 (type: string), _col1 (type: string), _col5 (type: string), _col6 (type: string)
outputColumnNames: _col0, _col1, _col2, _col3
- Select Operator
- SELECT * : (no compute)
- File Output Operator
- compressed: false
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ File Output Operator
+ compressed: false
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
Local Work:
Map Reduce Local Work
Union 3
@@ -326,7 +318,7 @@ STAGE PLANS:
alias: a
Statistics: Num rows: 0 Data size: 30 Basic stats: PARTIAL Column stats: NONE
Filter Operator
- predicate: ((key = '2') or (key = '3')) (type: boolean)
+ predicate: (not ((key = '2') or (key = '3'))) (type: boolean)
Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
Spark HashTable Sink Operator
condition expressions:
@@ -397,7 +389,7 @@ STAGE PLANS:
alias: b
Statistics: Num rows: 0 Data size: 30 Basic stats: PARTIAL Column stats: NONE
Filter Operator
- predicate: (key is not null and (not ((key = '2') or (key = '3')))) (type: boolean)
+ predicate: (key is not null and ((key = '2') or (key = '3'))) (type: boolean)
Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
Spark HashTable Sink Operator
condition expressions:
@@ -420,7 +412,7 @@ STAGE PLANS:
TableScan
alias: a
Filter Operator
- predicate: (key is not null and (not ((key = '2') or (key = '3')))) (type: boolean)
+ predicate: (key is not null and ((key = '2') or (key = '3'))) (type: boolean)
Map Join Operator
condition map:
Inner Join 0 to 1
@@ -436,15 +428,13 @@ STAGE PLANS:
Select Operator
expressions: _col0 (type: string), _col1 (type: string), _col5 (type: string), _col6 (type: string)
outputColumnNames: _col0, _col1, _col2, _col3
- Select Operator
- SELECT * : (no compute)
- File Output Operator
- compressed: false
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- name: default.dest1
+ File Output Operator
+ compressed: false
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ name: default.dest1
Local Work:
Map Reduce Local Work
Map 5
@@ -452,7 +442,7 @@ STAGE PLANS:
TableScan
alias: a
Filter Operator
- predicate: (key is not null and ((key = '2') or (key = '3'))) (type: boolean)
+ predicate: (key is not null and (not ((key = '2') or (key = '3')))) (type: boolean)
Map Join Operator
condition map:
Inner Join 0 to 1
@@ -468,15 +458,13 @@ STAGE PLANS:
Select Operator
expressions: _col0 (type: string), _col1 (type: string), _col5 (type: string), _col6 (type: string)
outputColumnNames: _col0, _col1, _col2, _col3
- Select Operator
- SELECT * : (no compute)
- File Output Operator
- compressed: false
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- name: default.dest1
+ File Output Operator
+ compressed: false
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ name: default.dest1
Local Work:
Map Reduce Local Work
Union 2
@@ -502,7 +490,7 @@ STAGE PLANS:
alias: b
Statistics: Num rows: 0 Data size: 30 Basic stats: PARTIAL Column stats: NONE
Filter Operator
- predicate: (key is not null and ((key = '2') or (key = '3'))) (type: boolean)
+ predicate: (key is not null and (not ((key = '2') or (key = '3')))) (type: boolean)
Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
Spark HashTable Sink Operator
condition expressions:
@@ -571,7 +559,7 @@ STAGE PLANS:
alias: a
Statistics: Num rows: 0 Data size: 30 Basic stats: PARTIAL Column stats: NONE
Filter Operator
- predicate: (not ((key = '2') or (key = '3'))) (type: boolean)
+ predicate: ((key = '2') or (key = '3')) (type: boolean)
Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
Spark HashTable Sink Operator
condition expressions:
@@ -594,7 +582,7 @@ STAGE PLANS:
TableScan
alias: b
Filter Operator
- predicate: (not ((key = '2') or (key = '3'))) (type: boolean)
+ predicate: ((key = '2') or (key = '3')) (type: boolean)
Map Join Operator
condition map:
Right Outer Join0 to 1
@@ -610,15 +598,13 @@ STAGE PLANS:
Select Operator
expressions: _col0 (type: string), _col1 (type: string), _col5 (type: string), _col6 (type: string)
outputColumnNames: _col0, _col1, _col2, _col3
- Select Operator
- SELECT * : (no compute)
- File Output Operator
- compressed: false
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- name: default.dest1
+ File Output Operator
+ compressed: false
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ name: default.dest1
Local Work:
Map Reduce Local Work
Map 4
@@ -626,7 +612,7 @@ STAGE PLANS:
TableScan
alias: b
Filter Operator
- predicate: ((key = '2') or (key = '3')) (type: boolean)
+ predicate: (not ((key = '2') or (key = '3'))) (type: boolean)
Map Join Operator
condition map:
Right Outer Join0 to 1
@@ -642,15 +628,13 @@ STAGE PLANS:
Select Operator
expressions: _col0 (type: string), _col1 (type: string), _col5 (type: string), _col6 (type: string)
outputColumnNames: _col0, _col1, _col2, _col3
- Select Operator
- SELECT * : (no compute)
- File Output Operator
- compressed: false
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- name: default.dest1
+ File Output Operator
+ compressed: false
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ name: default.dest1
Local Work:
Map Reduce Local Work
Union 3
@@ -676,7 +660,7 @@ STAGE PLANS:
alias: a
Statistics: Num rows: 0 Data size: 30 Basic stats: PARTIAL Column stats: NONE
Filter Operator
- predicate: ((key = '2') or (key = '3')) (type: boolean)
+ predicate: (not ((key = '2') or (key = '3'))) (type: boolean)
Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
Spark HashTable Sink Operator
condition expressions: