You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hive.apache.org by br...@apache.org on 2014/11/29 04:44:28 UTC
svn commit: r1642395 [11/22] - in /hive/branches/spark/ql/src:
java/org/apache/hadoop/hive/ql/exec/spark/
java/org/apache/hadoop/hive/ql/exec/spark/session/
test/results/clientpositive/ test/results/clientpositive/spark/
Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/mergejoins_mixed.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/mergejoins_mixed.q.out?rev=1642395&r1=1642394&r2=1642395&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/spark/mergejoins_mixed.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/spark/mergejoins_mixed.q.out Sat Nov 29 03:44:22 2014
@@ -28,17 +28,17 @@ STAGE PLANS:
Spark
#### A masked pattern was here ####
Vertices:
- Map 1
+ Map 2
Map Operator Tree:
TableScan
- alias: d
+ alias: b
Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
Spark HashTable Sink Operator
condition expressions:
0 {key} {value}
- 1 {key} {value}
+ 1 {value}
2 {key} {value}
- 3 {value}
+ 3 {key} {value}
keys:
0 key (type: string)
1 key (type: string)
@@ -46,16 +46,16 @@ STAGE PLANS:
3 key (type: string)
Local Work:
Map Reduce Local Work
- Map 2
+ Map 3
Map Operator Tree:
TableScan
- alias: b
+ alias: c
Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
Spark HashTable Sink Operator
condition expressions:
0 {key} {value}
- 1 {value}
- 2 {key} {value}
+ 1 {key} {value}
+ 2 {value}
3 {key} {value}
keys:
0 key (type: string)
@@ -64,17 +64,17 @@ STAGE PLANS:
3 key (type: string)
Local Work:
Map Reduce Local Work
- Map 3
+ Map 4
Map Operator Tree:
TableScan
- alias: c
+ alias: d
Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
Spark HashTable Sink Operator
condition expressions:
0 {key} {value}
1 {key} {value}
- 2 {value}
- 3 {key} {value}
+ 2 {key} {value}
+ 3 {value}
keys:
0 key (type: string)
1 key (type: string)
@@ -87,7 +87,7 @@ STAGE PLANS:
Spark
#### A masked pattern was here ####
Vertices:
- Map 4
+ Map 1
Map Operator Tree:
TableScan
alias: a
@@ -111,7 +111,7 @@ STAGE PLANS:
input vertices:
1 Map 2
2 Map 3
- 3 Map 1
+ 3 Map 4
Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
Select Operator
expressions: _col0 (type: string), _col1 (type: string), _col5 (type: string), _col6 (type: string), _col10 (type: string), _col11 (type: string), _col15 (type: string), _col16 (type: string)
@@ -149,15 +149,15 @@ STAGE PLANS:
Spark
#### A masked pattern was here ####
Vertices:
- Map 2
+ Map 1
Map Operator Tree:
TableScan
- alias: b
+ alias: a
Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
Spark HashTable Sink Operator
condition expressions:
- 0 {key} {value}
- 1 {value}
+ 0 {value}
+ 1 {key} {value}
2 {key} {value}
3 {key} {value}
keys:
@@ -167,16 +167,16 @@ STAGE PLANS:
3 key (type: string)
Local Work:
Map Reduce Local Work
- Map 3
+ Map 2
Map Operator Tree:
TableScan
- alias: c
+ alias: b
Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
Spark HashTable Sink Operator
condition expressions:
0 {key} {value}
- 1 {key} {value}
- 2 {value}
+ 1 {value}
+ 2 {key} {value}
3 {key} {value}
keys:
0 key (type: string)
@@ -185,16 +185,16 @@ STAGE PLANS:
3 key (type: string)
Local Work:
Map Reduce Local Work
- Map 4
+ Map 3
Map Operator Tree:
TableScan
- alias: a
+ alias: c
Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
Spark HashTable Sink Operator
condition expressions:
- 0 {value}
+ 0 {key} {value}
1 {key} {value}
- 2 {key} {value}
+ 2 {value}
3 {key} {value}
keys:
0 key (type: string)
@@ -208,7 +208,7 @@ STAGE PLANS:
Spark
#### A masked pattern was here ####
Vertices:
- Map 1
+ Map 4
Map Operator Tree:
TableScan
alias: d
@@ -230,7 +230,7 @@ STAGE PLANS:
3 key (type: string)
outputColumnNames: _col0, _col1, _col5, _col6, _col10, _col11, _col15, _col16
input vertices:
- 0 Map 4
+ 0 Map 1
1 Map 2
2 Map 3
Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
@@ -273,14 +273,14 @@ STAGE PLANS:
Map 1
Map Operator Tree:
TableScan
- alias: d
+ alias: a
Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
Spark HashTable Sink Operator
condition expressions:
- 0 {key} {value}
+ 0 {value}
1 {key} {value}
2 {key} {value}
- 3 {value}
+ 3 {key} {value}
keys:
0 key (type: string)
1 key (type: string)
@@ -309,14 +309,14 @@ STAGE PLANS:
Map 4
Map Operator Tree:
TableScan
- alias: a
+ alias: d
Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
Spark HashTable Sink Operator
condition expressions:
- 0 {value}
+ 0 {key} {value}
1 {key} {value}
2 {key} {value}
- 3 {key} {value}
+ 3 {value}
keys:
0 key (type: string)
1 key (type: string)
@@ -351,9 +351,9 @@ STAGE PLANS:
3 key (type: string)
outputColumnNames: _col0, _col1, _col5, _col6, _col10, _col11, _col15, _col16
input vertices:
- 0 Map 4
+ 0 Map 1
1 Map 2
- 3 Map 1
+ 3 Map 4
Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
Select Operator
expressions: _col0 (type: string), _col1 (type: string), _col5 (type: string), _col6 (type: string), _col10 (type: string), _col11 (type: string), _col15 (type: string), _col16 (type: string)
@@ -391,15 +391,15 @@ STAGE PLANS:
Spark
#### A masked pattern was here ####
Vertices:
- Map 2
+ Map 1
Map Operator Tree:
TableScan
- alias: b
+ alias: a
Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
Spark HashTable Sink Operator
condition expressions:
- 0 {key} {value}
- 1 {value}
+ 0 {value}
+ 1 {key} {value}
2 {key} {value}
3 {key} {value}
keys:
@@ -409,16 +409,16 @@ STAGE PLANS:
3 key (type: string)
Local Work:
Map Reduce Local Work
- Map 3
+ Map 2
Map Operator Tree:
TableScan
- alias: c
+ alias: b
Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
Spark HashTable Sink Operator
condition expressions:
0 {key} {value}
- 1 {key} {value}
- 2 {value}
+ 1 {value}
+ 2 {key} {value}
3 {key} {value}
keys:
0 key (type: string)
@@ -427,16 +427,16 @@ STAGE PLANS:
3 key (type: string)
Local Work:
Map Reduce Local Work
- Map 4
+ Map 3
Map Operator Tree:
TableScan
- alias: a
+ alias: c
Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
Spark HashTable Sink Operator
condition expressions:
- 0 {value}
+ 0 {key} {value}
1 {key} {value}
- 2 {key} {value}
+ 2 {value}
3 {key} {value}
keys:
0 key (type: string)
@@ -450,7 +450,7 @@ STAGE PLANS:
Spark
#### A masked pattern was here ####
Vertices:
- Map 1
+ Map 4
Map Operator Tree:
TableScan
alias: d
@@ -472,7 +472,7 @@ STAGE PLANS:
3 key (type: string)
outputColumnNames: _col0, _col1, _col5, _col6, _col10, _col11, _col15, _col16
input vertices:
- 0 Map 4
+ 0 Map 1
1 Map 2
2 Map 3
Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
@@ -514,50 +514,50 @@ STAGE PLANS:
Spark
#### A masked pattern was here ####
Vertices:
- Map 1
+ Map 2
Map Operator Tree:
TableScan
- alias: d
+ alias: b
Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
Spark HashTable Sink Operator
condition expressions:
0 {key} {value}
- 1 {key} {value}
- 2 {value}
+ 1 {value}
+ 2 {key} {value}
keys:
0 key (type: string)
1 key (type: string)
2 key (type: string)
Local Work:
Map Reduce Local Work
- Map 2
+ Map 3
Map Operator Tree:
TableScan
- alias: b
+ alias: c
Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
Spark HashTable Sink Operator
condition expressions:
- 0 {key} {value}
+ 0 {_col0} {_col1} {_col5} {_col6} {_col10} {_col11}
1 {value}
- 2 {key} {value}
keys:
- 0 key (type: string)
+ 0 _col6 (type: string)
1 key (type: string)
- 2 key (type: string)
Local Work:
Map Reduce Local Work
- Map 3
+ Map 4
Map Operator Tree:
TableScan
- alias: c
+ alias: d
Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
Spark HashTable Sink Operator
condition expressions:
- 0 {_col0} {_col1} {_col5} {_col6} {_col10} {_col11}
- 1 {value}
+ 0 {key} {value}
+ 1 {key} {value}
+ 2 {value}
keys:
- 0 _col6 (type: string)
+ 0 key (type: string)
1 key (type: string)
+ 2 key (type: string)
Local Work:
Map Reduce Local Work
@@ -565,7 +565,7 @@ STAGE PLANS:
Spark
#### A masked pattern was here ####
Vertices:
- Map 4
+ Map 1
Map Operator Tree:
TableScan
alias: a
@@ -585,7 +585,7 @@ STAGE PLANS:
outputColumnNames: _col0, _col1, _col5, _col6, _col10, _col11
input vertices:
1 Map 2
- 2 Map 1
+ 2 Map 4
Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
Map Join Operator
condition map:
@@ -637,15 +637,15 @@ STAGE PLANS:
Spark
#### A masked pattern was here ####
Vertices:
- Map 2
+ Map 1
Map Operator Tree:
TableScan
- alias: b
+ alias: a
Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
Spark HashTable Sink Operator
condition expressions:
- 0 {key} {value}
- 1 {value}
+ 0 {value}
+ 1 {key} {value}
2 {key} {value}
keys:
0 key (type: string)
@@ -653,15 +653,15 @@ STAGE PLANS:
2 key (type: string)
Local Work:
Map Reduce Local Work
- Map 4
+ Map 2
Map Operator Tree:
TableScan
- alias: a
+ alias: b
Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
Spark HashTable Sink Operator
condition expressions:
- 0 {value}
- 1 {key} {value}
+ 0 {key} {value}
+ 1 {value}
2 {key} {value}
keys:
0 key (type: string)
@@ -674,7 +674,7 @@ STAGE PLANS:
Spark
#### A masked pattern was here ####
Vertices:
- Map 1
+ Map 4
Map Operator Tree:
TableScan
alias: d
@@ -693,7 +693,7 @@ STAGE PLANS:
2 key (type: string)
outputColumnNames: _col0, _col1, _col5, _col6, _col10, _col11
input vertices:
- 0 Map 4
+ 0 Map 1
1 Map 2
Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
Spark HashTable Sink Operator
@@ -726,7 +726,7 @@ STAGE PLANS:
1 key (type: string)
outputColumnNames: _col0, _col1, _col5, _col6, _col10, _col11, _col15, _col16
input vertices:
- 0 Map 1
+ 0 Map 4
Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
Select Operator
expressions: _col0 (type: string), _col1 (type: string), _col5 (type: string), _col6 (type: string), _col15 (type: string), _col16 (type: string), _col10 (type: string), _col11 (type: string)
@@ -769,7 +769,7 @@ STAGE PLANS:
Map 1
Map Operator Tree:
TableScan
- alias: d
+ alias: a
Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
Reduce Output Operator
key expressions: key (type: string)
@@ -802,7 +802,7 @@ STAGE PLANS:
Map 6
Map Operator Tree:
TableScan
- alias: a
+ alias: d
Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
Reduce Output Operator
key expressions: key (type: string)
@@ -911,7 +911,7 @@ STAGE PLANS:
Spark
#### A masked pattern was here ####
Vertices:
- Map 4
+ Map 1
Map Operator Tree:
TableScan
alias: a
@@ -959,7 +959,7 @@ STAGE PLANS:
Spark
#### A masked pattern was here ####
Vertices:
- Map 1
+ Map 4
Map Operator Tree:
TableScan
alias: d
@@ -975,7 +975,7 @@ STAGE PLANS:
1 key (type: string)
outputColumnNames: _col0, _col1, _col5, _col6, _col10, _col11, _col15, _col16
input vertices:
- 0 Map 4
+ 0 Map 1
Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
Select Operator
expressions: _col0 (type: string), _col1 (type: string), _col5 (type: string), _col6 (type: string), _col10 (type: string), _col11 (type: string), _col15 (type: string), _col16 (type: string)
@@ -1054,17 +1054,6 @@ STAGE PLANS:
Map 1
Map Operator Tree:
TableScan
- alias: d
- Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
- Reduce Output Operator
- key expressions: key (type: string)
- sort order: +
- Map-reduce partition columns: key (type: string)
- Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
- value expressions: value (type: string)
- Map 5
- Map Operator Tree:
- TableScan
alias: a
Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
Filter Operator
@@ -1104,6 +1093,17 @@ STAGE PLANS:
value expressions: _col1 (type: string), _col5 (type: string), _col6 (type: string), _col10 (type: string), _col11 (type: string)
Local Work:
Map Reduce Local Work
+ Map 5
+ Map Operator Tree:
+ TableScan
+ alias: d
+ Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
+ Reduce Output Operator
+ key expressions: key (type: string)
+ sort order: +
+ Map-reduce partition columns: key (type: string)
+ Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
+ value expressions: value (type: string)
Reducer 2
Reduce Operator Tree:
Join Operator
@@ -1174,20 +1174,6 @@ STAGE PLANS:
Map 1
Map Operator Tree:
TableScan
- alias: d
- Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
- Spark HashTable Sink Operator
- condition expressions:
- 0 {_col0} {_col1} {_col5} {_col6} {_col10} {_col11}
- 1 {value}
- keys:
- 0 _col0 (type: string)
- 1 key (type: string)
- Local Work:
- Map Reduce Local Work
- Map 4
- Map Operator Tree:
- TableScan
alias: a
Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
Filter Operator
@@ -1215,6 +1201,20 @@ STAGE PLANS:
1 key (type: string)
Local Work:
Map Reduce Local Work
+ Map 4
+ Map Operator Tree:
+ TableScan
+ alias: d
+ Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
+ Spark HashTable Sink Operator
+ condition expressions:
+ 0 {_col0} {_col1} {_col5} {_col6} {_col10} {_col11}
+ 1 {value}
+ keys:
+ 0 _col0 (type: string)
+ 1 key (type: string)
+ Local Work:
+ Map Reduce Local Work
Stage: Stage-1
Spark
@@ -1236,7 +1236,7 @@ STAGE PLANS:
1 key (type: string)
outputColumnNames: _col0, _col1, _col5, _col6, _col10, _col11
input vertices:
- 0 Map 4
+ 0 Map 1
Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
Map Join Operator
condition map:
@@ -1249,7 +1249,7 @@ STAGE PLANS:
1 key (type: string)
outputColumnNames: _col0, _col1, _col5, _col6, _col10, _col11, _col15, _col16
input vertices:
- 1 Map 1
+ 1 Map 4
Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
Select Operator
expressions: _col0 (type: string), _col1 (type: string), _col5 (type: string), _col6 (type: string), _col10 (type: string), _col11 (type: string), _col15 (type: string), _col16 (type: string)
@@ -1288,7 +1288,7 @@ STAGE PLANS:
Spark
#### A masked pattern was here ####
Vertices:
- Map 3
+ Map 2
Map Operator Tree:
TableScan
alias: b
@@ -1310,7 +1310,7 @@ STAGE PLANS:
Spark
#### A masked pattern was here ####
Vertices:
- Map 5
+ Map 1
Map Operator Tree:
TableScan
alias: a
@@ -1329,7 +1329,7 @@ STAGE PLANS:
1 key (type: string)
outputColumnNames: _col0, _col1, _col5, _col6
input vertices:
- 1 Map 3
+ 1 Map 2
Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
Spark HashTable Sink Operator
condition expressions:
@@ -1344,21 +1344,10 @@ STAGE PLANS:
Stage: Stage-1
Spark
Edges:
- Reducer 2 <- Map 1 (PARTITION-LEVEL SORT, 1), Map 4 (PARTITION-LEVEL SORT, 1)
+ Reducer 4 <- Map 3 (PARTITION-LEVEL SORT, 1), Map 5 (PARTITION-LEVEL SORT, 1)
#### A masked pattern was here ####
Vertices:
- Map 1
- Map Operator Tree:
- TableScan
- alias: d
- Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
- Reduce Output Operator
- key expressions: key (type: string)
- sort order: +
- Map-reduce partition columns: key (type: string)
- Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
- value expressions: value (type: string)
- Map 4
+ Map 3
Map Operator Tree:
TableScan
alias: c
@@ -1374,7 +1363,7 @@ STAGE PLANS:
1 key (type: string)
outputColumnNames: _col0, _col1, _col5, _col6, _col10, _col11
input vertices:
- 0 Map 5
+ 0 Map 1
Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
Reduce Output Operator
key expressions: _col0 (type: string)
@@ -1384,7 +1373,18 @@ STAGE PLANS:
value expressions: _col1 (type: string), _col5 (type: string), _col6 (type: string), _col10 (type: string), _col11 (type: string)
Local Work:
Map Reduce Local Work
- Reducer 2
+ Map 5
+ Map Operator Tree:
+ TableScan
+ alias: d
+ Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
+ Reduce Output Operator
+ key expressions: key (type: string)
+ sort order: +
+ Map-reduce partition columns: key (type: string)
+ Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
+ value expressions: value (type: string)
+ Reducer 4
Reduce Operator Tree:
Join Operator
condition map:
@@ -1430,22 +1430,6 @@ STAGE PLANS:
Spark
#### A masked pattern was here ####
Vertices:
- Map 1
- Map Operator Tree:
- TableScan
- alias: d
- Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
- Spark HashTable Sink Operator
- condition expressions:
- 0 {_col0} {_col1} {_col5} {_col6}
- 1 {key} {value}
- 2 {value}
- keys:
- 0 _col6 (type: string)
- 1 key (type: string)
- 2 key (type: string)
- Local Work:
- Map Reduce Local Work
Map 2
Map Operator Tree:
TableScan
@@ -1479,12 +1463,28 @@ STAGE PLANS:
2 key (type: string)
Local Work:
Map Reduce Local Work
+ Map 4
+ Map Operator Tree:
+ TableScan
+ alias: d
+ Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
+ Spark HashTable Sink Operator
+ condition expressions:
+ 0 {_col0} {_col1} {_col5} {_col6}
+ 1 {key} {value}
+ 2 {value}
+ keys:
+ 0 _col6 (type: string)
+ 1 key (type: string)
+ 2 key (type: string)
+ Local Work:
+ Map Reduce Local Work
Stage: Stage-1
Spark
#### A masked pattern was here ####
Vertices:
- Map 4
+ Map 1
Map Operator Tree:
TableScan
alias: a
@@ -1520,7 +1520,7 @@ STAGE PLANS:
outputColumnNames: _col0, _col1, _col5, _col6, _col10, _col11, _col15, _col16
input vertices:
1 Map 3
- 2 Map 1
+ 2 Map 4
Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
Select Operator
expressions: _col0 (type: string), _col1 (type: string), _col5 (type: string), _col6 (type: string), _col10 (type: string), _col11 (type: string), _col15 (type: string), _col16 (type: string)
Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/multi_insert.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/multi_insert.q.out?rev=1642395&r1=1642394&r2=1642395&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/spark/multi_insert.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/spark/multi_insert.q.out Sat Nov 29 03:44:22 2014
@@ -598,15 +598,11 @@ STAGE PLANS:
Filter Operator
predicate: ((key < 10) or ((key > 10) and (key < 20))) (type: boolean)
Statistics: Num rows: 221 Data size: 2347 Basic stats: COMPLETE Column stats: NONE
- Select Operator
- expressions: key (type: string), value (type: string)
- outputColumnNames: key, value
+ Reduce Output Operator
+ key expressions: key (type: string), value (type: string)
+ sort order: ++
+ Map-reduce partition columns: key (type: string), value (type: string)
Statistics: Num rows: 221 Data size: 2347 Basic stats: COMPLETE Column stats: NONE
- Reduce Output Operator
- key expressions: key (type: string), value (type: string)
- sort order: ++
- Map-reduce partition columns: key (type: string), value (type: string)
- Statistics: Num rows: 221 Data size: 2347 Basic stats: COMPLETE Column stats: NONE
Reducer 2
Reduce Operator Tree:
Forward
@@ -619,18 +615,14 @@ STAGE PLANS:
mode: complete
outputColumnNames: _col0, _col1
Statistics: Num rows: 36 Data size: 382 Basic stats: COMPLETE Column stats: NONE
- Select Operator
- expressions: _col0 (type: string), _col1 (type: string)
- outputColumnNames: _col0, _col1
+ File Output Operator
+ compressed: false
Statistics: Num rows: 36 Data size: 382 Basic stats: COMPLETE Column stats: NONE
- File Output Operator
- compressed: false
- Statistics: Num rows: 36 Data size: 382 Basic stats: COMPLETE Column stats: NONE
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- name: default.src_multi1
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ name: default.src_multi1
Filter Operator
predicate: ((KEY._col0 > 10) and (KEY._col0 < 20)) (type: boolean)
Statistics: Num rows: 24 Data size: 254 Basic stats: COMPLETE Column stats: NONE
@@ -639,18 +631,14 @@ STAGE PLANS:
mode: complete
outputColumnNames: _col0, _col1
Statistics: Num rows: 12 Data size: 127 Basic stats: COMPLETE Column stats: NONE
- Select Operator
- expressions: _col0 (type: string), _col1 (type: string)
- outputColumnNames: _col0, _col1
+ File Output Operator
+ compressed: false
Statistics: Num rows: 12 Data size: 127 Basic stats: COMPLETE Column stats: NONE
- File Output Operator
- compressed: false
- Statistics: Num rows: 12 Data size: 127 Basic stats: COMPLETE Column stats: NONE
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- name: default.src_multi2
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ name: default.src_multi2
Stage: Stage-0
Move Operator
@@ -756,15 +744,11 @@ STAGE PLANS:
Filter Operator
predicate: ((key < 10) or ((key > 10) and (key < 20))) (type: boolean)
Statistics: Num rows: 221 Data size: 2347 Basic stats: COMPLETE Column stats: NONE
- Select Operator
- expressions: key (type: string), value (type: string)
- outputColumnNames: key, value
+ Reduce Output Operator
+ key expressions: key (type: string), value (type: string)
+ sort order: ++
+ Map-reduce partition columns: key (type: string), value (type: string)
Statistics: Num rows: 221 Data size: 2347 Basic stats: COMPLETE Column stats: NONE
- Reduce Output Operator
- key expressions: key (type: string), value (type: string)
- sort order: ++
- Map-reduce partition columns: key (type: string), value (type: string)
- Statistics: Num rows: 221 Data size: 2347 Basic stats: COMPLETE Column stats: NONE
Reducer 2
Reduce Operator Tree:
Forward
@@ -777,18 +761,14 @@ STAGE PLANS:
mode: complete
outputColumnNames: _col0, _col1
Statistics: Num rows: 36 Data size: 382 Basic stats: COMPLETE Column stats: NONE
- Select Operator
- expressions: _col0 (type: string), _col1 (type: string)
- outputColumnNames: _col0, _col1
+ File Output Operator
+ compressed: false
Statistics: Num rows: 36 Data size: 382 Basic stats: COMPLETE Column stats: NONE
- File Output Operator
- compressed: false
- Statistics: Num rows: 36 Data size: 382 Basic stats: COMPLETE Column stats: NONE
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- name: default.src_multi1
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ name: default.src_multi1
Filter Operator
predicate: ((KEY._col0 > 10) and (KEY._col0 < 20)) (type: boolean)
Statistics: Num rows: 24 Data size: 254 Basic stats: COMPLETE Column stats: NONE
@@ -797,18 +777,14 @@ STAGE PLANS:
mode: complete
outputColumnNames: _col0, _col1
Statistics: Num rows: 12 Data size: 127 Basic stats: COMPLETE Column stats: NONE
- Select Operator
- expressions: _col0 (type: string), _col1 (type: string)
- outputColumnNames: _col0, _col1
+ File Output Operator
+ compressed: false
Statistics: Num rows: 12 Data size: 127 Basic stats: COMPLETE Column stats: NONE
- File Output Operator
- compressed: false
- Statistics: Num rows: 12 Data size: 127 Basic stats: COMPLETE Column stats: NONE
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- name: default.src_multi2
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ name: default.src_multi2
Stage: Stage-0
Move Operator
@@ -914,15 +890,11 @@ STAGE PLANS:
Filter Operator
predicate: ((key < 10) or ((key > 10) and (key < 20))) (type: boolean)
Statistics: Num rows: 221 Data size: 2347 Basic stats: COMPLETE Column stats: NONE
- Select Operator
- expressions: key (type: string), value (type: string)
- outputColumnNames: key, value
+ Reduce Output Operator
+ key expressions: key (type: string), value (type: string)
+ sort order: ++
+ Map-reduce partition columns: key (type: string), value (type: string)
Statistics: Num rows: 221 Data size: 2347 Basic stats: COMPLETE Column stats: NONE
- Reduce Output Operator
- key expressions: key (type: string), value (type: string)
- sort order: ++
- Map-reduce partition columns: key (type: string), value (type: string)
- Statistics: Num rows: 221 Data size: 2347 Basic stats: COMPLETE Column stats: NONE
Reducer 2
Reduce Operator Tree:
Forward
@@ -935,18 +907,14 @@ STAGE PLANS:
mode: complete
outputColumnNames: _col0, _col1
Statistics: Num rows: 36 Data size: 382 Basic stats: COMPLETE Column stats: NONE
- Select Operator
- expressions: _col0 (type: string), _col1 (type: string)
- outputColumnNames: _col0, _col1
+ File Output Operator
+ compressed: false
Statistics: Num rows: 36 Data size: 382 Basic stats: COMPLETE Column stats: NONE
- File Output Operator
- compressed: false
- Statistics: Num rows: 36 Data size: 382 Basic stats: COMPLETE Column stats: NONE
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- name: default.src_multi1
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ name: default.src_multi1
Filter Operator
predicate: ((KEY._col0 > 10) and (KEY._col0 < 20)) (type: boolean)
Statistics: Num rows: 24 Data size: 254 Basic stats: COMPLETE Column stats: NONE
@@ -955,18 +923,14 @@ STAGE PLANS:
mode: complete
outputColumnNames: _col0, _col1
Statistics: Num rows: 12 Data size: 127 Basic stats: COMPLETE Column stats: NONE
- Select Operator
- expressions: _col0 (type: string), _col1 (type: string)
- outputColumnNames: _col0, _col1
+ File Output Operator
+ compressed: false
Statistics: Num rows: 12 Data size: 127 Basic stats: COMPLETE Column stats: NONE
- File Output Operator
- compressed: false
- Statistics: Num rows: 12 Data size: 127 Basic stats: COMPLETE Column stats: NONE
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- name: default.src_multi2
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ name: default.src_multi2
Stage: Stage-0
Move Operator
@@ -1072,15 +1036,11 @@ STAGE PLANS:
Filter Operator
predicate: ((key < 10) or ((key > 10) and (key < 20))) (type: boolean)
Statistics: Num rows: 221 Data size: 2347 Basic stats: COMPLETE Column stats: NONE
- Select Operator
- expressions: key (type: string), value (type: string)
- outputColumnNames: key, value
+ Reduce Output Operator
+ key expressions: key (type: string), value (type: string)
+ sort order: ++
+ Map-reduce partition columns: key (type: string), value (type: string)
Statistics: Num rows: 221 Data size: 2347 Basic stats: COMPLETE Column stats: NONE
- Reduce Output Operator
- key expressions: key (type: string), value (type: string)
- sort order: ++
- Map-reduce partition columns: key (type: string), value (type: string)
- Statistics: Num rows: 221 Data size: 2347 Basic stats: COMPLETE Column stats: NONE
Reducer 2
Reduce Operator Tree:
Forward
@@ -1093,18 +1053,14 @@ STAGE PLANS:
mode: complete
outputColumnNames: _col0, _col1
Statistics: Num rows: 36 Data size: 382 Basic stats: COMPLETE Column stats: NONE
- Select Operator
- expressions: _col0 (type: string), _col1 (type: string)
- outputColumnNames: _col0, _col1
+ File Output Operator
+ compressed: false
Statistics: Num rows: 36 Data size: 382 Basic stats: COMPLETE Column stats: NONE
- File Output Operator
- compressed: false
- Statistics: Num rows: 36 Data size: 382 Basic stats: COMPLETE Column stats: NONE
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- name: default.src_multi1
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ name: default.src_multi1
Filter Operator
predicate: ((KEY._col0 > 10) and (KEY._col0 < 20)) (type: boolean)
Statistics: Num rows: 24 Data size: 254 Basic stats: COMPLETE Column stats: NONE
@@ -1113,18 +1069,14 @@ STAGE PLANS:
mode: complete
outputColumnNames: _col0, _col1
Statistics: Num rows: 12 Data size: 127 Basic stats: COMPLETE Column stats: NONE
- Select Operator
- expressions: _col0 (type: string), _col1 (type: string)
- outputColumnNames: _col0, _col1
+ File Output Operator
+ compressed: false
Statistics: Num rows: 12 Data size: 127 Basic stats: COMPLETE Column stats: NONE
- File Output Operator
- compressed: false
- Statistics: Num rows: 12 Data size: 127 Basic stats: COMPLETE Column stats: NONE
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- name: default.src_multi2
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ name: default.src_multi2
Stage: Stage-0
Move Operator
@@ -1231,28 +1183,22 @@ STAGE PLANS:
outputColumnNames: _col0, _col1
Filter Operator
predicate: (_col0 < 10) (type: boolean)
- Select Operator
- expressions: _col0 (type: string), _col1 (type: string)
- outputColumnNames: _col0, _col1
- File Output Operator
- compressed: false
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- name: default.src_multi1
+ File Output Operator
+ compressed: false
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ name: default.src_multi1
Filter Operator
predicate: ((_col0 > 10) and (_col0 < 20)) (type: boolean)
- Select Operator
- expressions: _col0 (type: string), _col1 (type: string)
- outputColumnNames: _col0, _col1
- File Output Operator
- compressed: false
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- name: default.src_multi2
+ File Output Operator
+ compressed: false
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ name: default.src_multi2
Map 3
Map Operator Tree:
TableScan
@@ -1262,28 +1208,22 @@ STAGE PLANS:
outputColumnNames: _col0, _col1
Filter Operator
predicate: (_col0 < 10) (type: boolean)
- Select Operator
- expressions: _col0 (type: string), _col1 (type: string)
- outputColumnNames: _col0, _col1
- File Output Operator
- compressed: false
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- name: default.src_multi1
+ File Output Operator
+ compressed: false
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ name: default.src_multi1
Filter Operator
predicate: ((_col0 > 10) and (_col0 < 20)) (type: boolean)
- Select Operator
- expressions: _col0 (type: string), _col1 (type: string)
- outputColumnNames: _col0, _col1
- File Output Operator
- compressed: false
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- name: default.src_multi2
+ File Output Operator
+ compressed: false
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ name: default.src_multi2
Union 2
Vertex: Union 2
@@ -1418,28 +1358,22 @@ STAGE PLANS:
outputColumnNames: _col0, _col1
Filter Operator
predicate: (_col0 < 10) (type: boolean)
- Select Operator
- expressions: _col0 (type: string), _col1 (type: string)
- outputColumnNames: _col0, _col1
- File Output Operator
- compressed: false
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- name: default.src_multi1
+ File Output Operator
+ compressed: false
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ name: default.src_multi1
Filter Operator
predicate: ((_col0 > 10) and (_col0 < 20)) (type: boolean)
- Select Operator
- expressions: _col0 (type: string), _col1 (type: string)
- outputColumnNames: _col0, _col1
- File Output Operator
- compressed: false
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- name: default.src_multi2
+ File Output Operator
+ compressed: false
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ name: default.src_multi2
Map 3
Map Operator Tree:
TableScan
@@ -1449,28 +1383,22 @@ STAGE PLANS:
outputColumnNames: _col0, _col1
Filter Operator
predicate: (_col0 < 10) (type: boolean)
- Select Operator
- expressions: _col0 (type: string), _col1 (type: string)
- outputColumnNames: _col0, _col1
- File Output Operator
- compressed: false
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- name: default.src_multi1
+ File Output Operator
+ compressed: false
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ name: default.src_multi1
Filter Operator
predicate: ((_col0 > 10) and (_col0 < 20)) (type: boolean)
- Select Operator
- expressions: _col0 (type: string), _col1 (type: string)
- outputColumnNames: _col0, _col1
- File Output Operator
- compressed: false
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- name: default.src_multi2
+ File Output Operator
+ compressed: false
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ name: default.src_multi2
Union 2
Vertex: Union 2
@@ -1605,28 +1533,22 @@ STAGE PLANS:
outputColumnNames: _col0, _col1
Filter Operator
predicate: (_col0 < 10) (type: boolean)
- Select Operator
- expressions: _col0 (type: string), _col1 (type: string)
- outputColumnNames: _col0, _col1
- File Output Operator
- compressed: false
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- name: default.src_multi1
+ File Output Operator
+ compressed: false
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ name: default.src_multi1
Filter Operator
predicate: ((_col0 > 10) and (_col0 < 20)) (type: boolean)
- Select Operator
- expressions: _col0 (type: string), _col1 (type: string)
- outputColumnNames: _col0, _col1
- File Output Operator
- compressed: false
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- name: default.src_multi2
+ File Output Operator
+ compressed: false
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ name: default.src_multi2
Map 3
Map Operator Tree:
TableScan
@@ -1636,28 +1558,22 @@ STAGE PLANS:
outputColumnNames: _col0, _col1
Filter Operator
predicate: (_col0 < 10) (type: boolean)
- Select Operator
- expressions: _col0 (type: string), _col1 (type: string)
- outputColumnNames: _col0, _col1
- File Output Operator
- compressed: false
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- name: default.src_multi1
+ File Output Operator
+ compressed: false
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ name: default.src_multi1
Filter Operator
predicate: ((_col0 > 10) and (_col0 < 20)) (type: boolean)
- Select Operator
- expressions: _col0 (type: string), _col1 (type: string)
- outputColumnNames: _col0, _col1
- File Output Operator
- compressed: false
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- name: default.src_multi2
+ File Output Operator
+ compressed: false
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ name: default.src_multi2
Union 2
Vertex: Union 2
@@ -1792,28 +1708,22 @@ STAGE PLANS:
outputColumnNames: _col0, _col1
Filter Operator
predicate: (_col0 < 10) (type: boolean)
- Select Operator
- expressions: _col0 (type: string), _col1 (type: string)
- outputColumnNames: _col0, _col1
- File Output Operator
- compressed: false
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- name: default.src_multi1
+ File Output Operator
+ compressed: false
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ name: default.src_multi1
Filter Operator
predicate: ((_col0 > 10) and (_col0 < 20)) (type: boolean)
- Select Operator
- expressions: _col0 (type: string), _col1 (type: string)
- outputColumnNames: _col0, _col1
- File Output Operator
- compressed: false
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- name: default.src_multi2
+ File Output Operator
+ compressed: false
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ name: default.src_multi2
Map 3
Map Operator Tree:
TableScan
@@ -1823,28 +1733,22 @@ STAGE PLANS:
outputColumnNames: _col0, _col1
Filter Operator
predicate: (_col0 < 10) (type: boolean)
- Select Operator
- expressions: _col0 (type: string), _col1 (type: string)
- outputColumnNames: _col0, _col1
- File Output Operator
- compressed: false
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- name: default.src_multi1
+ File Output Operator
+ compressed: false
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ name: default.src_multi1
Filter Operator
predicate: ((_col0 > 10) and (_col0 < 20)) (type: boolean)
- Select Operator
- expressions: _col0 (type: string), _col1 (type: string)
- outputColumnNames: _col0, _col1
- File Output Operator
- compressed: false
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- name: default.src_multi2
+ File Output Operator
+ compressed: false
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ name: default.src_multi2
Union 2
Vertex: Union 2
Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/multi_insert_gby.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/multi_insert_gby.q.out?rev=1642395&r1=1642394&r2=1642395&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/spark/multi_insert_gby.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/spark/multi_insert_gby.q.out Sat Nov 29 03:44:22 2014
@@ -54,15 +54,11 @@ STAGE PLANS:
Filter Operator
predicate: ((key > 450) or (key > 500)) (type: boolean)
Statistics: Num rows: 332 Data size: 3527 Basic stats: COMPLETE Column stats: NONE
- Select Operator
- expressions: key (type: string)
- outputColumnNames: key
+ Reduce Output Operator
+ key expressions: key (type: string)
+ sort order: +
+ Map-reduce partition columns: key (type: string)
Statistics: Num rows: 332 Data size: 3527 Basic stats: COMPLETE Column stats: NONE
- Reduce Output Operator
- key expressions: key (type: string)
- sort order: +
- Map-reduce partition columns: key (type: string)
- Statistics: Num rows: 332 Data size: 3527 Basic stats: COMPLETE Column stats: NONE
Reducer 2
Reduce Operator Tree:
Forward
Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/multi_insert_gby2.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/multi_insert_gby2.q.out?rev=1642395&r1=1642394&r2=1642395&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/spark/multi_insert_gby2.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/spark/multi_insert_gby2.q.out Sat Nov 29 03:44:22 2014
@@ -95,18 +95,14 @@ STAGE PLANS:
mode: complete
outputColumnNames: _col0
Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE
- Select Operator
- expressions: _col0 (type: double)
- outputColumnNames: _col0
+ File Output Operator
+ compressed: false
Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE
- File Output Operator
- compressed: false
- Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- name: default.e2
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ name: default.e2
Stage: Stage-0
Move Operator
Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/multi_insert_gby3.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/multi_insert_gby3.q.out?rev=1642395&r1=1642394&r2=1642395&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/spark/multi_insert_gby3.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/spark/multi_insert_gby3.q.out Sat Nov 29 03:44:22 2014
@@ -1774,38 +1774,30 @@ STAGE PLANS:
expressions: KEY.reducesinkkey0 (type: string), VALUE._col0 (type: double), VALUE._col1 (type: string)
outputColumnNames: _col0, _col1, _col2
Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
- Select Operator
- expressions: _col0 (type: string), _col1 (type: double), _col2 (type: string)
- outputColumnNames: _col0, _col1, _col2
- Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
- Reduce Output Operator
- key expressions: _col0 (type: string), _col2 (type: string)
- sort order: ++
- Map-reduce partition columns: _col0 (type: string)
- Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
- value expressions: _col1 (type: double)
- Select Operator
- expressions: _col0 (type: string), _col1 (type: double), _col2 (type: string)
- outputColumnNames: _col0, _col1, _col2
+ Reduce Output Operator
+ key expressions: _col0 (type: string), _col2 (type: string)
+ sort order: ++
+ Map-reduce partition columns: _col0 (type: string)
Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
- Group By Operator
- aggregations: count(DISTINCT _col1)
- keys: _col0 (type: string), _col1 (type: double), _col2 (type: string)
- mode: complete
- outputColumnNames: _col0, _col1, _col2, _col3
+ value expressions: _col1 (type: double)
+ Group By Operator
+ aggregations: count(DISTINCT _col1)
+ keys: _col0 (type: string), _col1 (type: double), _col2 (type: string)
+ mode: complete
+ outputColumnNames: _col0, _col1, _col2, _col3
+ Statistics: Num rows: 250 Data size: 2656 Basic stats: COMPLETE Column stats: NONE
+ Select Operator
+ expressions: _col0 (type: string), UDFToDouble(_col3) (type: double)
+ outputColumnNames: _col0, _col1
Statistics: Num rows: 250 Data size: 2656 Basic stats: COMPLETE Column stats: NONE
- Select Operator
- expressions: _col0 (type: string), UDFToDouble(_col3) (type: double)
- outputColumnNames: _col0, _col1
+ File Output Operator
+ compressed: false
Statistics: Num rows: 250 Data size: 2656 Basic stats: COMPLETE Column stats: NONE
- File Output Operator
- compressed: false
- Statistics: Num rows: 250 Data size: 2656 Basic stats: COMPLETE Column stats: NONE
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- name: default.e3
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ name: default.e3
Reducer 3
Reduce Operator Tree:
Forward