You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hive.apache.org by br...@apache.org on 2014/11/29 04:44:28 UTC
svn commit: r1642395 [5/22] - in /hive/branches/spark/ql/src:
java/org/apache/hadoop/hive/ql/exec/spark/
java/org/apache/hadoop/hive/ql/exec/spark/session/
test/results/clientpositive/ test/results/clientpositive/spark/
Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/bucket_map_join_tez1.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/bucket_map_join_tez1.q.out?rev=1642395&r1=1642394&r2=1642395&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/spark/bucket_map_join_tez1.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/spark/bucket_map_join_tez1.q.out Sat Nov 29 03:44:22 2014
@@ -126,7 +126,7 @@ STAGE PLANS:
Spark
#### A masked pattern was here ####
Vertices:
- Map 2
+ Map 1
Map Operator Tree:
TableScan
alias: a
@@ -148,7 +148,7 @@ STAGE PLANS:
Spark
#### A masked pattern was here ####
Vertices:
- Map 1
+ Map 2
Map Operator Tree:
TableScan
alias: b
@@ -167,7 +167,7 @@ STAGE PLANS:
1 key (type: int)
outputColumnNames: _col0, _col1, _col7
input vertices:
- 0 Map 2
+ 0 Map 1
Statistics: Num rows: 275 Data size: 2921 Basic stats: COMPLETE Column stats: NONE
Select Operator
expressions: _col0 (type: int), _col1 (type: string), _col7 (type: string)
@@ -212,10 +212,10 @@ STAGE PLANS:
Stage: Stage-2
Spark
Edges:
- Reducer 3 <- Map 2 (GROUP, 3)
+ Reducer 2 <- Map 1 (GROUP, 3)
#### A masked pattern was here ####
Vertices:
- Map 2
+ Map 1
Map Operator Tree:
TableScan
alias: srcbucket_mapjoin
@@ -223,23 +223,19 @@ STAGE PLANS:
Filter Operator
predicate: key is not null (type: boolean)
Statistics: Num rows: 14 Data size: 1456 Basic stats: COMPLETE Column stats: NONE
- Select Operator
- expressions: key (type: int), value (type: string)
- outputColumnNames: key, value
+ Group By Operator
+ aggregations: sum(substr(value, 5))
+ keys: key (type: int)
+ mode: hash
+ outputColumnNames: _col0, _col1
Statistics: Num rows: 14 Data size: 1456 Basic stats: COMPLETE Column stats: NONE
- Group By Operator
- aggregations: sum(substr(value, 5))
- keys: key (type: int)
- mode: hash
- outputColumnNames: _col0, _col1
+ Reduce Output Operator
+ key expressions: _col0 (type: int)
+ sort order: +
+ Map-reduce partition columns: _col0 (type: int)
Statistics: Num rows: 14 Data size: 1456 Basic stats: COMPLETE Column stats: NONE
- Reduce Output Operator
- key expressions: _col0 (type: int)
- sort order: +
- Map-reduce partition columns: _col0 (type: int)
- Statistics: Num rows: 14 Data size: 1456 Basic stats: COMPLETE Column stats: NONE
- value expressions: _col1 (type: double)
- Reducer 3
+ value expressions: _col1 (type: double)
+ Reducer 2
Local Work:
Map Reduce Local Work
Reduce Operator Tree:
@@ -265,7 +261,7 @@ STAGE PLANS:
Spark
#### A masked pattern was here ####
Vertices:
- Map 1
+ Map 3
Map Operator Tree:
TableScan
alias: b
@@ -284,7 +280,7 @@ STAGE PLANS:
1 key (type: int)
outputColumnNames: _col0, _col1, _col3
input vertices:
- 0 Reducer 3
+ 0 Reducer 2
Statistics: Num rows: 133 Data size: 1411 Basic stats: COMPLETE Column stats: NONE
Select Operator
expressions: _col1 (type: int), _col0 (type: double), _col3 (type: string)
@@ -327,7 +323,7 @@ STAGE PLANS:
Spark
#### A masked pattern was here ####
Vertices:
- Map 1
+ Map 4
Map Operator Tree:
TableScan
alias: b
@@ -348,10 +344,10 @@ STAGE PLANS:
Stage: Stage-1
Spark
Edges:
- Reducer 3 <- Map 2 (GROUP, 3)
+ Reducer 2 <- Map 1 (GROUP, 3)
#### A masked pattern was here ####
Vertices:
- Map 2
+ Map 1
Map Operator Tree:
TableScan
alias: tab_part
@@ -370,27 +366,23 @@ STAGE PLANS:
1 key (type: int)
outputColumnNames: _col6, _col7
input vertices:
- 1 Map 4
+ 1 Map 3
Statistics: Num rows: 275 Data size: 2921 Basic stats: COMPLETE Column stats: NONE
- Select Operator
- expressions: _col6 (type: int), _col7 (type: string)
- outputColumnNames: _col6, _col7
+ Group By Operator
+ aggregations: sum(substr(_col7, 5))
+ keys: _col6 (type: int)
+ mode: hash
+ outputColumnNames: _col0, _col1
Statistics: Num rows: 275 Data size: 2921 Basic stats: COMPLETE Column stats: NONE
- Group By Operator
- aggregations: sum(substr(_col7, 5))
- keys: _col6 (type: int)
- mode: hash
- outputColumnNames: _col0, _col1
+ Reduce Output Operator
+ key expressions: _col0 (type: int)
+ sort order: +
+ Map-reduce partition columns: _col0 (type: int)
Statistics: Num rows: 275 Data size: 2921 Basic stats: COMPLETE Column stats: NONE
- Reduce Output Operator
- key expressions: _col0 (type: int)
- sort order: +
- Map-reduce partition columns: _col0 (type: int)
- Statistics: Num rows: 275 Data size: 2921 Basic stats: COMPLETE Column stats: NONE
- value expressions: _col1 (type: double)
+ value expressions: _col1 (type: double)
Local Work:
Map Reduce Local Work
- Reducer 3
+ Reducer 2
Local Work:
Map Reduce Local Work
Reduce Operator Tree:
@@ -415,7 +407,7 @@ STAGE PLANS:
1 key (type: int)
outputColumnNames: _col0, _col1, _col3
input vertices:
- 1 Map 1
+ 1 Map 4
Statistics: Num rows: 150 Data size: 1600 Basic stats: COMPLETE Column stats: NONE
Select Operator
expressions: _col1 (type: int), _col0 (type: double), _col3 (type: string)
@@ -433,7 +425,7 @@ STAGE PLANS:
Spark
#### A masked pattern was here ####
Vertices:
- Map 4
+ Map 3
Map Operator Tree:
TableScan
alias: tab
@@ -478,7 +470,7 @@ STAGE PLANS:
Spark
#### A masked pattern was here ####
Vertices:
- Map 4
+ Map 3
Map Operator Tree:
TableScan
alias: y
@@ -499,10 +491,10 @@ STAGE PLANS:
Stage: Stage-2
Spark
Edges:
- Reducer 3 <- Map 2 (GROUP, 3)
+ Reducer 2 <- Map 1 (GROUP, 3)
#### A masked pattern was here ####
Vertices:
- Map 2
+ Map 1
Map Operator Tree:
TableScan
alias: x
@@ -521,27 +513,23 @@ STAGE PLANS:
1 key (type: int)
outputColumnNames: _col0, _col1
input vertices:
- 1 Map 4
+ 1 Map 3
Statistics: Num rows: 133 Data size: 1411 Basic stats: COMPLETE Column stats: NONE
- Select Operator
- expressions: _col0 (type: int), _col1 (type: string)
+ Group By Operator
+ aggregations: sum(substr(_col1, 5))
+ keys: _col0 (type: int)
+ mode: hash
outputColumnNames: _col0, _col1
Statistics: Num rows: 133 Data size: 1411 Basic stats: COMPLETE Column stats: NONE
- Group By Operator
- aggregations: sum(substr(_col1, 5))
- keys: _col0 (type: int)
- mode: hash
- outputColumnNames: _col0, _col1
+ Reduce Output Operator
+ key expressions: _col0 (type: int)
+ sort order: +
+ Map-reduce partition columns: _col0 (type: int)
Statistics: Num rows: 133 Data size: 1411 Basic stats: COMPLETE Column stats: NONE
- Reduce Output Operator
- key expressions: _col0 (type: int)
- sort order: +
- Map-reduce partition columns: _col0 (type: int)
- Statistics: Num rows: 133 Data size: 1411 Basic stats: COMPLETE Column stats: NONE
- value expressions: _col1 (type: double)
+ value expressions: _col1 (type: double)
Local Work:
Map Reduce Local Work
- Reducer 3
+ Reducer 2
Local Work:
Map Reduce Local Work
Reduce Operator Tree:
@@ -567,7 +555,7 @@ STAGE PLANS:
Spark
#### A masked pattern was here ####
Vertices:
- Map 1
+ Map 4
Map Operator Tree:
TableScan
alias: b
@@ -586,7 +574,7 @@ STAGE PLANS:
1 key (type: int)
outputColumnNames: _col0, _col1, _col3
input vertices:
- 0 Reducer 3
+ 0 Reducer 2
Statistics: Num rows: 275 Data size: 2921 Basic stats: COMPLETE Column stats: NONE
Select Operator
expressions: _col1 (type: int), _col0 (type: double), _col3 (type: string)
@@ -628,7 +616,7 @@ STAGE PLANS:
Spark
#### A masked pattern was here ####
Vertices:
- Map 1
+ Map 2
Map Operator Tree:
TableScan
alias: b
@@ -647,7 +635,7 @@ STAGE PLANS:
2 key (type: int)
Local Work:
Map Reduce Local Work
- Map 2
+ Map 3
Map Operator Tree:
TableScan
alias: c
@@ -671,7 +659,7 @@ STAGE PLANS:
Spark
#### A masked pattern was here ####
Vertices:
- Map 3
+ Map 1
Map Operator Tree:
TableScan
alias: a
@@ -693,8 +681,8 @@ STAGE PLANS:
2 key (type: int)
outputColumnNames: _col0, _col1, _col7
input vertices:
- 1 Map 1
- 2 Map 2
+ 1 Map 2
+ 2 Map 3
Statistics: Num rows: 550 Data size: 5843 Basic stats: COMPLETE Column stats: NONE
Select Operator
expressions: _col0 (type: int), _col1 (type: string), _col7 (type: string)
@@ -734,37 +722,37 @@ STAGE PLANS:
Spark
#### A masked pattern was here ####
Vertices:
- Map 1
+ Map 2
Map Operator Tree:
TableScan
- alias: c
+ alias: y
Statistics: Num rows: 242 Data size: 2566 Basic stats: COMPLETE Column stats: NONE
Filter Operator
predicate: key is not null (type: boolean)
Statistics: Num rows: 121 Data size: 1283 Basic stats: COMPLETE Column stats: NONE
Spark HashTable Sink Operator
condition expressions:
- 0 {_col0} {_col1}
- 1 {value}
+ 0 {key} {value}
+ 1
keys:
- 0 _col0 (type: int)
+ 0 key (type: int)
1 key (type: int)
Local Work:
Map Reduce Local Work
Map 3
Map Operator Tree:
TableScan
- alias: y
+ alias: c
Statistics: Num rows: 242 Data size: 2566 Basic stats: COMPLETE Column stats: NONE
Filter Operator
predicate: key is not null (type: boolean)
Statistics: Num rows: 121 Data size: 1283 Basic stats: COMPLETE Column stats: NONE
Spark HashTable Sink Operator
condition expressions:
- 0 {key} {value}
- 1
+ 0 {_col0} {_col1}
+ 1 {value}
keys:
- 0 key (type: int)
+ 0 _col0 (type: int)
1 key (type: int)
Local Work:
Map Reduce Local Work
@@ -773,7 +761,7 @@ STAGE PLANS:
Spark
#### A masked pattern was here ####
Vertices:
- Map 2
+ Map 1
Map Operator Tree:
TableScan
alias: x
@@ -792,36 +780,32 @@ STAGE PLANS:
1 key (type: int)
outputColumnNames: _col0, _col1
input vertices:
- 1 Map 3
+ 1 Map 2
Statistics: Num rows: 275 Data size: 2921 Basic stats: COMPLETE Column stats: NONE
- Select Operator
- expressions: _col0 (type: int), _col1 (type: string)
- outputColumnNames: _col0, _col1
- Statistics: Num rows: 275 Data size: 2921 Basic stats: COMPLETE Column stats: NONE
- Map Join Operator
- condition map:
- Inner Join 0 to 1
- condition expressions:
- 0 {_col0} {_col1}
- 1 {value}
- keys:
- 0 _col0 (type: int)
- 1 key (type: int)
- outputColumnNames: _col0, _col1, _col3
- input vertices:
- 1 Map 1
+ Map Join Operator
+ condition map:
+ Inner Join 0 to 1
+ condition expressions:
+ 0 {_col0} {_col1}
+ 1 {value}
+ keys:
+ 0 _col0 (type: int)
+ 1 key (type: int)
+ outputColumnNames: _col0, _col1, _col3
+ input vertices:
+ 1 Map 3
+ Statistics: Num rows: 302 Data size: 3213 Basic stats: COMPLETE Column stats: NONE
+ Select Operator
+ expressions: _col0 (type: int), _col1 (type: string), _col3 (type: string)
+ outputColumnNames: _col0, _col1, _col2
Statistics: Num rows: 302 Data size: 3213 Basic stats: COMPLETE Column stats: NONE
- Select Operator
- expressions: _col0 (type: int), _col1 (type: string), _col3 (type: string)
- outputColumnNames: _col0, _col1, _col2
+ File Output Operator
+ compressed: false
Statistics: Num rows: 302 Data size: 3213 Basic stats: COMPLETE Column stats: NONE
- File Output Operator
- compressed: false
- Statistics: Num rows: 302 Data size: 3213 Basic stats: COMPLETE Column stats: NONE
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
Local Work:
Map Reduce Local Work
@@ -852,10 +836,10 @@ STAGE PLANS:
Stage: Stage-2
Spark
Edges:
- Reducer 3 <- Map 2 (GROUP, 3)
+ Reducer 2 <- Map 1 (GROUP, 3)
#### A masked pattern was here ####
Vertices:
- Map 2
+ Map 1
Map Operator Tree:
TableScan
alias: srcbucket_mapjoin
@@ -863,23 +847,19 @@ STAGE PLANS:
Filter Operator
predicate: key is not null (type: boolean)
Statistics: Num rows: 14 Data size: 1456 Basic stats: COMPLETE Column stats: NONE
- Select Operator
- expressions: key (type: int), value (type: string)
- outputColumnNames: key, value
+ Group By Operator
+ aggregations: sum(substr(value, 5))
+ keys: key (type: int)
+ mode: hash
+ outputColumnNames: _col0, _col1
Statistics: Num rows: 14 Data size: 1456 Basic stats: COMPLETE Column stats: NONE
- Group By Operator
- aggregations: sum(substr(value, 5))
- keys: key (type: int)
- mode: hash
- outputColumnNames: _col0, _col1
+ Reduce Output Operator
+ key expressions: _col0 (type: int)
+ sort order: +
+ Map-reduce partition columns: _col0 (type: int)
Statistics: Num rows: 14 Data size: 1456 Basic stats: COMPLETE Column stats: NONE
- Reduce Output Operator
- key expressions: _col0 (type: int)
- sort order: +
- Map-reduce partition columns: _col0 (type: int)
- Statistics: Num rows: 14 Data size: 1456 Basic stats: COMPLETE Column stats: NONE
- value expressions: _col1 (type: double)
- Reducer 3
+ value expressions: _col1 (type: double)
+ Reducer 2
Local Work:
Map Reduce Local Work
Reduce Operator Tree:
@@ -888,24 +868,20 @@ STAGE PLANS:
keys: KEY._col0 (type: int)
mode: mergepartial
outputColumnNames: _col0, _col1
- Statistics: Num rows: 7 Data size: 728 Basic stats: COMPLETE Column stats: NONE
- Select Operator
- expressions: _col0 (type: int), _col1 (type: double)
- outputColumnNames: _col0, _col1
- Statistics: Num rows: 7 Data size: 728 Basic stats: COMPLETE Column stats: NONE
- Spark HashTable Sink Operator
- condition expressions:
- 0 {_col1}
- 1 {value}
- keys:
- 0 _col0 (type: int)
- 1 key (type: int)
+ Statistics: Num rows: 14 Data size: 1456 Basic stats: COMPLETE Column stats: NONE
+ Spark HashTable Sink Operator
+ condition expressions:
+ 0 {_col1}
+ 1 {value}
+ keys:
+ 0 _col0 (type: int)
+ 1 key (type: int)
Stage: Stage-1
Spark
#### A masked pattern was here ####
Vertices:
- Map 1
+ Map 3
Map Operator Tree:
TableScan
alias: b
@@ -924,7 +900,7 @@ STAGE PLANS:
1 key (type: int)
outputColumnNames: _col0, _col1, _col3
input vertices:
- 0 Reducer 3
+ 0 Reducer 2
Statistics: Num rows: 275 Data size: 2921 Basic stats: COMPLETE Column stats: NONE
Select Operator
expressions: _col0 (type: int), _col1 (type: double), _col3 (type: string)
@@ -965,10 +941,10 @@ STAGE PLANS:
Stage: Stage-2
Spark
Edges:
- Reducer 3 <- Map 2 (GROUP, 3)
+ Reducer 2 <- Map 1 (GROUP, 3)
#### A masked pattern was here ####
Vertices:
- Map 2
+ Map 1
Map Operator Tree:
TableScan
alias: srcbucket_mapjoin
@@ -976,17 +952,13 @@ STAGE PLANS:
Filter Operator
predicate: key is not null (type: boolean)
Statistics: Num rows: 14 Data size: 1456 Basic stats: COMPLETE Column stats: NONE
- Select Operator
- expressions: key (type: int), value (type: string)
- outputColumnNames: key, value
+ Reduce Output Operator
+ key expressions: key (type: int)
+ sort order: +
+ Map-reduce partition columns: key (type: int)
Statistics: Num rows: 14 Data size: 1456 Basic stats: COMPLETE Column stats: NONE
- Reduce Output Operator
- key expressions: key (type: int)
- sort order: +
- Map-reduce partition columns: key (type: int)
- Statistics: Num rows: 14 Data size: 1456 Basic stats: COMPLETE Column stats: NONE
- value expressions: substr(value, 5) (type: string)
- Reducer 3
+ value expressions: substr(value, 5) (type: string)
+ Reducer 2
Local Work:
Map Reduce Local Work
Reduce Operator Tree:
@@ -995,24 +967,20 @@ STAGE PLANS:
keys: KEY._col0 (type: int)
mode: complete
outputColumnNames: _col0, _col1
- Statistics: Num rows: 7 Data size: 728 Basic stats: COMPLETE Column stats: NONE
- Select Operator
- expressions: _col0 (type: int), _col1 (type: double)
- outputColumnNames: _col0, _col1
- Statistics: Num rows: 7 Data size: 728 Basic stats: COMPLETE Column stats: NONE
- Spark HashTable Sink Operator
- condition expressions:
- 0 {_col1}
- 1 {value}
- keys:
- 0 _col0 (type: int)
- 1 key (type: int)
+ Statistics: Num rows: 14 Data size: 1456 Basic stats: COMPLETE Column stats: NONE
+ Spark HashTable Sink Operator
+ condition expressions:
+ 0 {_col1}
+ 1 {value}
+ keys:
+ 0 _col0 (type: int)
+ 1 key (type: int)
Stage: Stage-1
Spark
#### A masked pattern was here ####
Vertices:
- Map 1
+ Map 3
Map Operator Tree:
TableScan
alias: b
@@ -1031,7 +999,7 @@ STAGE PLANS:
1 key (type: int)
outputColumnNames: _col0, _col1, _col3
input vertices:
- 0 Reducer 3
+ 0 Reducer 2
Statistics: Num rows: 275 Data size: 2921 Basic stats: COMPLETE Column stats: NONE
Select Operator
expressions: _col0 (type: int), _col1 (type: double), _col3 (type: string)
@@ -1073,7 +1041,7 @@ STAGE PLANS:
Spark
#### A masked pattern was here ####
Vertices:
- Map 2
+ Map 1
Map Operator Tree:
TableScan
alias: a
@@ -1095,7 +1063,7 @@ STAGE PLANS:
Spark
#### A masked pattern was here ####
Vertices:
- Map 1
+ Map 2
Map Operator Tree:
TableScan
alias: b
@@ -1114,7 +1082,7 @@ STAGE PLANS:
1 value (type: string)
outputColumnNames: _col0, _col1, _col7
input vertices:
- 0 Map 2
+ 0 Map 1
Statistics: Num rows: 275 Data size: 2921 Basic stats: COMPLETE Column stats: NONE
Select Operator
expressions: _col0 (type: int), _col1 (type: string), _col7 (type: string)
@@ -1176,7 +1144,7 @@ STAGE PLANS:
Spark
#### A masked pattern was here ####
Vertices:
- Map 2
+ Map 1
Map Operator Tree:
TableScan
alias: a
@@ -1198,7 +1166,7 @@ STAGE PLANS:
Spark
#### A masked pattern was here ####
Vertices:
- Map 1
+ Map 2
Map Operator Tree:
TableScan
alias: b
@@ -1217,7 +1185,7 @@ STAGE PLANS:
1 key (type: int)
outputColumnNames: _col0, _col1, _col6
input vertices:
- 0 Map 2
+ 0 Map 1
Statistics: Num rows: 275 Data size: 2921 Basic stats: COMPLETE Column stats: NONE
Select Operator
expressions: _col0 (type: int), _col1 (type: string), _col6 (type: string)
@@ -1256,35 +1224,35 @@ STAGE PLANS:
Map 1
Map Operator Tree:
TableScan
- alias: b
+ alias: a
Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
Filter Operator
- predicate: value is not null (type: boolean)
- Statistics: Num rows: 250 Data size: 2656 Basic stats: COMPLETE Column stats: NONE
+ predicate: (key is not null and value is not null) (type: boolean)
+ Statistics: Num rows: 125 Data size: 1328 Basic stats: COMPLETE Column stats: NONE
Spark HashTable Sink Operator
condition expressions:
- 0 {_col0}
- 1 {key}
+ 0 {value}
+ 1
keys:
- 0 _col1 (type: string)
- 1 value (type: string)
+ 0 key (type: int)
+ 1 key (type: int)
Local Work:
Map Reduce Local Work
Map 3
Map Operator Tree:
TableScan
- alias: a
+ alias: b
Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
Filter Operator
- predicate: (key is not null and value is not null) (type: boolean)
- Statistics: Num rows: 125 Data size: 1328 Basic stats: COMPLETE Column stats: NONE
+ predicate: value is not null (type: boolean)
+ Statistics: Num rows: 250 Data size: 2656 Basic stats: COMPLETE Column stats: NONE
Spark HashTable Sink Operator
condition expressions:
- 0 {value}
- 1
+ 0 {_col0}
+ 1 {key}
keys:
- 0 key (type: int)
- 1 key (type: int)
+ 0 _col1 (type: string)
+ 1 value (type: string)
Local Work:
Map Reduce Local Work
@@ -1311,7 +1279,7 @@ STAGE PLANS:
1 key (type: int)
outputColumnNames: _col0, _col1
input vertices:
- 0 Map 3
+ 0 Map 1
Statistics: Num rows: 275 Data size: 2921 Basic stats: COMPLETE Column stats: NONE
Map Join Operator
condition map:
@@ -1324,7 +1292,7 @@ STAGE PLANS:
1 value (type: string)
outputColumnNames: _col0, _col12
input vertices:
- 1 Map 1
+ 1 Map 3
Statistics: Num rows: 302 Data size: 3213 Basic stats: COMPLETE Column stats: NONE
Select Operator
expressions: _col0 (type: int), _col12 (type: int)
Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/bucket_map_join_tez2.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/bucket_map_join_tez2.q.out?rev=1642395&r1=1642394&r2=1642395&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/spark/bucket_map_join_tez2.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/spark/bucket_map_join_tez2.q.out Sat Nov 29 03:44:22 2014
@@ -125,35 +125,35 @@ STAGE PLANS:
Map 1
Map Operator Tree:
TableScan
- alias: b
+ alias: a
Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
Filter Operator
- predicate: value is not null (type: boolean)
- Statistics: Num rows: 250 Data size: 2656 Basic stats: COMPLETE Column stats: NONE
+ predicate: (key is not null and value is not null) (type: boolean)
+ Statistics: Num rows: 125 Data size: 1328 Basic stats: COMPLETE Column stats: NONE
Spark HashTable Sink Operator
condition expressions:
- 0 {_col0}
- 1 {key}
+ 0 {value}
+ 1
keys:
- 0 _col1 (type: string)
- 1 value (type: string)
+ 0 key (type: int)
+ 1 key (type: int)
Local Work:
Map Reduce Local Work
Map 3
Map Operator Tree:
TableScan
- alias: a
+ alias: b
Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
Filter Operator
- predicate: (key is not null and value is not null) (type: boolean)
- Statistics: Num rows: 125 Data size: 1328 Basic stats: COMPLETE Column stats: NONE
+ predicate: value is not null (type: boolean)
+ Statistics: Num rows: 250 Data size: 2656 Basic stats: COMPLETE Column stats: NONE
Spark HashTable Sink Operator
condition expressions:
- 0 {value}
- 1
+ 0 {_col0}
+ 1 {key}
keys:
- 0 key (type: int)
- 1 key (type: int)
+ 0 _col1 (type: string)
+ 1 value (type: string)
Local Work:
Map Reduce Local Work
@@ -180,7 +180,7 @@ STAGE PLANS:
1 key (type: int)
outputColumnNames: _col0, _col1
input vertices:
- 0 Map 3
+ 0 Map 1
Statistics: Num rows: 275 Data size: 2921 Basic stats: COMPLETE Column stats: NONE
Map Join Operator
condition map:
@@ -193,7 +193,7 @@ STAGE PLANS:
1 value (type: string)
outputColumnNames: _col0, _col12
input vertices:
- 1 Map 1
+ 1 Map 3
Statistics: Num rows: 302 Data size: 3213 Basic stats: COMPLETE Column stats: NONE
Select Operator
expressions: _col0 (type: int), _col12 (type: int)
@@ -255,7 +255,7 @@ STAGE PLANS:
Spark
#### A masked pattern was here ####
Vertices:
- Map 2
+ Map 1
Map Operator Tree:
TableScan
alias: a
@@ -277,7 +277,7 @@ STAGE PLANS:
Spark
#### A masked pattern was here ####
Vertices:
- Map 1
+ Map 2
Map Operator Tree:
TableScan
alias: b
@@ -296,7 +296,7 @@ STAGE PLANS:
1 UDFToDouble(key) (type: double)
outputColumnNames: _col0, _col1, _col6
input vertices:
- 0 Map 2
+ 0 Map 1
Statistics: Num rows: 275 Data size: 2921 Basic stats: COMPLETE Column stats: NONE
Select Operator
expressions: _col0 (type: int), _col1 (type: string), _col6 (type: string)
@@ -591,91 +591,83 @@ STAGE DEPENDENCIES:
STAGE PLANS:
Stage: Stage-2
Spark
- Edges:
- Reducer 3 <- Map 2 (GROUP, 3)
#### A masked pattern was here ####
Vertices:
- Map 2
+ Map 3
Map Operator Tree:
TableScan
- alias: tab
+ alias: b
Statistics: Num rows: 242 Data size: 2566 Basic stats: COMPLETE Column stats: NONE
Filter Operator
predicate: key is not null (type: boolean)
Statistics: Num rows: 121 Data size: 1283 Basic stats: COMPLETE Column stats: NONE
- Select Operator
- expressions: key (type: int)
- outputColumnNames: key
- Statistics: Num rows: 121 Data size: 1283 Basic stats: COMPLETE Column stats: NONE
- Group By Operator
- keys: key (type: int)
- mode: hash
- outputColumnNames: _col0
- Statistics: Num rows: 121 Data size: 1283 Basic stats: COMPLETE Column stats: NONE
- Reduce Output Operator
- key expressions: _col0 (type: int)
- sort order: +
- Map-reduce partition columns: _col0 (type: int)
- Statistics: Num rows: 121 Data size: 1283 Basic stats: COMPLETE Column stats: NONE
- Reducer 3
+ Spark HashTable Sink Operator
+ condition expressions:
+ 0 {_col0}
+ 1
+ keys:
+ 0 _col0 (type: int)
+ 1 key (type: int)
Local Work:
Map Reduce Local Work
- Reduce Operator Tree:
- Group By Operator
- keys: KEY._col0 (type: int)
- mode: mergepartial
- outputColumnNames: _col0
- Statistics: Num rows: 60 Data size: 636 Basic stats: COMPLETE Column stats: NONE
- Select Operator
- expressions: _col0 (type: int)
- outputColumnNames: _col0
- Statistics: Num rows: 60 Data size: 636 Basic stats: COMPLETE Column stats: NONE
- Spark HashTable Sink Operator
- condition expressions:
- 0
- 1 {key}
- keys:
- 0 _col0 (type: int)
- 1 key (type: int)
Stage: Stage-1
Spark
+ Edges:
+ Reducer 2 <- Map 1 (GROUP, 3)
#### A masked pattern was here ####
Vertices:
Map 1
Map Operator Tree:
TableScan
- alias: b
+ alias: tab
Statistics: Num rows: 242 Data size: 2566 Basic stats: COMPLETE Column stats: NONE
Filter Operator
predicate: key is not null (type: boolean)
Statistics: Num rows: 121 Data size: 1283 Basic stats: COMPLETE Column stats: NONE
- Map Join Operator
- condition map:
- Inner Join 0 to 1
- condition expressions:
- 0 {_col0}
- 1 {key}
- keys:
- 0 _col0 (type: int)
- 1 key (type: int)
- outputColumnNames: _col0, _col1
- input vertices:
- 0 Reducer 3
- Statistics: Num rows: 133 Data size: 1411 Basic stats: COMPLETE Column stats: NONE
- Select Operator
- expressions: _col0 (type: int), _col1 (type: int)
- outputColumnNames: _col0, _col1
- Statistics: Num rows: 133 Data size: 1411 Basic stats: COMPLETE Column stats: NONE
- File Output Operator
- compressed: false
- Statistics: Num rows: 133 Data size: 1411 Basic stats: COMPLETE Column stats: NONE
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ Group By Operator
+ keys: key (type: int)
+ mode: hash
+ outputColumnNames: _col0
+ Statistics: Num rows: 121 Data size: 1283 Basic stats: COMPLETE Column stats: NONE
+ Reduce Output Operator
+ key expressions: _col0 (type: int)
+ sort order: +
+ Map-reduce partition columns: _col0 (type: int)
+ Statistics: Num rows: 121 Data size: 1283 Basic stats: COMPLETE Column stats: NONE
+ Reducer 2
Local Work:
Map Reduce Local Work
+ Reduce Operator Tree:
+ Group By Operator
+ keys: KEY._col0 (type: int)
+ mode: mergepartial
+ outputColumnNames: _col0
+ Statistics: Num rows: 121 Data size: 1283 Basic stats: COMPLETE Column stats: NONE
+ Map Join Operator
+ condition map:
+ Inner Join 0 to 1
+ condition expressions:
+ 0 {_col0}
+ 1 {key}
+ keys:
+ 0 _col0 (type: int)
+ 1 key (type: int)
+ outputColumnNames: _col0, _col1
+ input vertices:
+ 1 Map 3
+ Statistics: Num rows: 133 Data size: 1411 Basic stats: COMPLETE Column stats: NONE
+ Select Operator
+ expressions: _col0 (type: int), _col1 (type: int)
+ outputColumnNames: _col0, _col1
+ Statistics: Num rows: 133 Data size: 1411 Basic stats: COMPLETE Column stats: NONE
+ File Output Operator
+ compressed: false
+ Statistics: Num rows: 133 Data size: 1411 Basic stats: COMPLETE Column stats: NONE
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
Stage: Stage-0
Fetch Operator
@@ -695,91 +687,83 @@ STAGE DEPENDENCIES:
STAGE PLANS:
Stage: Stage-2
Spark
- Edges:
- Reducer 3 <- Map 2 (GROUP, 3)
#### A masked pattern was here ####
Vertices:
- Map 2
+ Map 3
Map Operator Tree:
TableScan
- alias: tab
+ alias: b
Statistics: Num rows: 242 Data size: 2566 Basic stats: COMPLETE Column stats: NONE
Filter Operator
- predicate: UDFToDouble(value) is not null (type: boolean)
+ predicate: UDFToDouble(key) is not null (type: boolean)
Statistics: Num rows: 121 Data size: 1283 Basic stats: COMPLETE Column stats: NONE
- Select Operator
- expressions: value (type: string)
- outputColumnNames: value
- Statistics: Num rows: 121 Data size: 1283 Basic stats: COMPLETE Column stats: NONE
- Group By Operator
- keys: value (type: string)
- mode: hash
- outputColumnNames: _col0
- Statistics: Num rows: 121 Data size: 1283 Basic stats: COMPLETE Column stats: NONE
- Reduce Output Operator
- key expressions: _col0 (type: string)
- sort order: +
- Map-reduce partition columns: _col0 (type: string)
- Statistics: Num rows: 121 Data size: 1283 Basic stats: COMPLETE Column stats: NONE
- Reducer 3
+ Spark HashTable Sink Operator
+ condition expressions:
+ 0 {_col0}
+ 1 {value}
+ keys:
+ 0 UDFToDouble(_col0) (type: double)
+ 1 UDFToDouble(key) (type: double)
Local Work:
Map Reduce Local Work
- Reduce Operator Tree:
- Group By Operator
- keys: KEY._col0 (type: string)
- mode: mergepartial
- outputColumnNames: _col0
- Statistics: Num rows: 60 Data size: 636 Basic stats: COMPLETE Column stats: NONE
- Select Operator
- expressions: _col0 (type: string)
- outputColumnNames: _col0
- Statistics: Num rows: 60 Data size: 636 Basic stats: COMPLETE Column stats: NONE
- Spark HashTable Sink Operator
- condition expressions:
- 0 {_col0}
- 1 {value}
- keys:
- 0 UDFToDouble(_col0) (type: double)
- 1 UDFToDouble(key) (type: double)
Stage: Stage-1
Spark
+ Edges:
+ Reducer 2 <- Map 1 (GROUP, 3)
#### A masked pattern was here ####
Vertices:
Map 1
Map Operator Tree:
TableScan
- alias: b
+ alias: tab
Statistics: Num rows: 242 Data size: 2566 Basic stats: COMPLETE Column stats: NONE
Filter Operator
- predicate: UDFToDouble(key) is not null (type: boolean)
+ predicate: UDFToDouble(value) is not null (type: boolean)
Statistics: Num rows: 121 Data size: 1283 Basic stats: COMPLETE Column stats: NONE
- Map Join Operator
- condition map:
- Inner Join 0 to 1
- condition expressions:
- 0 {_col0}
- 1 {value}
- keys:
- 0 UDFToDouble(_col0) (type: double)
- 1 UDFToDouble(key) (type: double)
- outputColumnNames: _col0, _col2
- input vertices:
- 0 Reducer 3
- Statistics: Num rows: 133 Data size: 1411 Basic stats: COMPLETE Column stats: NONE
- Select Operator
- expressions: _col0 (type: string), _col2 (type: string)
- outputColumnNames: _col0, _col1
- Statistics: Num rows: 133 Data size: 1411 Basic stats: COMPLETE Column stats: NONE
- File Output Operator
- compressed: false
- Statistics: Num rows: 133 Data size: 1411 Basic stats: COMPLETE Column stats: NONE
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ Group By Operator
+ keys: value (type: string)
+ mode: hash
+ outputColumnNames: _col0
+ Statistics: Num rows: 121 Data size: 1283 Basic stats: COMPLETE Column stats: NONE
+ Reduce Output Operator
+ key expressions: _col0 (type: string)
+ sort order: +
+ Map-reduce partition columns: _col0 (type: string)
+ Statistics: Num rows: 121 Data size: 1283 Basic stats: COMPLETE Column stats: NONE
+ Reducer 2
Local Work:
Map Reduce Local Work
+ Reduce Operator Tree:
+ Group By Operator
+ keys: KEY._col0 (type: string)
+ mode: mergepartial
+ outputColumnNames: _col0
+ Statistics: Num rows: 121 Data size: 1283 Basic stats: COMPLETE Column stats: NONE
+ Map Join Operator
+ condition map:
+ Inner Join 0 to 1
+ condition expressions:
+ 0 {_col0}
+ 1 {value}
+ keys:
+ 0 UDFToDouble(_col0) (type: double)
+ 1 UDFToDouble(key) (type: double)
+ outputColumnNames: _col0, _col2
+ input vertices:
+ 1 Map 3
+ Statistics: Num rows: 133 Data size: 1411 Basic stats: COMPLETE Column stats: NONE
+ Select Operator
+ expressions: _col0 (type: string), _col2 (type: string)
+ outputColumnNames: _col0, _col1
+ Statistics: Num rows: 133 Data size: 1411 Basic stats: COMPLETE Column stats: NONE
+ File Output Operator
+ compressed: false
+ Statistics: Num rows: 133 Data size: 1411 Basic stats: COMPLETE Column stats: NONE
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
Stage: Stage-0
Fetch Operator
Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin1.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin1.q.out?rev=1642395&r1=1642394&r2=1642395&view=diff
==============================================================================
Files hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin1.q.out (original) and hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin1.q.out Sat Nov 29 03:44:22 2014 differ
Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin10.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin10.q.out?rev=1642395&r1=1642394&r2=1642395&view=diff
==============================================================================
Files hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin10.q.out (original) and hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin10.q.out Sat Nov 29 03:44:22 2014 differ
Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin11.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin11.q.out?rev=1642395&r1=1642394&r2=1642395&view=diff
==============================================================================
Files hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin11.q.out (original) and hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin11.q.out Sat Nov 29 03:44:22 2014 differ
Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin12.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin12.q.out?rev=1642395&r1=1642394&r2=1642395&view=diff
==============================================================================
Files hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin12.q.out (original) and hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin12.q.out Sat Nov 29 03:44:22 2014 differ
Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin13.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin13.q.out?rev=1642395&r1=1642394&r2=1642395&view=diff
==============================================================================
Files hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin13.q.out (original) and hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin13.q.out Sat Nov 29 03:44:22 2014 differ
Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin2.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin2.q.out?rev=1642395&r1=1642394&r2=1642395&view=diff
==============================================================================
Files hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin2.q.out (original) and hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin2.q.out Sat Nov 29 03:44:22 2014 differ
Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin3.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin3.q.out?rev=1642395&r1=1642394&r2=1642395&view=diff
==============================================================================
Files hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin3.q.out (original) and hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin3.q.out Sat Nov 29 03:44:22 2014 differ
Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin4.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin4.q.out?rev=1642395&r1=1642394&r2=1642395&view=diff
==============================================================================
Files hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin4.q.out (original) and hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin4.q.out Sat Nov 29 03:44:22 2014 differ
Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin5.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin5.q.out?rev=1642395&r1=1642394&r2=1642395&view=diff
==============================================================================
Files hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin5.q.out (original) and hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin5.q.out Sat Nov 29 03:44:22 2014 differ
Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin7.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin7.q.out?rev=1642395&r1=1642394&r2=1642395&view=diff
==============================================================================
Files hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin7.q.out (original) and hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin7.q.out Sat Nov 29 03:44:22 2014 differ
Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin8.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin8.q.out?rev=1642395&r1=1642394&r2=1642395&view=diff
==============================================================================
Files hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin8.q.out (original) and hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin8.q.out Sat Nov 29 03:44:22 2014 differ
Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin9.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin9.q.out?rev=1642395&r1=1642394&r2=1642395&view=diff
==============================================================================
Files hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin9.q.out (original) and hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin9.q.out Sat Nov 29 03:44:22 2014 differ
Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin_negative.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin_negative.q.out?rev=1642395&r1=1642394&r2=1642395&view=diff
==============================================================================
Files hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin_negative.q.out (original) and hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin_negative.q.out Sat Nov 29 03:44:22 2014 differ
Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin_negative2.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin_negative2.q.out?rev=1642395&r1=1642394&r2=1642395&view=diff
==============================================================================
Files hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin_negative2.q.out (original) and hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin_negative2.q.out Sat Nov 29 03:44:22 2014 differ
Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin_negative3.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin_negative3.q.out?rev=1642395&r1=1642394&r2=1642395&view=diff
==============================================================================
Files hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin_negative3.q.out (original) and hive/branches/spark/ql/src/test/results/clientpositive/spark/bucketmapjoin_negative3.q.out Sat Nov 29 03:44:22 2014 differ
Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/column_access_stats.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/column_access_stats.q.out?rev=1642395&r1=1642394&r2=1642395&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/spark/column_access_stats.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/spark/column_access_stats.q.out Sat Nov 29 03:44:22 2014
@@ -381,7 +381,7 @@ STAGE PLANS:
Spark
#### A masked pattern was here ####
Vertices:
- Map 1
+ Map 2
Map Operator Tree:
TableScan
alias: t2
@@ -403,7 +403,7 @@ STAGE PLANS:
Spark
#### A masked pattern was here ####
Vertices:
- Map 2
+ Map 1
Map Operator Tree:
TableScan
alias: t1
@@ -422,7 +422,7 @@ STAGE PLANS:
1 key (type: string)
outputColumnNames: _col0
input vertices:
- 1 Map 1
+ 1 Map 2
Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
Select Operator
expressions: _col0 (type: string)
@@ -512,7 +512,7 @@ STAGE PLANS:
Spark
#### A masked pattern was here ####
Vertices:
- Map 1
+ Map 2
Map Operator Tree:
TableScan
alias: t2
@@ -534,7 +534,7 @@ STAGE PLANS:
Spark
#### A masked pattern was here ####
Vertices:
- Map 2
+ Map 1
Map Operator Tree:
TableScan
alias: t1
@@ -553,7 +553,7 @@ STAGE PLANS:
1 key (type: string)
outputColumnNames: _col0, _col5
input vertices:
- 1 Map 1
+ 1 Map 2
Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
Select Operator
expressions: _col0 (type: string), '3' (type: string), _col5 (type: string), '3' (type: string)
@@ -610,7 +610,7 @@ STAGE PLANS:
Spark
#### A masked pattern was here ####
Vertices:
- Map 1
+ Map 2
Map Operator Tree:
TableScan
alias: t2
@@ -636,7 +636,7 @@ STAGE PLANS:
Spark
#### A masked pattern was here ####
Vertices:
- Map 2
+ Map 1
Map Operator Tree:
TableScan
alias: t1
@@ -659,7 +659,7 @@ STAGE PLANS:
1 _col0 (type: string)
outputColumnNames: _col0
input vertices:
- 1 Map 1
+ 1 Map 2
Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
Select Operator
expressions: _col0 (type: string)
@@ -730,23 +730,6 @@ STAGE PLANS:
Spark
#### A masked pattern was here ####
Vertices:
- Map 1
- Map Operator Tree:
- TableScan
- alias: t3
- Statistics: Num rows: 0 Data size: 35 Basic stats: PARTIAL Column stats: NONE
- Filter Operator
- predicate: key is not null (type: boolean)
- Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
- Spark HashTable Sink Operator
- condition expressions:
- 0 {_col0}
- 1 {val}
- keys:
- 0 _col0 (type: string)
- 1 key (type: string)
- Local Work:
- Map Reduce Local Work
Map 2
Map Operator Tree:
TableScan
@@ -768,12 +751,29 @@ STAGE PLANS:
1 _col0 (type: string)
Local Work:
Map Reduce Local Work
+ Map 3
+ Map Operator Tree:
+ TableScan
+ alias: t3
+ Statistics: Num rows: 0 Data size: 35 Basic stats: PARTIAL Column stats: NONE
+ Filter Operator
+ predicate: key is not null (type: boolean)
+ Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
+ Spark HashTable Sink Operator
+ condition expressions:
+ 0 {_col0}
+ 1 {val}
+ keys:
+ 0 _col0 (type: string)
+ 1 key (type: string)
+ Local Work:
+ Map Reduce Local Work
Stage: Stage-1
Spark
#### A masked pattern was here ####
Vertices:
- Map 3
+ Map 1
Map Operator Tree:
TableScan
alias: t1
@@ -798,34 +798,30 @@ STAGE PLANS:
input vertices:
1 Map 2
Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
- Select Operator
- expressions: _col0 (type: string)
- outputColumnNames: _col0
+ Map Join Operator
+ condition map:
+ Inner Join 0 to 1
+ condition expressions:
+ 0 {_col0}
+ 1 {key} {val}
+ keys:
+ 0 _col0 (type: string)
+ 1 key (type: string)
+ outputColumnNames: _col0, _col1, _col2
+ input vertices:
+ 1 Map 3
Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
- Map Join Operator
- condition map:
- Inner Join 0 to 1
- condition expressions:
- 0 {_col0}
- 1 {key} {val}
- keys:
- 0 _col0 (type: string)
- 1 key (type: string)
+ Select Operator
+ expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string)
outputColumnNames: _col0, _col1, _col2
- input vertices:
- 1 Map 1
Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
- Select Operator
- expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string)
- outputColumnNames: _col0, _col1, _col2
+ File Output Operator
+ compressed: false
Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
- File Output Operator
- compressed: false
- Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
Local Work:
Map Reduce Local Work
Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/cross_join.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/cross_join.q.out?rev=1642395&r1=1642394&r2=1642395&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/spark/cross_join.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/spark/cross_join.q.out Sat Nov 29 03:44:22 2014
@@ -14,7 +14,7 @@ STAGE PLANS:
Spark
#### A masked pattern was here ####
Vertices:
- Map 1
+ Map 2
Map Operator Tree:
TableScan
alias: src2
@@ -33,7 +33,7 @@ STAGE PLANS:
Spark
#### A masked pattern was here ####
Vertices:
- Map 2
+ Map 1
Map Operator Tree:
TableScan
alias: src
@@ -49,7 +49,7 @@ STAGE PLANS:
1
outputColumnNames: _col0
input vertices:
- 1 Map 1
+ 1 Map 2
Statistics: Num rows: 550 Data size: 5843 Basic stats: COMPLETE Column stats: NONE
Select Operator
expressions: _col0 (type: string)
@@ -87,7 +87,7 @@ STAGE PLANS:
Spark
#### A masked pattern was here ####
Vertices:
- Map 1
+ Map 2
Map Operator Tree:
TableScan
alias: src2
@@ -106,7 +106,7 @@ STAGE PLANS:
Spark
#### A masked pattern was here ####
Vertices:
- Map 2
+ Map 1
Map Operator Tree:
TableScan
alias: src
@@ -122,7 +122,7 @@ STAGE PLANS:
1
outputColumnNames: _col0
input vertices:
- 1 Map 1
+ 1 Map 2
Statistics: Num rows: 550 Data size: 5843 Basic stats: COMPLETE Column stats: NONE
Select Operator
expressions: _col0 (type: string)
@@ -160,7 +160,7 @@ STAGE PLANS:
Spark
#### A masked pattern was here ####
Vertices:
- Map 1
+ Map 2
Map Operator Tree:
TableScan
alias: src2
@@ -182,7 +182,7 @@ STAGE PLANS:
Spark
#### A masked pattern was here ####
Vertices:
- Map 2
+ Map 1
Map Operator Tree:
TableScan
alias: src
@@ -201,7 +201,7 @@ STAGE PLANS:
1 key (type: string)
outputColumnNames: _col0
input vertices:
- 1 Map 1
+ 1 Map 2
Statistics: Num rows: 275 Data size: 2921 Basic stats: COMPLETE Column stats: NONE
Select Operator
expressions: _col0 (type: string)
Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/ctas.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/ctas.q.out?rev=1642395&r1=1642394&r2=1642395&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/spark/ctas.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/spark/ctas.q.out Sat Nov 29 03:44:22 2014
@@ -1,10 +1,10 @@
-PREHOOK: query: -- EXCLUDE_HADOOP_MAJOR_VERSIONS(0.20, 0.20S)
+PREHOOK: query: -- EXCLUDE_HADOOP_MAJOR_VERSIONS( 0.20S)
create table nzhang_Tmp(a int, b string)
PREHOOK: type: CREATETABLE
PREHOOK: Output: database:default
PREHOOK: Output: default@nzhang_Tmp
-POSTHOOK: query: -- EXCLUDE_HADOOP_MAJOR_VERSIONS(0.20, 0.20S)
+POSTHOOK: query: -- EXCLUDE_HADOOP_MAJOR_VERSIONS( 0.20S)
create table nzhang_Tmp(a int, b string)
POSTHOOK: type: CREATETABLE
Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/groupby4.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/groupby4.q.out?rev=1642395&r1=1642394&r2=1642395&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/spark/groupby4.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/spark/groupby4.q.out Sat Nov 29 03:44:22 2014
@@ -64,18 +64,14 @@ STAGE PLANS:
mode: final
outputColumnNames: _col0
Statistics: Num rows: 250 Data size: 2656 Basic stats: COMPLETE Column stats: NONE
- Select Operator
- expressions: _col0 (type: string)
- outputColumnNames: _col0
+ File Output Operator
+ compressed: false
Statistics: Num rows: 250 Data size: 2656 Basic stats: COMPLETE Column stats: NONE
- File Output Operator
- compressed: false
- Statistics: Num rows: 250 Data size: 2656 Basic stats: COMPLETE Column stats: NONE
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- name: default.dest1
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ name: default.dest1
Stage: Stage-0
Move Operator
Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/groupby7_noskew_multi_single_reducer.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/groupby7_noskew_multi_single_reducer.q.out?rev=1642395&r1=1642394&r2=1642395&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/spark/groupby7_noskew_multi_single_reducer.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/spark/groupby7_noskew_multi_single_reducer.q.out Sat Nov 29 03:44:22 2014
@@ -112,17 +112,13 @@ STAGE PLANS:
mode: complete
outputColumnNames: _col0, _col1
Statistics: Num rows: 250 Data size: 2656 Basic stats: COMPLETE Column stats: NONE
- Select Operator
- expressions: _col0 (type: string), _col1 (type: double)
- outputColumnNames: _col0, _col1
- Statistics: Num rows: 250 Data size: 2656 Basic stats: COMPLETE Column stats: NONE
- Limit
- Number of rows: 10
+ Limit
+ Number of rows: 10
+ Statistics: Num rows: 10 Data size: 100 Basic stats: COMPLETE Column stats: NONE
+ Reduce Output Operator
+ sort order:
Statistics: Num rows: 10 Data size: 100 Basic stats: COMPLETE Column stats: NONE
- Reduce Output Operator
- sort order:
- Statistics: Num rows: 10 Data size: 100 Basic stats: COMPLETE Column stats: NONE
- value expressions: _col0 (type: string), _col1 (type: double)
+ value expressions: _col0 (type: string), _col1 (type: double)
Reducer 6
Reduce Operator Tree:
Forward
@@ -133,17 +129,13 @@ STAGE PLANS:
mode: complete
outputColumnNames: _col0, _col1
Statistics: Num rows: 250 Data size: 2656 Basic stats: COMPLETE Column stats: NONE
- Select Operator
- expressions: _col0 (type: string), _col1 (type: double)
- outputColumnNames: _col0, _col1
- Statistics: Num rows: 250 Data size: 2656 Basic stats: COMPLETE Column stats: NONE
- Limit
- Number of rows: 10
+ Limit
+ Number of rows: 10
+ Statistics: Num rows: 10 Data size: 100 Basic stats: COMPLETE Column stats: NONE
+ Reduce Output Operator
+ sort order:
Statistics: Num rows: 10 Data size: 100 Basic stats: COMPLETE Column stats: NONE
- Reduce Output Operator
- sort order:
- Statistics: Num rows: 10 Data size: 100 Basic stats: COMPLETE Column stats: NONE
- value expressions: _col0 (type: string), _col1 (type: double)
+ value expressions: _col0 (type: string), _col1 (type: double)
Stage: Stage-0
Move Operator
Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/groupby_complex_types.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/groupby_complex_types.q.out?rev=1642395&r1=1642394&r2=1642395&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/spark/groupby_complex_types.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/spark/groupby_complex_types.q.out Sat Nov 29 03:44:22 2014
@@ -127,18 +127,14 @@ STAGE PLANS:
mode: mergepartial
outputColumnNames: _col0, _col1
Statistics: Num rows: 250 Data size: 2656 Basic stats: COMPLETE Column stats: NONE
- Select Operator
- expressions: _col0 (type: array<string>), _col1 (type: bigint)
- outputColumnNames: _col0, _col1
+ File Output Operator
+ compressed: false
Statistics: Num rows: 250 Data size: 2656 Basic stats: COMPLETE Column stats: NONE
- File Output Operator
- compressed: false
- Statistics: Num rows: 250 Data size: 2656 Basic stats: COMPLETE Column stats: NONE
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- name: default.dest1
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ name: default.dest1
Reducer 3
Reduce Operator Tree:
Group By Operator
@@ -147,18 +143,14 @@ STAGE PLANS:
mode: mergepartial
outputColumnNames: _col0, _col1
Statistics: Num rows: 250 Data size: 2656 Basic stats: COMPLETE Column stats: NONE
- Select Operator
- expressions: _col0 (type: map<string,string>), _col1 (type: bigint)
- outputColumnNames: _col0, _col1
+ File Output Operator
+ compressed: false
Statistics: Num rows: 250 Data size: 2656 Basic stats: COMPLETE Column stats: NONE
- File Output Operator
- compressed: false
- Statistics: Num rows: 250 Data size: 2656 Basic stats: COMPLETE Column stats: NONE
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- name: default.dest2
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ name: default.dest2
Reducer 4
Reduce Operator Tree:
Group By Operator
@@ -167,18 +159,14 @@ STAGE PLANS:
mode: mergepartial
outputColumnNames: _col0, _col1
Statistics: Num rows: 250 Data size: 2656 Basic stats: COMPLETE Column stats: NONE
- Select Operator
- expressions: _col0 (type: struct<col1:string,col2:string>), _col1 (type: bigint)
- outputColumnNames: _col0, _col1
+ File Output Operator
+ compressed: false
Statistics: Num rows: 250 Data size: 2656 Basic stats: COMPLETE Column stats: NONE
- File Output Operator
- compressed: false
- Statistics: Num rows: 250 Data size: 2656 Basic stats: COMPLETE Column stats: NONE
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- name: default.dest3
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ name: default.dest3
Stage: Stage-0
Move Operator
Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/groupby_complex_types_multi_single_reducer.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/groupby_complex_types_multi_single_reducer.q.out?rev=1642395&r1=1642394&r2=1642395&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/spark/groupby_complex_types_multi_single_reducer.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/spark/groupby_complex_types_multi_single_reducer.q.out Sat Nov 29 03:44:22 2014
@@ -95,17 +95,13 @@ STAGE PLANS:
mode: mergepartial
outputColumnNames: _col0, _col1
Statistics: Num rows: 250 Data size: 2656 Basic stats: COMPLETE Column stats: NONE
- Select Operator
- expressions: _col0 (type: array<string>), _col1 (type: bigint)
- outputColumnNames: _col0, _col1
- Statistics: Num rows: 250 Data size: 2656 Basic stats: COMPLETE Column stats: NONE
- Limit
- Number of rows: 10
+ Limit
+ Number of rows: 10
+ Statistics: Num rows: 10 Data size: 100 Basic stats: COMPLETE Column stats: NONE
+ Reduce Output Operator
+ sort order:
Statistics: Num rows: 10 Data size: 100 Basic stats: COMPLETE Column stats: NONE
- Reduce Output Operator
- sort order:
- Statistics: Num rows: 10 Data size: 100 Basic stats: COMPLETE Column stats: NONE
- value expressions: _col0 (type: array<string>), _col1 (type: bigint)
+ value expressions: _col0 (type: array<string>), _col1 (type: bigint)
Reducer 3
Reduce Operator Tree:
Select Operator
@@ -131,17 +127,13 @@ STAGE PLANS:
mode: mergepartial
outputColumnNames: _col0, _col1
Statistics: Num rows: 250 Data size: 2656 Basic stats: COMPLETE Column stats: NONE
- Select Operator
- expressions: _col0 (type: map<string,string>), _col1 (type: bigint)
- outputColumnNames: _col0, _col1
- Statistics: Num rows: 250 Data size: 2656 Basic stats: COMPLETE Column stats: NONE
- Limit
- Number of rows: 10
+ Limit
+ Number of rows: 10
+ Statistics: Num rows: 10 Data size: 100 Basic stats: COMPLETE Column stats: NONE
+ Reduce Output Operator
+ sort order:
Statistics: Num rows: 10 Data size: 100 Basic stats: COMPLETE Column stats: NONE
- Reduce Output Operator
- sort order:
- Statistics: Num rows: 10 Data size: 100 Basic stats: COMPLETE Column stats: NONE
- value expressions: _col0 (type: map<string,string>), _col1 (type: bigint)
+ value expressions: _col0 (type: map<string,string>), _col1 (type: bigint)
Reducer 5
Reduce Operator Tree:
Select Operator
Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/groupby_multi_single_reducer2.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/groupby_multi_single_reducer2.q.out?rev=1642395&r1=1642394&r2=1642395&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/spark/groupby_multi_single_reducer2.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/spark/groupby_multi_single_reducer2.q.out Sat Nov 29 03:44:22 2014
@@ -46,16 +46,12 @@ STAGE PLANS:
Filter Operator
predicate: ((substr(key, 1, 1) >= 5) or (substr(key, 1, 1) < 5)) (type: boolean)
Statistics: Num rows: 332 Data size: 3527 Basic stats: COMPLETE Column stats: NONE
- Select Operator
- expressions: key (type: string), value (type: string)
- outputColumnNames: key, value
+ Reduce Output Operator
+ key expressions: substr(key, 1, 1) (type: string), key (type: string)
+ sort order: ++
+ Map-reduce partition columns: substr(key, 1, 1) (type: string)
Statistics: Num rows: 332 Data size: 3527 Basic stats: COMPLETE Column stats: NONE
- Reduce Output Operator
- key expressions: substr(key, 1, 1) (type: string), key (type: string)
- sort order: ++
- Map-reduce partition columns: substr(key, 1, 1) (type: string)
- Statistics: Num rows: 332 Data size: 3527 Basic stats: COMPLETE Column stats: NONE
- value expressions: value (type: string)
+ value expressions: value (type: string)
Reducer 2
Reduce Operator Tree:
Forward