You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hive.apache.org by br...@apache.org on 2014/12/29 03:14:00 UTC
svn commit: r1648280 [24/25] - in /hive/branches/spark: data/conf/spark/
ql/src/java/org/apache/hadoop/hive/ql/exec/spark/session/
ql/src/java/org/apache/hadoop/hive/ql/optimizer/optiq/
ql/src/test/results/clientpositive/spark/
Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/subquery_in.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/subquery_in.q.out?rev=1648280&r1=1648279&r2=1648280&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/spark/subquery_in.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/spark/subquery_in.q.out Mon Dec 29 02:13:57 2014
@@ -15,16 +15,31 @@ from src
where src.key in (select key from src s1 where s1.key > '9')
POSTHOOK: type: QUERY
STAGE DEPENDENCIES:
- Stage-2 is a root stage
- Stage-1 depends on stages: Stage-2
+ Stage-1 is a root stage
Stage-0 depends on stages: Stage-1
STAGE PLANS:
- Stage: Stage-2
+ Stage: Stage-1
Spark
+ Edges:
+ Reducer 2 <- Map 1 (PARTITION-LEVEL SORT, 1), Map 3 (PARTITION-LEVEL SORT, 1)
#### A masked pattern was here ####
Vertices:
- Map 2
+ Map 1
+ Map Operator Tree:
+ TableScan
+ alias: src
+ Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
+ Filter Operator
+ predicate: key is not null (type: boolean)
+ Statistics: Num rows: 250 Data size: 2656 Basic stats: COMPLETE Column stats: NONE
+ Reduce Output Operator
+ key expressions: key (type: string)
+ sort order: +
+ Map-reduce partition columns: key (type: string)
+ Statistics: Num rows: 250 Data size: 2656 Basic stats: COMPLETE Column stats: NONE
+ value expressions: value (type: string)
+ Map 3
Map Operator Tree:
TableScan
alias: s1
@@ -41,54 +56,32 @@ STAGE PLANS:
mode: hash
outputColumnNames: _col0
Statistics: Num rows: 166 Data size: 1763 Basic stats: COMPLETE Column stats: NONE
- Spark HashTable Sink Operator
- condition expressions:
- 0 {key} {value}
- 1
- keys:
- 0 key (type: string)
- 1 _col0 (type: string)
- Local Work:
- Map Reduce Local Work
-
- Stage: Stage-1
- Spark
-#### A masked pattern was here ####
- Vertices:
- Map 1
- Map Operator Tree:
- TableScan
- alias: src
- Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
- Filter Operator
- predicate: key is not null (type: boolean)
- Statistics: Num rows: 250 Data size: 2656 Basic stats: COMPLETE Column stats: NONE
- Map Join Operator
- condition map:
- Left Semi Join 0 to 1
- condition expressions:
- 0 {key} {value}
- 1
- keys:
- 0 key (type: string)
- 1 _col0 (type: string)
- outputColumnNames: _col0, _col1
- input vertices:
- 1 Map 2
- Statistics: Num rows: 275 Data size: 2921 Basic stats: COMPLETE Column stats: NONE
- Select Operator
- expressions: _col0 (type: string), _col1 (type: string)
- outputColumnNames: _col0, _col1
- Statistics: Num rows: 275 Data size: 2921 Basic stats: COMPLETE Column stats: NONE
- File Output Operator
- compressed: false
- Statistics: Num rows: 275 Data size: 2921 Basic stats: COMPLETE Column stats: NONE
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- Local Work:
- Map Reduce Local Work
+ Reduce Output Operator
+ key expressions: _col0 (type: string)
+ sort order: +
+ Map-reduce partition columns: _col0 (type: string)
+ Statistics: Num rows: 166 Data size: 1763 Basic stats: COMPLETE Column stats: NONE
+ Reducer 2
+ Reduce Operator Tree:
+ Join Operator
+ condition map:
+ Left Semi Join 0 to 1
+ keys:
+ 0 key (type: string)
+ 1 _col0 (type: string)
+ outputColumnNames: _col0, _col1
+ Statistics: Num rows: 275 Data size: 2921 Basic stats: COMPLETE Column stats: NONE
+ Select Operator
+ expressions: _col0 (type: string), _col1 (type: string)
+ outputColumnNames: _col0, _col1
+ Statistics: Num rows: 275 Data size: 2921 Basic stats: COMPLETE Column stats: NONE
+ File Output Operator
+ compressed: false
+ Statistics: Num rows: 275 Data size: 2921 Basic stats: COMPLETE Column stats: NONE
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
Stage: Stage-0
Fetch Operator
@@ -140,16 +133,30 @@ where b.key in
)
POSTHOOK: type: QUERY
STAGE DEPENDENCIES:
- Stage-2 is a root stage
- Stage-1 depends on stages: Stage-2
+ Stage-1 is a root stage
Stage-0 depends on stages: Stage-1
STAGE PLANS:
- Stage: Stage-2
+ Stage: Stage-1
Spark
+ Edges:
+ Reducer 2 <- Map 1 (PARTITION-LEVEL SORT, 3), Map 3 (PARTITION-LEVEL SORT, 3)
#### A masked pattern was here ####
Vertices:
- Map 2
+ Map 1
+ Map Operator Tree:
+ TableScan
+ alias: b
+ Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
+ Filter Operator
+ predicate: (key is not null and value is not null) (type: boolean)
+ Statistics: Num rows: 125 Data size: 1328 Basic stats: COMPLETE Column stats: NONE
+ Reduce Output Operator
+ key expressions: key (type: string), value (type: string)
+ sort order: ++
+ Map-reduce partition columns: key (type: string), value (type: string)
+ Statistics: Num rows: 125 Data size: 1328 Basic stats: COMPLETE Column stats: NONE
+ Map 3
Map Operator Tree:
TableScan
alias: a
@@ -166,54 +173,32 @@ STAGE PLANS:
mode: hash
outputColumnNames: _col0, _col1
Statistics: Num rows: 83 Data size: 881 Basic stats: COMPLETE Column stats: NONE
- Spark HashTable Sink Operator
- condition expressions:
- 0 {key} {value}
- 1
- keys:
- 0 key (type: string), value (type: string)
- 1 _col0 (type: string), _col1 (type: string)
- Local Work:
- Map Reduce Local Work
-
- Stage: Stage-1
- Spark
-#### A masked pattern was here ####
- Vertices:
- Map 1
- Map Operator Tree:
- TableScan
- alias: b
- Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
- Filter Operator
- predicate: (key is not null and value is not null) (type: boolean)
- Statistics: Num rows: 125 Data size: 1328 Basic stats: COMPLETE Column stats: NONE
- Map Join Operator
- condition map:
- Left Semi Join 0 to 1
- condition expressions:
- 0 {key} {value}
- 1
- keys:
- 0 key (type: string), value (type: string)
- 1 _col0 (type: string), _col1 (type: string)
- outputColumnNames: _col0, _col1
- input vertices:
- 1 Map 2
- Statistics: Num rows: 137 Data size: 1460 Basic stats: COMPLETE Column stats: NONE
- Select Operator
- expressions: _col0 (type: string), _col1 (type: string)
- outputColumnNames: _col0, _col1
- Statistics: Num rows: 137 Data size: 1460 Basic stats: COMPLETE Column stats: NONE
- File Output Operator
- compressed: false
- Statistics: Num rows: 137 Data size: 1460 Basic stats: COMPLETE Column stats: NONE
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- Local Work:
- Map Reduce Local Work
+ Reduce Output Operator
+ key expressions: _col0 (type: string), _col1 (type: string)
+ sort order: ++
+ Map-reduce partition columns: _col0 (type: string), _col1 (type: string)
+ Statistics: Num rows: 83 Data size: 881 Basic stats: COMPLETE Column stats: NONE
+ Reducer 2
+ Reduce Operator Tree:
+ Join Operator
+ condition map:
+ Left Semi Join 0 to 1
+ keys:
+ 0 key (type: string), value (type: string)
+ 1 _col0 (type: string), _col1 (type: string)
+ outputColumnNames: _col0, _col1
+ Statistics: Num rows: 137 Data size: 1460 Basic stats: COMPLETE Column stats: NONE
+ Select Operator
+ expressions: _col0 (type: string), _col1 (type: string)
+ outputColumnNames: _col0, _col1
+ Statistics: Num rows: 137 Data size: 1460 Basic stats: COMPLETE Column stats: NONE
+ File Output Operator
+ compressed: false
+ Statistics: Num rows: 137 Data size: 1460 Basic stats: COMPLETE Column stats: NONE
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
Stage: Stage-0
Fetch Operator
@@ -273,19 +258,33 @@ part where part.p_size in
)
POSTHOOK: type: QUERY
STAGE DEPENDENCIES:
- Stage-2 is a root stage
- Stage-1 depends on stages: Stage-2
+ Stage-1 is a root stage
Stage-0 depends on stages: Stage-1
STAGE PLANS:
- Stage: Stage-2
+ Stage: Stage-1
Spark
Edges:
- Reducer 3 <- Map 2 (PARTITION-LEVEL SORT, 3)
- Reducer 4 <- Reducer 3 (GROUP, 1)
+ Reducer 4 <- Map 3 (PARTITION-LEVEL SORT, 3)
+ Reducer 5 <- Reducer 4 (GROUP, 1)
+ Reducer 2 <- Map 1 (PARTITION-LEVEL SORT, 3), Reducer 5 (PARTITION-LEVEL SORT, 3)
#### A masked pattern was here ####
Vertices:
- Map 2
+ Map 1
+ Map Operator Tree:
+ TableScan
+ alias: part
+ Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
+ Filter Operator
+ predicate: UDFToDouble(p_size) is not null (type: boolean)
+ Statistics: Num rows: 13 Data size: 1573 Basic stats: COMPLETE Column stats: NONE
+ Reduce Output Operator
+ key expressions: UDFToDouble(p_size) (type: double)
+ sort order: +
+ Map-reduce partition columns: UDFToDouble(p_size) (type: double)
+ Statistics: Num rows: 13 Data size: 1573 Basic stats: COMPLETE Column stats: NONE
+ value expressions: p_name (type: string), p_size (type: int)
+ Map 3
Map Operator Tree:
TableScan
alias: part
@@ -296,7 +295,28 @@ STAGE PLANS:
Map-reduce partition columns: p_mfgr (type: string)
Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
value expressions: p_mfgr (type: string), p_size (type: int)
- Reducer 3
+ Reducer 2
+ Reduce Operator Tree:
+ Join Operator
+ condition map:
+ Left Semi Join 0 to 1
+ keys:
+ 0 UDFToDouble(p_size) (type: double)
+ 1 _col0 (type: double)
+ outputColumnNames: _col1, _col5
+ Statistics: Num rows: 14 Data size: 1730 Basic stats: COMPLETE Column stats: NONE
+ Select Operator
+ expressions: _col1 (type: string), _col5 (type: int)
+ outputColumnNames: _col0, _col1
+ Statistics: Num rows: 14 Data size: 1730 Basic stats: COMPLETE Column stats: NONE
+ File Output Operator
+ compressed: false
+ Statistics: Num rows: 14 Data size: 1730 Basic stats: COMPLETE Column stats: NONE
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ Reducer 4
Reduce Operator Tree:
Extract
Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
@@ -318,9 +338,7 @@ STAGE PLANS:
sort order:
Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL Column stats: NONE
value expressions: _col0 (type: struct<count:bigint,sum:double,input:int>)
- Reducer 4
- Local Work:
- Map Reduce Local Work
+ Reducer 5
Reduce Operator Tree:
Group By Operator
aggregations: avg(VALUE._col0)
@@ -335,52 +353,11 @@ STAGE PLANS:
mode: hash
outputColumnNames: _col0
Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE
- Spark HashTable Sink Operator
- condition expressions:
- 0 {p_name} {p_size}
- 1
- keys:
- 0 UDFToDouble(p_size) (type: double)
- 1 _col0 (type: double)
-
- Stage: Stage-1
- Spark
-#### A masked pattern was here ####
- Vertices:
- Map 1
- Map Operator Tree:
- TableScan
- alias: part
- Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
- Filter Operator
- predicate: UDFToDouble(p_size) is not null (type: boolean)
- Statistics: Num rows: 13 Data size: 1573 Basic stats: COMPLETE Column stats: NONE
- Map Join Operator
- condition map:
- Left Semi Join 0 to 1
- condition expressions:
- 0 {p_name} {p_size}
- 1
- keys:
- 0 UDFToDouble(p_size) (type: double)
- 1 _col0 (type: double)
- outputColumnNames: _col1, _col5
- input vertices:
- 1 Reducer 4
- Statistics: Num rows: 14 Data size: 1730 Basic stats: COMPLETE Column stats: NONE
- Select Operator
- expressions: _col1 (type: string), _col5 (type: int)
- outputColumnNames: _col0, _col1
- Statistics: Num rows: 14 Data size: 1730 Basic stats: COMPLETE Column stats: NONE
- File Output Operator
- compressed: false
- Statistics: Num rows: 14 Data size: 1730 Basic stats: COMPLETE Column stats: NONE
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- Local Work:
- Map Reduce Local Work
+ Reduce Output Operator
+ key expressions: _col0 (type: double)
+ sort order: +
+ Map-reduce partition columns: _col0 (type: double)
+ Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE
Stage: Stage-0
Fetch Operator
@@ -429,19 +406,33 @@ from part b where b.p_size in
)
POSTHOOK: type: QUERY
STAGE DEPENDENCIES:
- Stage-2 is a root stage
- Stage-1 depends on stages: Stage-2
+ Stage-1 is a root stage
Stage-0 depends on stages: Stage-1
STAGE PLANS:
- Stage: Stage-2
+ Stage: Stage-1
Spark
Edges:
- Reducer 3 <- Map 2 (PARTITION-LEVEL SORT, 3)
- Reducer 4 <- Reducer 3 (GROUP, 3)
+ Reducer 4 <- Map 3 (PARTITION-LEVEL SORT, 3)
+ Reducer 5 <- Reducer 4 (GROUP, 3)
+ Reducer 2 <- Map 1 (PARTITION-LEVEL SORT, 3), Reducer 5 (PARTITION-LEVEL SORT, 3)
#### A masked pattern was here ####
Vertices:
- Map 2
+ Map 1
+ Map Operator Tree:
+ TableScan
+ alias: b
+ Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
+ Filter Operator
+ predicate: (p_size is not null and p_mfgr is not null) (type: boolean)
+ Statistics: Num rows: 7 Data size: 847 Basic stats: COMPLETE Column stats: NONE
+ Reduce Output Operator
+ key expressions: p_size (type: int), p_mfgr (type: string)
+ sort order: ++
+ Map-reduce partition columns: p_size (type: int), p_mfgr (type: string)
+ Statistics: Num rows: 7 Data size: 847 Basic stats: COMPLETE Column stats: NONE
+ value expressions: p_name (type: string)
+ Map 3
Map Operator Tree:
TableScan
alias: part
@@ -452,7 +443,28 @@ STAGE PLANS:
Map-reduce partition columns: p_mfgr (type: string)
Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
value expressions: p_mfgr (type: string), p_size (type: int)
- Reducer 3
+ Reducer 2
+ Reduce Operator Tree:
+ Join Operator
+ condition map:
+ Left Semi Join 0 to 1
+ keys:
+ 0 p_size (type: int), p_mfgr (type: string)
+ 1 _col0 (type: int), _col1 (type: string)
+ outputColumnNames: _col1, _col2, _col5
+ Statistics: Num rows: 7 Data size: 931 Basic stats: COMPLETE Column stats: NONE
+ Select Operator
+ expressions: _col2 (type: string), _col1 (type: string), _col5 (type: int)
+ outputColumnNames: _col0, _col1, _col2
+ Statistics: Num rows: 7 Data size: 931 Basic stats: COMPLETE Column stats: NONE
+ File Output Operator
+ compressed: false
+ Statistics: Num rows: 7 Data size: 931 Basic stats: COMPLETE Column stats: NONE
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ Reducer 4
Reduce Operator Tree:
Extract
Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
@@ -477,9 +489,7 @@ STAGE PLANS:
Map-reduce partition columns: _col0 (type: string)
Statistics: Num rows: 4 Data size: 484 Basic stats: COMPLETE Column stats: NONE
value expressions: _col1 (type: int)
- Reducer 4
- Local Work:
- Map Reduce Local Work
+ Reducer 5
Reduce Operator Tree:
Group By Operator
aggregations: min(VALUE._col0)
@@ -499,52 +509,11 @@ STAGE PLANS:
mode: hash
outputColumnNames: _col0, _col1
Statistics: Num rows: 2 Data size: 242 Basic stats: COMPLETE Column stats: NONE
- Spark HashTable Sink Operator
- condition expressions:
- 0 {p_name} {p_mfgr} {p_size}
- 1
- keys:
- 0 p_size (type: int), p_mfgr (type: string)
- 1 _col0 (type: int), _col1 (type: string)
-
- Stage: Stage-1
- Spark
-#### A masked pattern was here ####
- Vertices:
- Map 1
- Map Operator Tree:
- TableScan
- alias: b
- Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
- Filter Operator
- predicate: (p_size is not null and p_mfgr is not null) (type: boolean)
- Statistics: Num rows: 7 Data size: 847 Basic stats: COMPLETE Column stats: NONE
- Map Join Operator
- condition map:
- Left Semi Join 0 to 1
- condition expressions:
- 0 {p_name} {p_mfgr} {p_size}
- 1
- keys:
- 0 p_size (type: int), p_mfgr (type: string)
- 1 _col0 (type: int), _col1 (type: string)
- outputColumnNames: _col1, _col2, _col5
- input vertices:
- 1 Reducer 4
- Statistics: Num rows: 7 Data size: 931 Basic stats: COMPLETE Column stats: NONE
- Select Operator
- expressions: _col2 (type: string), _col1 (type: string), _col5 (type: int)
- outputColumnNames: _col0, _col1, _col2
- Statistics: Num rows: 7 Data size: 931 Basic stats: COMPLETE Column stats: NONE
- File Output Operator
- compressed: false
- Statistics: Num rows: 7 Data size: 931 Basic stats: COMPLETE Column stats: NONE
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- Local Work:
- Map Reduce Local Work
+ Reduce Output Operator
+ key expressions: _col0 (type: int), _col1 (type: string)
+ sort order: ++
+ Map-reduce partition columns: _col0 (type: int), _col1 (type: string)
+ Statistics: Num rows: 2 Data size: 242 Basic stats: COMPLETE Column stats: NONE
Stage: Stage-0
Fetch Operator
@@ -597,18 +566,31 @@ where b.key in
)
POSTHOOK: type: QUERY
STAGE DEPENDENCIES:
- Stage-2 is a root stage
- Stage-1 depends on stages: Stage-2
+ Stage-1 is a root stage
Stage-0 depends on stages: Stage-1
STAGE PLANS:
- Stage: Stage-2
+ Stage: Stage-1
Spark
Edges:
- Reducer 3 <- Map 2 (GROUP, 3)
+ Reducer 4 <- Map 3 (GROUP, 3)
+ Reducer 2 <- Map 1 (PARTITION-LEVEL SORT, 3), Reducer 4 (PARTITION-LEVEL SORT, 3)
#### A masked pattern was here ####
Vertices:
- Map 2
+ Map 1
+ Map Operator Tree:
+ TableScan
+ alias: b
+ Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
+ Filter Operator
+ predicate: (key is not null and value is not null) (type: boolean)
+ Statistics: Num rows: 125 Data size: 1328 Basic stats: COMPLETE Column stats: NONE
+ Reduce Output Operator
+ key expressions: key (type: string), value (type: string)
+ sort order: ++
+ Map-reduce partition columns: key (type: string), value (type: string)
+ Statistics: Num rows: 125 Data size: 1328 Basic stats: COMPLETE Column stats: NONE
+ Map 3
Map Operator Tree:
TableScan
alias: a
@@ -626,9 +608,28 @@ STAGE PLANS:
sort order: ++
Map-reduce partition columns: _col0 (type: string), _col1 (type: string)
Statistics: Num rows: 83 Data size: 881 Basic stats: COMPLETE Column stats: NONE
- Reducer 3
- Local Work:
- Map Reduce Local Work
+ Reducer 2
+ Reduce Operator Tree:
+ Join Operator
+ condition map:
+ Left Semi Join 0 to 1
+ keys:
+ 0 key (type: string), value (type: string)
+ 1 _col0 (type: string), _col1 (type: string)
+ outputColumnNames: _col0, _col1
+ Statistics: Num rows: 137 Data size: 1460 Basic stats: COMPLETE Column stats: NONE
+ Select Operator
+ expressions: _col0 (type: string), _col1 (type: string)
+ outputColumnNames: _col0, _col1
+ Statistics: Num rows: 137 Data size: 1460 Basic stats: COMPLETE Column stats: NONE
+ File Output Operator
+ compressed: false
+ Statistics: Num rows: 137 Data size: 1460 Basic stats: COMPLETE Column stats: NONE
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ Reducer 4
Reduce Operator Tree:
Group By Operator
keys: KEY._col0 (type: string), KEY._col1 (type: string)
@@ -640,52 +641,11 @@ STAGE PLANS:
mode: hash
outputColumnNames: _col0, _col1
Statistics: Num rows: 41 Data size: 435 Basic stats: COMPLETE Column stats: NONE
- Spark HashTable Sink Operator
- condition expressions:
- 0 {key} {value}
- 1
- keys:
- 0 key (type: string), value (type: string)
- 1 _col0 (type: string), _col1 (type: string)
-
- Stage: Stage-1
- Spark
-#### A masked pattern was here ####
- Vertices:
- Map 1
- Map Operator Tree:
- TableScan
- alias: b
- Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
- Filter Operator
- predicate: (key is not null and value is not null) (type: boolean)
- Statistics: Num rows: 125 Data size: 1328 Basic stats: COMPLETE Column stats: NONE
- Map Join Operator
- condition map:
- Left Semi Join 0 to 1
- condition expressions:
- 0 {key} {value}
- 1
- keys:
- 0 key (type: string), value (type: string)
- 1 _col0 (type: string), _col1 (type: string)
- outputColumnNames: _col0, _col1
- input vertices:
- 1 Reducer 3
- Statistics: Num rows: 137 Data size: 1460 Basic stats: COMPLETE Column stats: NONE
- Select Operator
- expressions: _col0 (type: string), _col1 (type: string)
- outputColumnNames: _col0, _col1
- Statistics: Num rows: 137 Data size: 1460 Basic stats: COMPLETE Column stats: NONE
- File Output Operator
- compressed: false
- Statistics: Num rows: 137 Data size: 1460 Basic stats: COMPLETE Column stats: NONE
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- Local Work:
- Map Reduce Local Work
+ Reduce Output Operator
+ key expressions: _col0 (type: string), _col1 (type: string)
+ sort order: ++
+ Map-reduce partition columns: _col0 (type: string), _col1 (type: string)
+ Statistics: Num rows: 41 Data size: 435 Basic stats: COMPLETE Column stats: NONE
Stage: Stage-0
Fetch Operator
@@ -761,16 +721,37 @@ where li.l_linenumber = 1 and
li.l_orderkey in (select l_orderkey from lineitem where l_shipmode = 'AIR')
POSTHOOK: type: QUERY
STAGE DEPENDENCIES:
- Stage-2 is a root stage
- Stage-1 depends on stages: Stage-2
+ Stage-1 is a root stage
Stage-0 depends on stages: Stage-1
STAGE PLANS:
- Stage: Stage-2
+ Stage: Stage-1
Spark
+ Edges:
+ Reducer 2 <- Map 1 (GROUP, 3)
+ Reducer 3 <- Map 5 (PARTITION-LEVEL SORT, 3), Reducer 2 (PARTITION-LEVEL SORT, 3)
+ Reducer 4 <- Map 6 (PARTITION-LEVEL SORT, 3), Reducer 3 (PARTITION-LEVEL SORT, 3)
#### A masked pattern was here ####
Vertices:
- Map 3
+ Map 1
+ Map Operator Tree:
+ TableScan
+ alias: lineitem
+ Statistics: Num rows: 100 Data size: 11999 Basic stats: COMPLETE Column stats: NONE
+ Filter Operator
+ predicate: l_partkey is not null (type: boolean)
+ Statistics: Num rows: 50 Data size: 5999 Basic stats: COMPLETE Column stats: NONE
+ Group By Operator
+ keys: l_partkey (type: int)
+ mode: hash
+ outputColumnNames: _col0
+ Statistics: Num rows: 50 Data size: 5999 Basic stats: COMPLETE Column stats: NONE
+ Reduce Output Operator
+ key expressions: _col0 (type: int)
+ sort order: +
+ Map-reduce partition columns: _col0 (type: int)
+ Statistics: Num rows: 50 Data size: 5999 Basic stats: COMPLETE Column stats: NONE
+ Map 5
Map Operator Tree:
TableScan
alias: li
@@ -778,16 +759,13 @@ STAGE PLANS:
Filter Operator
predicate: ((l_partkey is not null and l_orderkey is not null) and (l_linenumber = 1)) (type: boolean)
Statistics: Num rows: 12 Data size: 1439 Basic stats: COMPLETE Column stats: NONE
- Spark HashTable Sink Operator
- condition expressions:
- 0 {_col0}
- 1 {l_orderkey} {l_suppkey}
- keys:
- 0 _col0 (type: int)
- 1 l_partkey (type: int)
- Local Work:
- Map Reduce Local Work
- Map 4
+ Reduce Output Operator
+ key expressions: l_partkey (type: int)
+ sort order: +
+ Map-reduce partition columns: l_partkey (type: int)
+ Statistics: Num rows: 12 Data size: 1439 Basic stats: COMPLETE Column stats: NONE
+ value expressions: l_orderkey (type: int), l_suppkey (type: int)
+ Map 6
Map Operator Tree:
TableScan
alias: lineitem
@@ -804,86 +782,60 @@ STAGE PLANS:
mode: hash
outputColumnNames: _col0
Statistics: Num rows: 25 Data size: 2999 Basic stats: COMPLETE Column stats: NONE
- Spark HashTable Sink Operator
- condition expressions:
- 0 {_col0} {_col3}
- 1
- keys:
- 0 _col1 (type: int)
- 1 _col0 (type: int)
- Local Work:
- Map Reduce Local Work
-
- Stage: Stage-1
- Spark
- Edges:
- Reducer 2 <- Map 1 (GROUP, 3)
-#### A masked pattern was here ####
- Vertices:
- Map 1
- Map Operator Tree:
- TableScan
- alias: lineitem
- Statistics: Num rows: 100 Data size: 11999 Basic stats: COMPLETE Column stats: NONE
- Filter Operator
- predicate: l_partkey is not null (type: boolean)
- Statistics: Num rows: 50 Data size: 5999 Basic stats: COMPLETE Column stats: NONE
- Group By Operator
- keys: l_partkey (type: int)
- mode: hash
- outputColumnNames: _col0
- Statistics: Num rows: 50 Data size: 5999 Basic stats: COMPLETE Column stats: NONE
- Reduce Output Operator
- key expressions: _col0 (type: int)
- sort order: +
- Map-reduce partition columns: _col0 (type: int)
- Statistics: Num rows: 50 Data size: 5999 Basic stats: COMPLETE Column stats: NONE
+ Reduce Output Operator
+ key expressions: _col0 (type: int)
+ sort order: +
+ Map-reduce partition columns: _col0 (type: int)
+ Statistics: Num rows: 25 Data size: 2999 Basic stats: COMPLETE Column stats: NONE
Reducer 2
- Local Work:
- Map Reduce Local Work
Reduce Operator Tree:
Group By Operator
keys: KEY._col0 (type: int)
mode: mergepartial
outputColumnNames: _col0
- Statistics: Num rows: 50 Data size: 5999 Basic stats: COMPLETE Column stats: NONE
- Map Join Operator
- condition map:
- Inner Join 0 to 1
- condition expressions:
- 0 {_col0}
- 1 {l_orderkey} {l_suppkey}
- keys:
- 0 _col0 (type: int)
- 1 l_partkey (type: int)
- outputColumnNames: _col0, _col1, _col3
- input vertices:
- 1 Map 3
- Statistics: Num rows: 55 Data size: 6598 Basic stats: COMPLETE Column stats: NONE
- Map Join Operator
- condition map:
- Left Semi Join 0 to 1
- condition expressions:
- 0 {_col0} {_col3}
- 1
- keys:
- 0 _col1 (type: int)
- 1 _col0 (type: int)
- outputColumnNames: _col0, _col3
- input vertices:
- 1 Map 4
- Statistics: Num rows: 60 Data size: 7257 Basic stats: COMPLETE Column stats: NONE
- Select Operator
- expressions: _col0 (type: int), _col3 (type: int)
- outputColumnNames: _col0, _col1
- Statistics: Num rows: 60 Data size: 7257 Basic stats: COMPLETE Column stats: NONE
- File Output Operator
- compressed: false
- Statistics: Num rows: 60 Data size: 7257 Basic stats: COMPLETE Column stats: NONE
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ Statistics: Num rows: 25 Data size: 2999 Basic stats: COMPLETE Column stats: NONE
+ Reduce Output Operator
+ key expressions: _col0 (type: int)
+ sort order: +
+ Map-reduce partition columns: _col0 (type: int)
+ Statistics: Num rows: 25 Data size: 2999 Basic stats: COMPLETE Column stats: NONE
+ Reducer 3
+ Reduce Operator Tree:
+ Join Operator
+ condition map:
+ Inner Join 0 to 1
+ keys:
+ 0 _col0 (type: int)
+ 1 l_partkey (type: int)
+ outputColumnNames: _col0, _col1, _col3
+ Statistics: Num rows: 27 Data size: 3298 Basic stats: COMPLETE Column stats: NONE
+ Reduce Output Operator
+ key expressions: _col1 (type: int)
+ sort order: +
+ Map-reduce partition columns: _col1 (type: int)
+ Statistics: Num rows: 27 Data size: 3298 Basic stats: COMPLETE Column stats: NONE
+ value expressions: _col0 (type: int), _col3 (type: int)
+ Reducer 4
+ Reduce Operator Tree:
+ Join Operator
+ condition map:
+ Left Semi Join 0 to 1
+ keys:
+ 0 _col1 (type: int)
+ 1 _col0 (type: int)
+ outputColumnNames: _col0, _col3
+ Statistics: Num rows: 29 Data size: 3627 Basic stats: COMPLETE Column stats: NONE
+ Select Operator
+ expressions: _col0 (type: int), _col3 (type: int)
+ outputColumnNames: _col0, _col1
+ Statistics: Num rows: 29 Data size: 3627 Basic stats: COMPLETE Column stats: NONE
+ File Output Operator
+ compressed: false
+ Statistics: Num rows: 29 Data size: 3627 Basic stats: COMPLETE Column stats: NONE
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
Stage: Stage-0
Fetch Operator
@@ -931,5 +883,3 @@ where li.l_linenumber = 1 and
POSTHOOK: type: QUERY
POSTHOOK: Input: default@lineitem
#### A masked pattern was here ####
-108570 8571
-4297 1798
Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/subquery_multiinsert.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/subquery_multiinsert.q.out?rev=1648280&r1=1648279&r2=1648280&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/spark/subquery_multiinsert.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/spark/subquery_multiinsert.q.out Mon Dec 29 02:13:57 2014
@@ -161,8 +161,8 @@ STAGE PLANS:
Join Operator
condition map:
Left Semi Join 0 to 1
- condition expressions:
- 0 {VALUE._col0} {VALUE._col1}
+ keys:
+ 0
1
outputColumnNames: _col0, _col1
Statistics: Num rows: 550 Data size: 5843 Basic stats: COMPLETE Column stats: NONE
@@ -177,9 +177,9 @@ STAGE PLANS:
Join Operator
condition map:
Left Outer Join0 to 1
- condition expressions:
- 0 {KEY.reducesinkkey0} {VALUE._col0}
- 1 {KEY.reducesinkkey0}
+ keys:
+ 0 _col0 (type: string)
+ 1 _col0 (type: string)
outputColumnNames: _col0, _col1, _col5
Statistics: Num rows: 605 Data size: 6427 Basic stats: COMPLETE Column stats: NONE
Filter Operator
@@ -213,9 +213,9 @@ STAGE PLANS:
Join Operator
condition map:
Left Semi Join 0 to 1
- condition expressions:
- 0 {KEY.reducesinkkey0} {KEY.reducesinkkey1}
- 1
+ keys:
+ 0 key (type: string), value (type: string)
+ 1 _col0 (type: string), _col1 (type: string)
outputColumnNames: _col0, _col1
Statistics: Num rows: 550 Data size: 5843 Basic stats: COMPLETE Column stats: NONE
File Output Operator
@@ -525,9 +525,6 @@ STAGE PLANS:
outputColumnNames: _col0, _col1
Statistics: Num rows: 83 Data size: 881 Basic stats: COMPLETE Column stats: NONE
Spark HashTable Sink Operator
- condition expressions:
- 0 {key} {value}
- 1
keys:
0 key (type: string), value (type: string)
1 _col0 (type: string), _col1 (type: string)
@@ -546,9 +543,6 @@ STAGE PLANS:
outputColumnNames: _col0
Statistics: Num rows: 166 Data size: 1763 Basic stats: COMPLETE Column stats: NONE
Spark HashTable Sink Operator
- condition expressions:
- 0 {_col0} {_col1}
- 1
keys:
0 _col0 (type: string)
1 _col0 (type: string)
@@ -595,9 +589,6 @@ STAGE PLANS:
outputColumnNames: _col0
Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE
Spark HashTable Sink Operator
- condition expressions:
- 0 {key} {value}
- 1
keys:
0
1
@@ -616,9 +607,6 @@ STAGE PLANS:
Map Join Operator
condition map:
Left Semi Join 0 to 1
- condition expressions:
- 0 {key} {value}
- 1
keys:
0
1
@@ -629,9 +617,6 @@ STAGE PLANS:
Map Join Operator
condition map:
Left Outer Join0 to 1
- condition expressions:
- 0 {_col0} {_col1}
- 1 {_col0}
keys:
0 _col0 (type: string)
1 _col0 (type: string)
@@ -654,9 +639,6 @@ STAGE PLANS:
Map Join Operator
condition map:
Left Semi Join 0 to 1
- condition expressions:
- 0 {key} {value}
- 1
keys:
0 key (type: string), value (type: string)
1 _col0 (type: string), _col1 (type: string)
Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/temp_table_join1.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/temp_table_join1.q.out?rev=1648280&r1=1648279&r2=1648280&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/spark/temp_table_join1.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/spark/temp_table_join1.q.out Mon Dec 29 02:13:57 2014
@@ -75,9 +75,9 @@ STAGE PLANS:
Join Operator
condition map:
Inner Join 0 to 1
- condition expressions:
- 0 {KEY.reducesinkkey0}
- 1 {VALUE._col0}
+ keys:
+ 0 key (type: string)
+ 1 key (type: string)
outputColumnNames: _col0, _col6
Statistics: Num rows: 5 Data size: 57 Basic stats: COMPLETE Column stats: NONE
Select Operator
@@ -171,9 +171,9 @@ STAGE PLANS:
Join Operator
condition map:
Inner Join 0 to 1
- condition expressions:
- 0 {KEY.reducesinkkey0}
- 1 {VALUE._col0}
+ keys:
+ 0 key (type: string)
+ 1 key (type: string)
outputColumnNames: _col0, _col6
Statistics: Num rows: 5 Data size: 57 Basic stats: COMPLETE Column stats: NONE
Select Operator
@@ -269,9 +269,9 @@ STAGE PLANS:
Join Operator
condition map:
Inner Join 0 to 1
- condition expressions:
- 0 {KEY.reducesinkkey0}
- 1 {VALUE._col0}
+ keys:
+ 0 key (type: string)
+ 1 key (type: string)
outputColumnNames: _col0, _col6
Statistics: Num rows: 5 Data size: 57 Basic stats: COMPLETE Column stats: NONE
Select Operator
Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/tez_join_tests.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/tez_join_tests.q.out?rev=1648280&r1=1648279&r2=1648280&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/spark/tez_join_tests.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/spark/tez_join_tests.q.out Mon Dec 29 02:13:57 2014
@@ -25,54 +25,62 @@ STAGE PLANS:
Map 1
Map Operator Tree:
TableScan
- alias: a
- Statistics: Num rows: 25 Data size: 191 Basic stats: COMPLETE Column stats: NONE
- Reduce Output Operator
- key expressions: key (type: string)
- sort order: +
- Map-reduce partition columns: key (type: string)
- Statistics: Num rows: 25 Data size: 191 Basic stats: COMPLETE Column stats: NONE
- Map 6
- Map Operator Tree:
- TableScan
alias: b
Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
- Reduce Output Operator
- key expressions: key (type: string)
- sort order: +
- Map-reduce partition columns: key (type: string)
+ Select Operator
+ expressions: key (type: string), value (type: string)
+ outputColumnNames: _col0, _col1
Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
- value expressions: value (type: string)
+ Reduce Output Operator
+ key expressions: _col0 (type: string)
+ sort order: +
+ Map-reduce partition columns: _col0 (type: string)
+ Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
+ value expressions: _col1 (type: string)
+ Map 6
+ Map Operator Tree:
+ TableScan
+ alias: a
+ Statistics: Num rows: 25 Data size: 191 Basic stats: COMPLETE Column stats: NONE
+ Select Operator
+ expressions: key (type: string)
+ outputColumnNames: _col0
+ Statistics: Num rows: 25 Data size: 191 Basic stats: COMPLETE Column stats: NONE
+ Reduce Output Operator
+ key expressions: _col0 (type: string)
+ sort order: +
+ Map-reduce partition columns: _col0 (type: string)
+ Statistics: Num rows: 25 Data size: 191 Basic stats: COMPLETE Column stats: NONE
Map 7
Map Operator Tree:
TableScan
- alias: c
+ alias: b
Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
- Reduce Output Operator
- key expressions: value (type: string)
- sort order: +
- Map-reduce partition columns: value (type: string)
+ Select Operator
+ expressions: key (type: string), value (type: string)
+ outputColumnNames: _col0, _col1
Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
- value expressions: key (type: string)
+ Reduce Output Operator
+ key expressions: _col1 (type: string)
+ sort order: +
+ Map-reduce partition columns: _col1 (type: string)
+ Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
+ value expressions: _col0 (type: string)
Reducer 2
Reduce Operator Tree:
Join Operator
condition map:
- Left Outer Join0 to 1
- condition expressions:
- 0
- 1 {KEY.reducesinkkey0} {VALUE._col0}
- outputColumnNames: _col5, _col6
+ Right Outer Join0 to 1
+ keys:
+ 0 _col0 (type: string)
+ 1 _col0 (type: string)
+ outputColumnNames: _col0, _col1
Statistics: Num rows: 550 Data size: 5843 Basic stats: COMPLETE Column stats: NONE
- Select Operator
- expressions: _col5 (type: string), _col6 (type: string)
- outputColumnNames: _col0, _col1
+ Reduce Output Operator
+ key expressions: _col0 (type: string)
+ sort order: +
Statistics: Num rows: 550 Data size: 5843 Basic stats: COMPLETE Column stats: NONE
- Reduce Output Operator
- key expressions: _col0 (type: string)
- sort order: +
- Statistics: Num rows: 550 Data size: 5843 Basic stats: COMPLETE Column stats: NONE
- value expressions: _col1 (type: string)
+ value expressions: _col1 (type: string)
Reducer 3
Reduce Operator Tree:
Select Operator
@@ -90,9 +98,9 @@ STAGE PLANS:
Join Operator
condition map:
Right Outer Join0 to 1
- condition expressions:
- 0 {VALUE._col0} {KEY.reducesinkkey0}
- 1 {VALUE._col0} {KEY.reducesinkkey0}
+ keys:
+ 0 _col1 (type: string)
+ 1 _col1 (type: string)
outputColumnNames: _col0, _col1, _col2, _col3
Statistics: Num rows: 605 Data size: 6427 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/tez_joins_explain.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/tez_joins_explain.q.out?rev=1648280&r1=1648279&r2=1648280&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/spark/tez_joins_explain.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/spark/tez_joins_explain.q.out Mon Dec 29 02:13:57 2014
@@ -25,54 +25,62 @@ STAGE PLANS:
Map 1
Map Operator Tree:
TableScan
- alias: a
- Statistics: Num rows: 25 Data size: 191 Basic stats: COMPLETE Column stats: NONE
- Reduce Output Operator
- key expressions: key (type: string)
- sort order: +
- Map-reduce partition columns: key (type: string)
- Statistics: Num rows: 25 Data size: 191 Basic stats: COMPLETE Column stats: NONE
- Map 6
- Map Operator Tree:
- TableScan
alias: b
Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
- Reduce Output Operator
- key expressions: key (type: string)
- sort order: +
- Map-reduce partition columns: key (type: string)
+ Select Operator
+ expressions: key (type: string), value (type: string)
+ outputColumnNames: _col0, _col1
Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
- value expressions: value (type: string)
+ Reduce Output Operator
+ key expressions: _col0 (type: string)
+ sort order: +
+ Map-reduce partition columns: _col0 (type: string)
+ Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
+ value expressions: _col1 (type: string)
+ Map 6
+ Map Operator Tree:
+ TableScan
+ alias: a
+ Statistics: Num rows: 25 Data size: 191 Basic stats: COMPLETE Column stats: NONE
+ Select Operator
+ expressions: key (type: string)
+ outputColumnNames: _col0
+ Statistics: Num rows: 25 Data size: 191 Basic stats: COMPLETE Column stats: NONE
+ Reduce Output Operator
+ key expressions: _col0 (type: string)
+ sort order: +
+ Map-reduce partition columns: _col0 (type: string)
+ Statistics: Num rows: 25 Data size: 191 Basic stats: COMPLETE Column stats: NONE
Map 7
Map Operator Tree:
TableScan
- alias: c
+ alias: b
Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
- Reduce Output Operator
- key expressions: value (type: string)
- sort order: +
- Map-reduce partition columns: value (type: string)
+ Select Operator
+ expressions: key (type: string), value (type: string)
+ outputColumnNames: _col0, _col1
Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
- value expressions: key (type: string)
+ Reduce Output Operator
+ key expressions: _col1 (type: string)
+ sort order: +
+ Map-reduce partition columns: _col1 (type: string)
+ Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
+ value expressions: _col0 (type: string)
Reducer 2
Reduce Operator Tree:
Join Operator
condition map:
- Left Outer Join0 to 1
- condition expressions:
- 0
- 1 {KEY.reducesinkkey0} {VALUE._col0}
- outputColumnNames: _col5, _col6
+ Right Outer Join0 to 1
+ keys:
+ 0 _col0 (type: string)
+ 1 _col0 (type: string)
+ outputColumnNames: _col0, _col1
Statistics: Num rows: 550 Data size: 5843 Basic stats: COMPLETE Column stats: NONE
- Select Operator
- expressions: _col5 (type: string), _col6 (type: string)
- outputColumnNames: _col0, _col1
+ Reduce Output Operator
+ key expressions: _col0 (type: string)
+ sort order: +
Statistics: Num rows: 550 Data size: 5843 Basic stats: COMPLETE Column stats: NONE
- Reduce Output Operator
- key expressions: _col0 (type: string)
- sort order: +
- Statistics: Num rows: 550 Data size: 5843 Basic stats: COMPLETE Column stats: NONE
- value expressions: _col1 (type: string)
+ value expressions: _col1 (type: string)
Reducer 3
Reduce Operator Tree:
Select Operator
@@ -90,9 +98,9 @@ STAGE PLANS:
Join Operator
condition map:
Right Outer Join0 to 1
- condition expressions:
- 0 {VALUE._col0} {KEY.reducesinkkey0}
- 1 {VALUE._col0} {KEY.reducesinkkey0}
+ keys:
+ 0 _col1 (type: string)
+ 1 _col1 (type: string)
outputColumnNames: _col0, _col1, _col2, _col3
Statistics: Num rows: 605 Data size: 6427 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/transform1.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/transform1.q.out?rev=1648280&r1=1648279&r2=1648280&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/spark/transform1.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/spark/transform1.q.out Mon Dec 29 02:13:57 2014
@@ -53,10 +53,12 @@ STAGE PLANS:
PREHOOK: query: SELECT transform(*) USING 'cat' AS (col array<bigint>) FROM transform1_t1
PREHOOK: type: QUERY
+PREHOOK: Input: cat
PREHOOK: Input: default@transform1_t1
#### A masked pattern was here ####
POSTHOOK: query: SELECT transform(*) USING 'cat' AS (col array<bigint>) FROM transform1_t1
POSTHOOK: type: QUERY
+POSTHOOK: Input: cat
POSTHOOK: Input: default@transform1_t1
#### A masked pattern was here ####
PREHOOK: query: create table transform1_t2(col array<int>)
@@ -125,10 +127,12 @@ STAGE PLANS:
PREHOOK: query: SELECT transform('0\0021\0022') USING 'cat' AS (col array<int>) FROM transform1_t2
PREHOOK: type: QUERY
+PREHOOK: Input: cat
PREHOOK: Input: default@transform1_t2
#### A masked pattern was here ####
POSTHOOK: query: SELECT transform('0\0021\0022') USING 'cat' AS (col array<int>) FROM transform1_t2
POSTHOOK: type: QUERY
+POSTHOOK: Input: cat
POSTHOOK: Input: default@transform1_t2
#### A masked pattern was here ####
[0,1,2]
Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/transform_ppr1.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/transform_ppr1.q.out?rev=1648280&r1=1648279&r2=1648280&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/spark/transform_ppr1.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/spark/transform_ppr1.q.out Mon Dec 29 02:13:57 2014
@@ -152,7 +152,7 @@ STAGE PLANS:
COLUMN_STATS_ACCURATE true
bucket_count -1
columns key,value
- columns.comments default default
+ columns.comments 'default','default'
columns.types string:string
#### A masked pattern was here ####
name default.srcpart
@@ -173,7 +173,7 @@ STAGE PLANS:
properties:
bucket_count -1
columns key,value
- columns.comments default default
+ columns.comments 'default','default'
columns.types string:string
#### A masked pattern was here ####
name default.srcpart
@@ -198,7 +198,7 @@ STAGE PLANS:
COLUMN_STATS_ACCURATE true
bucket_count -1
columns key,value
- columns.comments default default
+ columns.comments 'default','default'
columns.types string:string
#### A masked pattern was here ####
name default.srcpart
@@ -219,7 +219,7 @@ STAGE PLANS:
properties:
bucket_count -1
columns key,value
- columns.comments default default
+ columns.comments 'default','default'
columns.types string:string
#### A masked pattern was here ####
name default.srcpart
@@ -244,7 +244,7 @@ STAGE PLANS:
COLUMN_STATS_ACCURATE true
bucket_count -1
columns key,value
- columns.comments default default
+ columns.comments 'default','default'
columns.types string:string
#### A masked pattern was here ####
name default.srcpart
@@ -265,7 +265,7 @@ STAGE PLANS:
properties:
bucket_count -1
columns key,value
- columns.comments default default
+ columns.comments 'default','default'
columns.types string:string
#### A masked pattern was here ####
name default.srcpart
@@ -290,7 +290,7 @@ STAGE PLANS:
COLUMN_STATS_ACCURATE true
bucket_count -1
columns key,value
- columns.comments default default
+ columns.comments 'default','default'
columns.types string:string
#### A masked pattern was here ####
name default.srcpart
@@ -311,7 +311,7 @@ STAGE PLANS:
properties:
bucket_count -1
columns key,value
- columns.comments default default
+ columns.comments 'default','default'
columns.types string:string
#### A masked pattern was here ####
name default.srcpart
@@ -372,6 +372,7 @@ PREHOOK: query: FROM (
) tmap
SELECT tmap.tkey, tmap.tvalue WHERE tmap.tkey < 100 AND tmap.ds = '2008-04-08'
PREHOOK: type: QUERY
+PREHOOK: Input: cat
PREHOOK: Input: default@srcpart
PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11
PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12
@@ -386,6 +387,7 @@ POSTHOOK: query: FROM (
) tmap
SELECT tmap.tkey, tmap.tvalue WHERE tmap.tkey < 100 AND tmap.ds = '2008-04-08'
POSTHOOK: type: QUERY
+POSTHOOK: Input: cat
POSTHOOK: Input: default@srcpart
POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11
POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12
Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/transform_ppr2.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/transform_ppr2.q.out?rev=1648280&r1=1648279&r2=1648280&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/spark/transform_ppr2.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/spark/transform_ppr2.q.out Mon Dec 29 02:13:57 2014
@@ -154,7 +154,7 @@ STAGE PLANS:
COLUMN_STATS_ACCURATE true
bucket_count -1
columns key,value
- columns.comments default default
+ columns.comments 'default','default'
columns.types string:string
#### A masked pattern was here ####
name default.srcpart
@@ -175,7 +175,7 @@ STAGE PLANS:
properties:
bucket_count -1
columns key,value
- columns.comments default default
+ columns.comments 'default','default'
columns.types string:string
#### A masked pattern was here ####
name default.srcpart
@@ -200,7 +200,7 @@ STAGE PLANS:
COLUMN_STATS_ACCURATE true
bucket_count -1
columns key,value
- columns.comments default default
+ columns.comments 'default','default'
columns.types string:string
#### A masked pattern was here ####
name default.srcpart
@@ -221,7 +221,7 @@ STAGE PLANS:
properties:
bucket_count -1
columns key,value
- columns.comments default default
+ columns.comments 'default','default'
columns.types string:string
#### A masked pattern was here ####
name default.srcpart
@@ -281,6 +281,7 @@ PREHOOK: query: FROM (
) tmap
SELECT tmap.tkey, tmap.tvalue WHERE tmap.tkey < 100
PREHOOK: type: QUERY
+PREHOOK: Input: cat
PREHOOK: Input: default@srcpart
PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11
PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12
@@ -294,6 +295,7 @@ POSTHOOK: query: FROM (
) tmap
SELECT tmap.tkey, tmap.tvalue WHERE tmap.tkey < 100
POSTHOOK: type: QUERY
+POSTHOOK: Input: cat
POSTHOOK: Input: default@srcpart
POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11
POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12
Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/union_ppr.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/union_ppr.q.out?rev=1648280&r1=1648279&r2=1648280&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/spark/union_ppr.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/spark/union_ppr.q.out Mon Dec 29 02:13:57 2014
@@ -149,7 +149,7 @@ STAGE PLANS:
COLUMN_STATS_ACCURATE true
bucket_count -1
columns key,value
- columns.comments default default
+ columns.comments 'default','default'
columns.types string:string
#### A masked pattern was here ####
name default.srcpart
@@ -170,7 +170,7 @@ STAGE PLANS:
properties:
bucket_count -1
columns key,value
- columns.comments default default
+ columns.comments 'default','default'
columns.types string:string
#### A masked pattern was here ####
name default.srcpart
@@ -195,7 +195,7 @@ STAGE PLANS:
COLUMN_STATS_ACCURATE true
bucket_count -1
columns key,value
- columns.comments default default
+ columns.comments 'default','default'
columns.types string:string
#### A masked pattern was here ####
name default.srcpart
@@ -216,7 +216,7 @@ STAGE PLANS:
properties:
bucket_count -1
columns key,value
- columns.comments default default
+ columns.comments 'default','default'
columns.types string:string
#### A masked pattern was here ####
name default.srcpart
@@ -266,7 +266,7 @@ STAGE PLANS:
COLUMN_STATS_ACCURATE true
bucket_count -1
columns key,value
- columns.comments default default
+ columns.comments 'default','default'
columns.types string:string
#### A masked pattern was here ####
name default.srcpart
@@ -287,7 +287,7 @@ STAGE PLANS:
properties:
bucket_count -1
columns key,value
- columns.comments default default
+ columns.comments 'default','default'
columns.types string:string
#### A masked pattern was here ####
name default.srcpart
@@ -312,7 +312,7 @@ STAGE PLANS:
COLUMN_STATS_ACCURATE true
bucket_count -1
columns key,value
- columns.comments default default
+ columns.comments 'default','default'
columns.types string:string
#### A masked pattern was here ####
name default.srcpart
@@ -333,7 +333,7 @@ STAGE PLANS:
properties:
bucket_count -1
columns key,value
- columns.comments default default
+ columns.comments 'default','default'
columns.types string:string
#### A masked pattern was here ####
name default.srcpart
Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/vector_cast_constant.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/vector_cast_constant.q.out?rev=1648280&r1=1648279&r2=1648280&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/spark/vector_cast_constant.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/spark/vector_cast_constant.q.out Mon Dec 29 02:13:57 2014
@@ -152,18 +152,18 @@ STAGE PLANS:
keys: KEY._col0 (type: int)
mode: mergepartial
outputColumnNames: _col0, _col1, _col2, _col3
- Statistics: Num rows: 1049 Data size: 311170 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 524 Data size: 155436 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
key expressions: _col0 (type: int)
sort order: +
- Statistics: Num rows: 1049 Data size: 311170 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 524 Data size: 155436 Basic stats: COMPLETE Column stats: NONE
value expressions: _col1 (type: double), _col2 (type: double), _col3 (type: decimal(14,4))
Reducer 3
Reduce Operator Tree:
Select Operator
expressions: KEY.reducesinkkey0 (type: int), VALUE._col0 (type: double), VALUE._col1 (type: double), VALUE._col2 (type: decimal(14,4))
outputColumnNames: _col0, _col1, _col2, _col3
- Statistics: Num rows: 1049 Data size: 311170 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 524 Data size: 155436 Basic stats: COMPLETE Column stats: NONE
Limit
Number of rows: 10
Statistics: Num rows: 10 Data size: 2960 Basic stats: COMPLETE Column stats: NONE
Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/vector_decimal_mapjoin.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/vector_decimal_mapjoin.q.out?rev=1648280&r1=1648279&r2=1648280&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/spark/vector_decimal_mapjoin.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/spark/vector_decimal_mapjoin.q.out Mon Dec 29 02:13:57 2014
@@ -49,9 +49,6 @@ STAGE PLANS:
predicate: (cint = 6981) (type: boolean)
Statistics: Num rows: 6144 Data size: 1082530 Basic stats: COMPLETE Column stats: NONE
Spark HashTable Sink Operator
- condition expressions:
- 0 {cdecimal1}
- 1 {cdecimal2}
keys:
0 6981 (type: int)
1 6981 (type: int)
@@ -73,9 +70,6 @@ STAGE PLANS:
Map Join Operator
condition map:
Inner Join 0 to 1
- condition expressions:
- 0 {cdecimal1}
- 1 {cdecimal2}
keys:
0 6981 (type: int)
1 6981 (type: int)
Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/vector_left_outer_join.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/vector_left_outer_join.q.out?rev=1648280&r1=1648279&r2=1648280&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/spark/vector_left_outer_join.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/spark/vector_left_outer_join.q.out Mon Dec 29 02:13:57 2014
@@ -29,29 +29,31 @@ STAGE PLANS:
Map 3
Map Operator Tree:
TableScan
- alias: cd
+ alias: c
Statistics: Num rows: 12288 Data size: 377237 Basic stats: COMPLETE Column stats: NONE
- Spark HashTable Sink Operator
- condition expressions:
- 0 {ctinyint}
- 1
- keys:
- 0 cint (type: int)
- 1 cint (type: int)
+ Select Operator
+ expressions: cint (type: int)
+ outputColumnNames: _col0
+ Statistics: Num rows: 12288 Data size: 377237 Basic stats: COMPLETE Column stats: NONE
+ Spark HashTable Sink Operator
+ keys:
+ 0 _col1 (type: int)
+ 1 _col0 (type: int)
Local Work:
Map Reduce Local Work
Map 4
Map Operator Tree:
TableScan
- alias: hd
+ alias: c
Statistics: Num rows: 12288 Data size: 377237 Basic stats: COMPLETE Column stats: NONE
- Spark HashTable Sink Operator
- condition expressions:
- 0
- 1
- keys:
- 0 _col0 (type: tinyint)
- 1 ctinyint (type: tinyint)
+ Select Operator
+ expressions: ctinyint (type: tinyint)
+ outputColumnNames: _col0
+ Statistics: Num rows: 12288 Data size: 377237 Basic stats: COMPLETE Column stats: NONE
+ Spark HashTable Sink Operator
+ keys:
+ 0 _col0 (type: tinyint)
+ 1 _col0 (type: tinyint)
Local Work:
Map Reduce Local Work
@@ -66,40 +68,38 @@ STAGE PLANS:
TableScan
alias: c
Statistics: Num rows: 12288 Data size: 377237 Basic stats: COMPLETE Column stats: NONE
- Map Join Operator
- condition map:
- Left Outer Join0 to 1
- condition expressions:
- 0 {ctinyint}
- 1
- keys:
- 0 cint (type: int)
- 1 cint (type: int)
- outputColumnNames: _col0
- input vertices:
- 1 Map 3
- Statistics: Num rows: 13516 Data size: 414960 Basic stats: COMPLETE Column stats: NONE
+ Select Operator
+ expressions: ctinyint (type: tinyint), cint (type: int)
+ outputColumnNames: _col0, _col1
+ Statistics: Num rows: 12288 Data size: 377237 Basic stats: COMPLETE Column stats: NONE
Map Join Operator
condition map:
Left Outer Join0 to 1
- condition expressions:
- 0
- 1
keys:
- 0 _col0 (type: tinyint)
- 1 ctinyint (type: tinyint)
+ 0 _col1 (type: int)
+ 1 _col0 (type: int)
+ outputColumnNames: _col0
input vertices:
- 1 Map 4
- Statistics: Num rows: 14867 Data size: 456456 Basic stats: COMPLETE Column stats: NONE
- Group By Operator
- aggregations: count()
- mode: hash
- outputColumnNames: _col0
- Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE
- Reduce Output Operator
- sort order:
+ 1 Map 3
+ Statistics: Num rows: 13516 Data size: 414960 Basic stats: COMPLETE Column stats: NONE
+ Map Join Operator
+ condition map:
+ Left Outer Join0 to 1
+ keys:
+ 0 _col0 (type: tinyint)
+ 1 _col0 (type: tinyint)
+ input vertices:
+ 1 Map 4
+ Statistics: Num rows: 14867 Data size: 456456 Basic stats: COMPLETE Column stats: NONE
+ Group By Operator
+ aggregations: count()
+ mode: hash
+ outputColumnNames: _col0
Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE
- value expressions: _col0 (type: bigint)
+ Reduce Output Operator
+ sort order:
+ Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE
+ value expressions: _col0 (type: bigint)
Local Work:
Map Reduce Local Work
Execution mode: vectorized