You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hive.apache.org by br...@apache.org on 2014/11/07 21:41:45 UTC
svn commit: r1637444 [15/20] - in /hive/branches/spark: ./
cli/src/test/org/apache/hadoop/hive/cli/ common/
common/src/java/org/apache/hadoop/hive/common/type/
common/src/java/org/apache/hadoop/hive/conf/
common/src/test/org/apache/hadoop/hive/conf/ co...
Modified: hive/branches/spark/ql/src/test/results/clientpositive/ql_rewrite_gbtoidx.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/ql_rewrite_gbtoidx.q.out?rev=1637444&r1=1637443&r2=1637444&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/ql_rewrite_gbtoidx.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/ql_rewrite_gbtoidx.q.out Fri Nov 7 20:41:34 2014
@@ -1,8 +1,8 @@
-PREHOOK: query: DROP TABLE lineitem
+PREHOOK: query: DROP TABLE IF EXISTS lineitem_ix
PREHOOK: type: DROPTABLE
-POSTHOOK: query: DROP TABLE lineitem
+POSTHOOK: query: DROP TABLE IF EXISTS lineitem_ix
POSTHOOK: type: DROPTABLE
-PREHOOK: query: CREATE TABLE lineitem (L_ORDERKEY INT,
+PREHOOK: query: CREATE TABLE lineitem_ix (L_ORDERKEY INT,
L_PARTKEY INT,
L_SUPPKEY INT,
L_LINENUMBER INT,
@@ -22,8 +22,8 @@ ROW FORMAT DELIMITED
FIELDS TERMINATED BY '|'
PREHOOK: type: CREATETABLE
PREHOOK: Output: database:default
-PREHOOK: Output: default@lineitem
-POSTHOOK: query: CREATE TABLE lineitem (L_ORDERKEY INT,
+PREHOOK: Output: default@lineitem_ix
+POSTHOOK: query: CREATE TABLE lineitem_ix (L_ORDERKEY INT,
L_PARTKEY INT,
L_SUPPKEY INT,
L_LINENUMBER INT,
@@ -43,40 +43,40 @@ ROW FORMAT DELIMITED
FIELDS TERMINATED BY '|'
POSTHOOK: type: CREATETABLE
POSTHOOK: Output: database:default
-POSTHOOK: Output: default@lineitem
-PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/lineitem.txt' OVERWRITE INTO TABLE lineitem
+POSTHOOK: Output: default@lineitem_ix
+PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/lineitem.txt' OVERWRITE INTO TABLE lineitem_ix
PREHOOK: type: LOAD
#### A masked pattern was here ####
-PREHOOK: Output: default@lineitem
-POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/lineitem.txt' OVERWRITE INTO TABLE lineitem
+PREHOOK: Output: default@lineitem_ix
+POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/lineitem.txt' OVERWRITE INTO TABLE lineitem_ix
POSTHOOK: type: LOAD
#### A masked pattern was here ####
-POSTHOOK: Output: default@lineitem
-PREHOOK: query: CREATE INDEX lineitem_lshipdate_idx ON TABLE lineitem(l_shipdate) AS 'org.apache.hadoop.hive.ql.index.AggregateIndexHandler' WITH DEFERRED REBUILD IDXPROPERTIES("AGGREGATES"="count(l_shipdate)")
+POSTHOOK: Output: default@lineitem_ix
+PREHOOK: query: CREATE INDEX lineitem_ix_lshipdate_idx ON TABLE lineitem_ix(l_shipdate) AS 'org.apache.hadoop.hive.ql.index.AggregateIndexHandler' WITH DEFERRED REBUILD IDXPROPERTIES("AGGREGATES"="count(l_shipdate)")
PREHOOK: type: CREATEINDEX
-PREHOOK: Input: default@lineitem
-POSTHOOK: query: CREATE INDEX lineitem_lshipdate_idx ON TABLE lineitem(l_shipdate) AS 'org.apache.hadoop.hive.ql.index.AggregateIndexHandler' WITH DEFERRED REBUILD IDXPROPERTIES("AGGREGATES"="count(l_shipdate)")
+PREHOOK: Input: default@lineitem_ix
+POSTHOOK: query: CREATE INDEX lineitem_ix_lshipdate_idx ON TABLE lineitem_ix(l_shipdate) AS 'org.apache.hadoop.hive.ql.index.AggregateIndexHandler' WITH DEFERRED REBUILD IDXPROPERTIES("AGGREGATES"="count(l_shipdate)")
POSTHOOK: type: CREATEINDEX
-POSTHOOK: Input: default@lineitem
-POSTHOOK: Output: default@default__lineitem_lineitem_lshipdate_idx__
-PREHOOK: query: ALTER INDEX lineitem_lshipdate_idx ON lineitem REBUILD
+POSTHOOK: Input: default@lineitem_ix
+POSTHOOK: Output: default@default__lineitem_ix_lineitem_ix_lshipdate_idx__
+PREHOOK: query: ALTER INDEX lineitem_ix_lshipdate_idx ON lineitem_ix REBUILD
PREHOOK: type: ALTERINDEX_REBUILD
-PREHOOK: Input: default@lineitem
-PREHOOK: Output: default@default__lineitem_lineitem_lshipdate_idx__
-POSTHOOK: query: ALTER INDEX lineitem_lshipdate_idx ON lineitem REBUILD
+PREHOOK: Input: default@lineitem_ix
+PREHOOK: Output: default@default__lineitem_ix_lineitem_ix_lshipdate_idx__
+POSTHOOK: query: ALTER INDEX lineitem_ix_lshipdate_idx ON lineitem_ix REBUILD
POSTHOOK: type: ALTERINDEX_REBUILD
-POSTHOOK: Input: default@lineitem
-POSTHOOK: Output: default@default__lineitem_lineitem_lshipdate_idx__
-POSTHOOK: Lineage: default__lineitem_lineitem_lshipdate_idx__._bucketname SIMPLE [(lineitem)lineitem.FieldSchema(name:INPUT__FILE__NAME, type:string, comment:), ]
-POSTHOOK: Lineage: default__lineitem_lineitem_lshipdate_idx__._count_of_l_shipdate EXPRESSION [(lineitem)lineitem.FieldSchema(name:l_shipdate, type:string, comment:null), ]
-POSTHOOK: Lineage: default__lineitem_lineitem_lshipdate_idx__._offsets EXPRESSION [(lineitem)lineitem.FieldSchema(name:BLOCK__OFFSET__INSIDE__FILE, type:bigint, comment:), ]
-POSTHOOK: Lineage: default__lineitem_lineitem_lshipdate_idx__.l_shipdate SIMPLE [(lineitem)lineitem.FieldSchema(name:l_shipdate, type:string, comment:null), ]
+POSTHOOK: Input: default@lineitem_ix
+POSTHOOK: Output: default@default__lineitem_ix_lineitem_ix_lshipdate_idx__
+POSTHOOK: Lineage: default__lineitem_ix_lineitem_ix_lshipdate_idx__._bucketname SIMPLE [(lineitem_ix)lineitem_ix.FieldSchema(name:INPUT__FILE__NAME, type:string, comment:), ]
+POSTHOOK: Lineage: default__lineitem_ix_lineitem_ix_lshipdate_idx__._count_of_l_shipdate EXPRESSION [(lineitem_ix)lineitem_ix.FieldSchema(name:l_shipdate, type:string, comment:null), ]
+POSTHOOK: Lineage: default__lineitem_ix_lineitem_ix_lshipdate_idx__._offsets EXPRESSION [(lineitem_ix)lineitem_ix.FieldSchema(name:BLOCK__OFFSET__INSIDE__FILE, type:bigint, comment:), ]
+POSTHOOK: Lineage: default__lineitem_ix_lineitem_ix_lshipdate_idx__.l_shipdate SIMPLE [(lineitem_ix)lineitem_ix.FieldSchema(name:l_shipdate, type:string, comment:null), ]
PREHOOK: query: explain select l_shipdate, count(l_shipdate)
-from lineitem
+from lineitem_ix
group by l_shipdate
PREHOOK: type: QUERY
POSTHOOK: query: explain select l_shipdate, count(l_shipdate)
-from lineitem
+from lineitem_ix
group by l_shipdate
POSTHOOK: type: QUERY
STAGE DEPENDENCIES:
@@ -88,7 +88,7 @@ STAGE PLANS:
Map Reduce
Map Operator Tree:
TableScan
- alias: lineitem
+ alias: lineitem_ix
Statistics: Num rows: 120 Data size: 12099 Basic stats: COMPLETE Column stats: NONE
Select Operator
expressions: l_shipdate (type: string)
@@ -132,18 +132,18 @@ STAGE PLANS:
ListSink
PREHOOK: query: select l_shipdate, count(l_shipdate)
-from lineitem
+from lineitem_ix
group by l_shipdate
order by l_shipdate
PREHOOK: type: QUERY
-PREHOOK: Input: default@lineitem
+PREHOOK: Input: default@lineitem_ix
#### A masked pattern was here ####
POSTHOOK: query: select l_shipdate, count(l_shipdate)
-from lineitem
+from lineitem_ix
group by l_shipdate
order by l_shipdate
POSTHOOK: type: QUERY
-POSTHOOK: Input: default@lineitem
+POSTHOOK: Input: default@lineitem_ix
#### A masked pattern was here ####
1992-04-27 1
1992-07-02 1
@@ -241,11 +241,11 @@ POSTHOOK: Input: default@lineitem
1998-10-23 1
1998-10-30 1
PREHOOK: query: explain select l_shipdate, count(l_shipdate)
-from lineitem
+from lineitem_ix
group by l_shipdate
PREHOOK: type: QUERY
POSTHOOK: query: explain select l_shipdate, count(l_shipdate)
-from lineitem
+from lineitem_ix
group by l_shipdate
POSTHOOK: type: QUERY
STAGE DEPENDENCIES:
@@ -257,23 +257,23 @@ STAGE PLANS:
Map Reduce
Map Operator Tree:
TableScan
- alias: default.default__lineitem_lineitem_lshipdate_idx__
- Statistics: Num rows: 95 Data size: 8675 Basic stats: COMPLETE Column stats: NONE
+ alias: default.default__lineitem_ix_lineitem_ix_lshipdate_idx__
+ Statistics: Num rows: 95 Data size: 8960 Basic stats: COMPLETE Column stats: NONE
Select Operator
expressions: l_shipdate (type: string), _count_of_l_shipdate (type: bigint)
outputColumnNames: l_shipdate, _count_of_l_shipdate
- Statistics: Num rows: 95 Data size: 8675 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 95 Data size: 8960 Basic stats: COMPLETE Column stats: NONE
Group By Operator
aggregations: sum(_count_of_l_shipdate)
keys: l_shipdate (type: string)
mode: hash
outputColumnNames: _col0, _col1
- Statistics: Num rows: 95 Data size: 8675 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 95 Data size: 8960 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
key expressions: _col0 (type: string)
sort order: +
Map-reduce partition columns: _col0 (type: string)
- Statistics: Num rows: 95 Data size: 8675 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 95 Data size: 8960 Basic stats: COMPLETE Column stats: NONE
value expressions: _col1 (type: bigint)
Reduce Operator Tree:
Group By Operator
@@ -281,14 +281,14 @@ STAGE PLANS:
keys: KEY._col0 (type: string)
mode: mergepartial
outputColumnNames: _col0, _col1
- Statistics: Num rows: 47 Data size: 4291 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 47 Data size: 4432 Basic stats: COMPLETE Column stats: NONE
Select Operator
expressions: _col0 (type: string), _col1 (type: bigint)
outputColumnNames: _col0, _col1
- Statistics: Num rows: 47 Data size: 4291 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 47 Data size: 4432 Basic stats: COMPLETE Column stats: NONE
File Output Operator
compressed: false
- Statistics: Num rows: 47 Data size: 4291 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 47 Data size: 4432 Basic stats: COMPLETE Column stats: NONE
table:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
@@ -301,20 +301,20 @@ STAGE PLANS:
ListSink
PREHOOK: query: select l_shipdate, count(l_shipdate)
-from lineitem
+from lineitem_ix
group by l_shipdate
order by l_shipdate
PREHOOK: type: QUERY
-PREHOOK: Input: default@default__lineitem_lineitem_lshipdate_idx__
-PREHOOK: Input: default@lineitem
+PREHOOK: Input: default@default__lineitem_ix_lineitem_ix_lshipdate_idx__
+PREHOOK: Input: default@lineitem_ix
#### A masked pattern was here ####
POSTHOOK: query: select l_shipdate, count(l_shipdate)
-from lineitem
+from lineitem_ix
group by l_shipdate
order by l_shipdate
POSTHOOK: type: QUERY
-POSTHOOK: Input: default@default__lineitem_lineitem_lshipdate_idx__
-POSTHOOK: Input: default@lineitem
+POSTHOOK: Input: default@default__lineitem_ix_lineitem_ix_lshipdate_idx__
+POSTHOOK: Input: default@lineitem_ix
#### A masked pattern was here ####
1992-04-27 1
1992-07-02 1
@@ -414,14 +414,14 @@ POSTHOOK: Input: default@lineitem
PREHOOK: query: explain select year(l_shipdate) as year,
month(l_shipdate) as month,
count(l_shipdate) as monthly_shipments
-from lineitem
+from lineitem_ix
group by year(l_shipdate), month(l_shipdate)
order by year, month
PREHOOK: type: QUERY
POSTHOOK: query: explain select year(l_shipdate) as year,
month(l_shipdate) as month,
count(l_shipdate) as monthly_shipments
-from lineitem
+from lineitem_ix
group by year(l_shipdate), month(l_shipdate)
order by year, month
POSTHOOK: type: QUERY
@@ -435,7 +435,7 @@ STAGE PLANS:
Map Reduce
Map Operator Tree:
TableScan
- alias: lineitem
+ alias: lineitem_ix
Statistics: Num rows: 120 Data size: 12099 Basic stats: COMPLETE Column stats: NONE
Select Operator
expressions: l_shipdate (type: string)
@@ -502,20 +502,20 @@ STAGE PLANS:
PREHOOK: query: select year(l_shipdate) as year,
month(l_shipdate) as month,
count(l_shipdate) as monthly_shipments
-from lineitem
+from lineitem_ix
group by year(l_shipdate), month(l_shipdate)
order by year, month
PREHOOK: type: QUERY
-PREHOOK: Input: default@lineitem
+PREHOOK: Input: default@lineitem_ix
#### A masked pattern was here ####
POSTHOOK: query: select year(l_shipdate) as year,
month(l_shipdate) as month,
count(l_shipdate) as monthly_shipments
-from lineitem
+from lineitem_ix
group by year(l_shipdate), month(l_shipdate)
order by year, month
POSTHOOK: type: QUERY
-POSTHOOK: Input: default@lineitem
+POSTHOOK: Input: default@lineitem_ix
#### A masked pattern was here ####
1992 4 1
1992 7 3
@@ -563,14 +563,14 @@ POSTHOOK: Input: default@lineitem
PREHOOK: query: explain select year(l_shipdate) as year,
month(l_shipdate) as month,
count(l_shipdate) as monthly_shipments
-from lineitem
+from lineitem_ix
group by year(l_shipdate), month(l_shipdate)
order by year, month
PREHOOK: type: QUERY
POSTHOOK: query: explain select year(l_shipdate) as year,
month(l_shipdate) as month,
count(l_shipdate) as monthly_shipments
-from lineitem
+from lineitem_ix
group by year(l_shipdate), month(l_shipdate)
order by year, month
POSTHOOK: type: QUERY
@@ -584,23 +584,23 @@ STAGE PLANS:
Map Reduce
Map Operator Tree:
TableScan
- alias: default.default__lineitem_lineitem_lshipdate_idx__
- Statistics: Num rows: 95 Data size: 8675 Basic stats: COMPLETE Column stats: NONE
+ alias: default.default__lineitem_ix_lineitem_ix_lshipdate_idx__
+ Statistics: Num rows: 95 Data size: 8960 Basic stats: COMPLETE Column stats: NONE
Select Operator
expressions: l_shipdate (type: string), _count_of_l_shipdate (type: bigint)
outputColumnNames: l_shipdate, _count_of_l_shipdate
- Statistics: Num rows: 95 Data size: 8675 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 95 Data size: 8960 Basic stats: COMPLETE Column stats: NONE
Group By Operator
aggregations: sum(_count_of_l_shipdate)
keys: year(l_shipdate) (type: int), month(l_shipdate) (type: int)
mode: hash
outputColumnNames: _col0, _col1, _col2
- Statistics: Num rows: 95 Data size: 8675 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 95 Data size: 8960 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
key expressions: _col0 (type: int), _col1 (type: int)
sort order: ++
Map-reduce partition columns: _col0 (type: int), _col1 (type: int)
- Statistics: Num rows: 95 Data size: 8675 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 95 Data size: 8960 Basic stats: COMPLETE Column stats: NONE
value expressions: _col2 (type: bigint)
Reduce Operator Tree:
Group By Operator
@@ -608,11 +608,11 @@ STAGE PLANS:
keys: KEY._col0 (type: int), KEY._col1 (type: int)
mode: mergepartial
outputColumnNames: _col0, _col1, _col2
- Statistics: Num rows: 47 Data size: 4291 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 47 Data size: 4432 Basic stats: COMPLETE Column stats: NONE
Select Operator
expressions: _col0 (type: int), _col1 (type: int), _col2 (type: bigint)
outputColumnNames: _col0, _col1, _col2
- Statistics: Num rows: 47 Data size: 4291 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 47 Data size: 4432 Basic stats: COMPLETE Column stats: NONE
File Output Operator
compressed: false
table:
@@ -627,16 +627,16 @@ STAGE PLANS:
Reduce Output Operator
key expressions: _col0 (type: int), _col1 (type: int)
sort order: ++
- Statistics: Num rows: 47 Data size: 4291 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 47 Data size: 4432 Basic stats: COMPLETE Column stats: NONE
value expressions: _col2 (type: bigint)
Reduce Operator Tree:
Select Operator
expressions: KEY.reducesinkkey0 (type: int), KEY.reducesinkkey1 (type: int), VALUE._col0 (type: bigint)
outputColumnNames: _col0, _col1, _col2
- Statistics: Num rows: 47 Data size: 4291 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 47 Data size: 4432 Basic stats: COMPLETE Column stats: NONE
File Output Operator
compressed: false
- Statistics: Num rows: 47 Data size: 4291 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 47 Data size: 4432 Basic stats: COMPLETE Column stats: NONE
table:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
@@ -651,22 +651,22 @@ STAGE PLANS:
PREHOOK: query: select year(l_shipdate) as year,
month(l_shipdate) as month,
count(l_shipdate) as monthly_shipments
-from lineitem
+from lineitem_ix
group by year(l_shipdate), month(l_shipdate)
order by year, month
PREHOOK: type: QUERY
-PREHOOK: Input: default@default__lineitem_lineitem_lshipdate_idx__
-PREHOOK: Input: default@lineitem
+PREHOOK: Input: default@default__lineitem_ix_lineitem_ix_lshipdate_idx__
+PREHOOK: Input: default@lineitem_ix
#### A masked pattern was here ####
POSTHOOK: query: select year(l_shipdate) as year,
month(l_shipdate) as month,
count(l_shipdate) as monthly_shipments
-from lineitem
+from lineitem_ix
group by year(l_shipdate), month(l_shipdate)
order by year, month
POSTHOOK: type: QUERY
-POSTHOOK: Input: default@default__lineitem_lineitem_lshipdate_idx__
-POSTHOOK: Input: default@lineitem
+POSTHOOK: Input: default@default__lineitem_ix_lineitem_ix_lshipdate_idx__
+POSTHOOK: Input: default@lineitem_ix
#### A masked pattern was here ####
1992 4 1
1992 7 3
@@ -718,14 +718,14 @@ lastyear.monthly_shipments as monthly_sh
from (select year(l_shipdate) as year,
month(l_shipdate) as month,
count(l_shipdate) as monthly_shipments
- from lineitem
+ from lineitem_ix
where year(l_shipdate) = 1997
group by year(l_shipdate), month(l_shipdate)
) lastyear join
(select year(l_shipdate) as year,
month(l_shipdate) as month,
count(l_shipdate) as monthly_shipments
- from lineitem
+ from lineitem_ix
where year(l_shipdate) = 1998
group by year(l_shipdate), month(l_shipdate)
) thisyear
@@ -738,14 +738,14 @@ lastyear.monthly_shipments as monthly_sh
from (select year(l_shipdate) as year,
month(l_shipdate) as month,
count(l_shipdate) as monthly_shipments
- from lineitem
+ from lineitem_ix
where year(l_shipdate) = 1997
group by year(l_shipdate), month(l_shipdate)
) lastyear join
(select year(l_shipdate) as year,
month(l_shipdate) as month,
count(l_shipdate) as monthly_shipments
- from lineitem
+ from lineitem_ix
where year(l_shipdate) = 1998
group by year(l_shipdate), month(l_shipdate)
) thisyear
@@ -762,29 +762,29 @@ STAGE PLANS:
Map Reduce
Map Operator Tree:
TableScan
- alias: thisyear:default.default__lineitem_lineitem_lshipdate_idx__
- Statistics: Num rows: 95 Data size: 8675 Basic stats: COMPLETE Column stats: NONE
+ alias: lastyear:default.default__lineitem_ix_lineitem_ix_lshipdate_idx__
+ Statistics: Num rows: 95 Data size: 8960 Basic stats: COMPLETE Column stats: NONE
Filter Operator
- predicate: (year(l_shipdate) = 1998) (type: boolean)
- Statistics: Num rows: 47 Data size: 4291 Basic stats: COMPLETE Column stats: NONE
+ predicate: (year(l_shipdate) = 1997) (type: boolean)
+ Statistics: Num rows: 47 Data size: 4432 Basic stats: COMPLETE Column stats: NONE
Select Operator
expressions: l_shipdate (type: string), _count_of_l_shipdate (type: bigint)
outputColumnNames: l_shipdate, _count_of_l_shipdate
- Statistics: Num rows: 47 Data size: 4291 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 47 Data size: 4432 Basic stats: COMPLETE Column stats: NONE
Group By Operator
aggregations: sum(_count_of_l_shipdate)
keys: year(l_shipdate) (type: int), month(l_shipdate) (type: int)
mode: hash
outputColumnNames: _col0, _col1, _col2
- Statistics: Num rows: 23 Data size: 2099 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 23 Data size: 2168 Basic stats: COMPLETE Column stats: NONE
Filter Operator
predicate: _col1 is not null (type: boolean)
- Statistics: Num rows: 12 Data size: 1095 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 12 Data size: 1131 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
key expressions: _col0 (type: int), _col1 (type: int)
sort order: ++
Map-reduce partition columns: _col0 (type: int), _col1 (type: int)
- Statistics: Num rows: 12 Data size: 1095 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 12 Data size: 1131 Basic stats: COMPLETE Column stats: NONE
value expressions: _col2 (type: bigint)
Reduce Operator Tree:
Group By Operator
@@ -792,11 +792,11 @@ STAGE PLANS:
keys: KEY._col0 (type: int), KEY._col1 (type: int)
mode: mergepartial
outputColumnNames: _col0, _col1, _col2
- Statistics: Num rows: 6 Data size: 547 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 6 Data size: 565 Basic stats: COMPLETE Column stats: NONE
Select Operator
expressions: _col1 (type: int), _col2 (type: bigint)
outputColumnNames: _col1, _col2
- Statistics: Num rows: 6 Data size: 547 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 6 Data size: 565 Basic stats: COMPLETE Column stats: NONE
File Output Operator
compressed: false
table:
@@ -812,14 +812,14 @@ STAGE PLANS:
key expressions: _col1 (type: int)
sort order: +
Map-reduce partition columns: _col1 (type: int)
- Statistics: Num rows: 6 Data size: 547 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 6 Data size: 565 Basic stats: COMPLETE Column stats: NONE
value expressions: _col2 (type: bigint)
TableScan
Reduce Output Operator
key expressions: _col1 (type: int)
sort order: +
Map-reduce partition columns: _col1 (type: int)
- Statistics: Num rows: 6 Data size: 547 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 6 Data size: 565 Basic stats: COMPLETE Column stats: NONE
value expressions: _col2 (type: bigint)
Reduce Operator Tree:
Join Operator
@@ -829,14 +829,14 @@ STAGE PLANS:
0 {KEY.reducesinkkey0} {VALUE._col1}
1 {KEY.reducesinkkey0} {VALUE._col1}
outputColumnNames: _col1, _col2, _col4, _col5
- Statistics: Num rows: 6 Data size: 601 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 6 Data size: 621 Basic stats: COMPLETE Column stats: NONE
Select Operator
expressions: _col1 (type: int), _col4 (type: int), ((_col5 - _col2) / _col2) (type: double)
outputColumnNames: _col0, _col1, _col2
- Statistics: Num rows: 6 Data size: 601 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 6 Data size: 621 Basic stats: COMPLETE Column stats: NONE
File Output Operator
compressed: false
- Statistics: Num rows: 6 Data size: 601 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 6 Data size: 621 Basic stats: COMPLETE Column stats: NONE
table:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
@@ -846,29 +846,29 @@ STAGE PLANS:
Map Reduce
Map Operator Tree:
TableScan
- alias: lastyear:default.default__lineitem_lineitem_lshipdate_idx__
- Statistics: Num rows: 95 Data size: 8675 Basic stats: COMPLETE Column stats: NONE
+ alias: thisyear:default.default__lineitem_ix_lineitem_ix_lshipdate_idx__
+ Statistics: Num rows: 95 Data size: 8960 Basic stats: COMPLETE Column stats: NONE
Filter Operator
- predicate: (year(l_shipdate) = 1997) (type: boolean)
- Statistics: Num rows: 47 Data size: 4291 Basic stats: COMPLETE Column stats: NONE
+ predicate: (year(l_shipdate) = 1998) (type: boolean)
+ Statistics: Num rows: 47 Data size: 4432 Basic stats: COMPLETE Column stats: NONE
Select Operator
expressions: l_shipdate (type: string), _count_of_l_shipdate (type: bigint)
outputColumnNames: l_shipdate, _count_of_l_shipdate
- Statistics: Num rows: 47 Data size: 4291 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 47 Data size: 4432 Basic stats: COMPLETE Column stats: NONE
Group By Operator
aggregations: sum(_count_of_l_shipdate)
keys: year(l_shipdate) (type: int), month(l_shipdate) (type: int)
mode: hash
outputColumnNames: _col0, _col1, _col2
- Statistics: Num rows: 23 Data size: 2099 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 23 Data size: 2168 Basic stats: COMPLETE Column stats: NONE
Filter Operator
predicate: _col1 is not null (type: boolean)
- Statistics: Num rows: 12 Data size: 1095 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 12 Data size: 1131 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
key expressions: _col0 (type: int), _col1 (type: int)
sort order: ++
Map-reduce partition columns: _col0 (type: int), _col1 (type: int)
- Statistics: Num rows: 12 Data size: 1095 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 12 Data size: 1131 Basic stats: COMPLETE Column stats: NONE
value expressions: _col2 (type: bigint)
Reduce Operator Tree:
Group By Operator
@@ -876,11 +876,11 @@ STAGE PLANS:
keys: KEY._col0 (type: int), KEY._col1 (type: int)
mode: mergepartial
outputColumnNames: _col0, _col1, _col2
- Statistics: Num rows: 6 Data size: 547 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 6 Data size: 565 Basic stats: COMPLETE Column stats: NONE
Select Operator
expressions: _col1 (type: int), _col2 (type: bigint)
outputColumnNames: _col1, _col2
- Statistics: Num rows: 6 Data size: 547 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 6 Data size: 565 Basic stats: COMPLETE Column stats: NONE
File Output Operator
compressed: false
table:
@@ -895,44 +895,44 @@ STAGE PLANS:
ListSink
PREHOOK: query: explain select l_shipdate, cnt
-from (select l_shipdate, count(l_shipdate) as cnt from lineitem group by l_shipdate
+from (select l_shipdate, count(l_shipdate) as cnt from lineitem_ix group by l_shipdate
union all
select l_shipdate, l_orderkey as cnt
-from lineitem) dummy
+from lineitem_ix) dummy
PREHOOK: type: QUERY
POSTHOOK: query: explain select l_shipdate, cnt
-from (select l_shipdate, count(l_shipdate) as cnt from lineitem group by l_shipdate
+from (select l_shipdate, count(l_shipdate) as cnt from lineitem_ix group by l_shipdate
union all
select l_shipdate, l_orderkey as cnt
-from lineitem) dummy
+from lineitem_ix) dummy
POSTHOOK: type: QUERY
STAGE DEPENDENCIES:
- Stage-3 is a root stage
- Stage-2 depends on stages: Stage-3
+ Stage-1 is a root stage
+ Stage-2 depends on stages: Stage-1
Stage-0 depends on stages: Stage-2
STAGE PLANS:
- Stage: Stage-3
+ Stage: Stage-1
Map Reduce
Map Operator Tree:
TableScan
- alias: null-subquery1:dummy-subquery1:default.default__lineitem_lineitem_lshipdate_idx__
- Statistics: Num rows: 95 Data size: 8675 Basic stats: COMPLETE Column stats: NONE
+ alias: null-subquery1:dummy-subquery1:default.default__lineitem_ix_lineitem_ix_lshipdate_idx__
+ Statistics: Num rows: 95 Data size: 8960 Basic stats: COMPLETE Column stats: NONE
Select Operator
expressions: l_shipdate (type: string), _count_of_l_shipdate (type: bigint)
outputColumnNames: l_shipdate, _count_of_l_shipdate
- Statistics: Num rows: 95 Data size: 8675 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 95 Data size: 8960 Basic stats: COMPLETE Column stats: NONE
Group By Operator
aggregations: sum(_count_of_l_shipdate)
keys: l_shipdate (type: string)
mode: hash
outputColumnNames: _col0, _col1
- Statistics: Num rows: 95 Data size: 8675 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 95 Data size: 8960 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
key expressions: _col0 (type: string)
sort order: +
Map-reduce partition columns: _col0 (type: string)
- Statistics: Num rows: 95 Data size: 8675 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 95 Data size: 8960 Basic stats: COMPLETE Column stats: NONE
value expressions: _col1 (type: bigint)
Reduce Operator Tree:
Group By Operator
@@ -940,11 +940,11 @@ STAGE PLANS:
keys: KEY._col0 (type: string)
mode: mergepartial
outputColumnNames: _col0, _col1
- Statistics: Num rows: 47 Data size: 4291 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 47 Data size: 4432 Basic stats: COMPLETE Column stats: NONE
Select Operator
expressions: _col0 (type: string), _col1 (type: bigint)
outputColumnNames: _col0, _col1
- Statistics: Num rows: 47 Data size: 4291 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 47 Data size: 4432 Basic stats: COMPLETE Column stats: NONE
File Output Operator
compressed: false
table:
@@ -956,39 +956,39 @@ STAGE PLANS:
Map Reduce
Map Operator Tree:
TableScan
- alias: lineitem
+ Union
+ Statistics: Num rows: 163 Data size: 16531 Basic stats: COMPLETE Column stats: NONE
+ Select Operator
+ expressions: _col0 (type: string), _col1 (type: bigint)
+ outputColumnNames: _col0, _col1
+ Statistics: Num rows: 163 Data size: 16531 Basic stats: COMPLETE Column stats: NONE
+ File Output Operator
+ compressed: false
+ Statistics: Num rows: 163 Data size: 16531 Basic stats: COMPLETE Column stats: NONE
+ table:
+ input format: org.apache.hadoop.mapred.TextInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ TableScan
+ alias: lineitem_ix
Statistics: Num rows: 116 Data size: 12099 Basic stats: COMPLETE Column stats: NONE
Select Operator
expressions: l_shipdate (type: string), UDFToLong(l_orderkey) (type: bigint)
outputColumnNames: _col0, _col1
Statistics: Num rows: 116 Data size: 12099 Basic stats: COMPLETE Column stats: NONE
Union
- Statistics: Num rows: 163 Data size: 16390 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 163 Data size: 16531 Basic stats: COMPLETE Column stats: NONE
Select Operator
expressions: _col0 (type: string), _col1 (type: bigint)
outputColumnNames: _col0, _col1
- Statistics: Num rows: 163 Data size: 16390 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 163 Data size: 16531 Basic stats: COMPLETE Column stats: NONE
File Output Operator
compressed: false
- Statistics: Num rows: 163 Data size: 16390 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 163 Data size: 16531 Basic stats: COMPLETE Column stats: NONE
table:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- TableScan
- Union
- Statistics: Num rows: 163 Data size: 16390 Basic stats: COMPLETE Column stats: NONE
- Select Operator
- expressions: _col0 (type: string), _col1 (type: bigint)
- outputColumnNames: _col0, _col1
- Statistics: Num rows: 163 Data size: 16390 Basic stats: COMPLETE Column stats: NONE
- File Output Operator
- compressed: false
- Statistics: Num rows: 163 Data size: 16390 Basic stats: COMPLETE Column stats: NONE
- table:
- input format: org.apache.hadoop.mapred.TextInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
Stage: Stage-0
Fetch Operator
Modified: hive/branches/spark/ql/src/test/results/clientpositive/reducesink_dedup.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/reducesink_dedup.q.out?rev=1637444&r1=1637443&r2=1637444&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/reducesink_dedup.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/reducesink_dedup.q.out Fri Nov 7 20:41:34 2014
@@ -1,37 +1,3 @@
-PREHOOK: query: DROP TABLE part
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: DROP TABLE part
-POSTHOOK: type: DROPTABLE
-PREHOOK: query: -- data setup
-CREATE TABLE part(
- p_partkey INT,
- p_name STRING,
- p_mfgr STRING,
- p_brand STRING,
- p_type STRING,
- p_size INT,
- p_container STRING,
- p_retailprice DOUBLE,
- p_comment STRING
-)
-PREHOOK: type: CREATETABLE
-PREHOOK: Output: database:default
-PREHOOK: Output: default@part
-POSTHOOK: query: -- data setup
-CREATE TABLE part(
- p_partkey INT,
- p_name STRING,
- p_mfgr STRING,
- p_brand STRING,
- p_type STRING,
- p_size INT,
- p_container STRING,
- p_retailprice DOUBLE,
- p_comment STRING
-)
-POSTHOOK: type: CREATETABLE
-POSTHOOK: Output: database:default
-POSTHOOK: Output: default@part
PREHOOK: query: select p_name
from (select p_name from part distribute by 1 sort by 1) p
distribute by 1 sort by 1
@@ -44,3 +10,29 @@ distribute by 1 sort by 1
POSTHOOK: type: QUERY
POSTHOOK: Input: default@part
#### A masked pattern was here ####
+almond azure blanched chiffon midnight
+almond aquamarine dodger light gainsboro
+almond antique sky peru orange
+almond antique medium spring khaki
+almond antique blue firebrick mint
+almond azure aquamarine papaya violet
+almond aquamarine yellow dodger mint
+almond aquamarine floral ivory bisque
+almond antique violet mint lemon
+almond antique gainsboro frosted violet
+almond antique olive coral navajo
+almond antique misty red olive
+almond antique metallic orange dim
+almond antique forest lavender goldenrod
+almond antique chartreuse khaki white
+almond aquamarine sandy cyan gainsboro
+almond aquamarine rose maroon antique
+almond aquamarine midnight light salmon
+almond antique violet turquoise frosted
+almond antique violet chocolate turquoise
+almond aquamarine pink moccasin thistle
+almond aquamarine burnished black steel
+almond antique salmon chartreuse burlywood
+almond antique chartreuse lavender yellow
+almond antique burnished rose metallic
+almond antique burnished rose metallic
Modified: hive/branches/spark/ql/src/test/results/clientpositive/rename_column.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/rename_column.q.out?rev=1637444&r1=1637443&r2=1637444&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/rename_column.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/rename_column.q.out Fri Nov 7 20:41:34 2014
@@ -147,8 +147,14 @@ PREHOOK: type: SHOWTABLES
POSTHOOK: query: SHOW TABLES
POSTHOOK: type: SHOWTABLES
alltypesorc
+cbo_t1
+cbo_t2
+cbo_t3
+lineitem
+part
src
src1
+src_cbo
src_json
src_sequencefile
src_thrift
Modified: hive/branches/spark/ql/src/test/results/clientpositive/scriptfile1_win.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/scriptfile1_win.q.out?rev=1637444&r1=1637443&r2=1637444&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/scriptfile1_win.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/scriptfile1_win.q.out Fri Nov 7 20:41:34 2014
@@ -3,6 +3,7 @@ PREHOOK: query: -- INCLUDE_OS_WINDOWS
CREATE TABLE dest1(key INT, value STRING)
PREHOOK: type: CREATETABLE
PREHOOK: Output: database:default
+PREHOOK: Output: default@dest1
POSTHOOK: query: -- INCLUDE_OS_WINDOWS
CREATE TABLE dest1(key INT, value STRING)
@@ -39,8 +40,6 @@ POSTHOOK: query: SELECT dest1.* FROM des
POSTHOOK: type: QUERY
POSTHOOK: Input: default@dest1
#### A masked pattern was here ####
-POSTHOOK: Lineage: dest1.key SCRIPT [(src)src.FieldSchema(name:key, type:string, comment:default), (src)src.FieldSchema(name:value, type:string, comment:default), ]
-POSTHOOK: Lineage: dest1.value SCRIPT [(src)src.FieldSchema(name:key, type:string, comment:default), (src)src.FieldSchema(name:value, type:string, comment:default), ]
NULL NULL
NULL NULL
10 val_10
Modified: hive/branches/spark/ql/src/test/results/clientpositive/show_tables.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/show_tables.q.out?rev=1637444&r1=1637443&r2=1637444&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/show_tables.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/show_tables.q.out Fri Nov 7 20:41:34 2014
@@ -130,10 +130,16 @@ PREHOOK: type: SHOWTABLES
POSTHOOK: query: SHOW TABLES FROM default
POSTHOOK: type: SHOWTABLES
alltypesorc
+cbo_t1
+cbo_t2
+cbo_t3
+lineitem
+part
shtb_test1
shtb_test2
src
src1
+src_cbo
src_json
src_sequencefile
src_thrift
@@ -152,10 +158,16 @@ PREHOOK: type: SHOWTABLES
POSTHOOK: query: SHOW TABLES IN default
POSTHOOK: type: SHOWTABLES
alltypesorc
+cbo_t1
+cbo_t2
+cbo_t3
+lineitem
+part
shtb_test1
shtb_test2
src
src1
+src_cbo
src_json
src_sequencefile
src_thrift
Modified: hive/branches/spark/ql/src/test/results/clientpositive/subquery_in.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/subquery_in.q.out?rev=1637444&r1=1637443&r2=1637444&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/subquery_in.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/subquery_in.q.out Fri Nov 7 20:41:34 2014
@@ -1,110 +1,14 @@
PREHOOK: query: -- SORT_QUERY_RESULTS
-DROP TABLE part
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: -- SORT_QUERY_RESULTS
-
-DROP TABLE part
-POSTHOOK: type: DROPTABLE
-PREHOOK: query: -- data setup
-CREATE TABLE part(
- p_partkey INT,
- p_name STRING,
- p_mfgr STRING,
- p_brand STRING,
- p_type STRING,
- p_size INT,
- p_container STRING,
- p_retailprice DOUBLE,
- p_comment STRING
-)
-PREHOOK: type: CREATETABLE
-PREHOOK: Output: database:default
-PREHOOK: Output: default@part
-POSTHOOK: query: -- data setup
-CREATE TABLE part(
- p_partkey INT,
- p_name STRING,
- p_mfgr STRING,
- p_brand STRING,
- p_type STRING,
- p_size INT,
- p_container STRING,
- p_retailprice DOUBLE,
- p_comment STRING
-)
-POSTHOOK: type: CREATETABLE
-POSTHOOK: Output: database:default
-POSTHOOK: Output: default@part
-PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/part_tiny.txt' overwrite into table part
-PREHOOK: type: LOAD
-#### A masked pattern was here ####
-PREHOOK: Output: default@part
-POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/part_tiny.txt' overwrite into table part
-POSTHOOK: type: LOAD
-#### A masked pattern was here ####
-POSTHOOK: Output: default@part
-PREHOOK: query: DROP TABLE lineitem
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: DROP TABLE lineitem
-POSTHOOK: type: DROPTABLE
-PREHOOK: query: CREATE TABLE lineitem (L_ORDERKEY INT,
- L_PARTKEY INT,
- L_SUPPKEY INT,
- L_LINENUMBER INT,
- L_QUANTITY DOUBLE,
- L_EXTENDEDPRICE DOUBLE,
- L_DISCOUNT DOUBLE,
- L_TAX DOUBLE,
- L_RETURNFLAG STRING,
- L_LINESTATUS STRING,
- l_shipdate STRING,
- L_COMMITDATE STRING,
- L_RECEIPTDATE STRING,
- L_SHIPINSTRUCT STRING,
- L_SHIPMODE STRING,
- L_COMMENT STRING)
-ROW FORMAT DELIMITED
-FIELDS TERMINATED BY '|'
-PREHOOK: type: CREATETABLE
-PREHOOK: Output: database:default
-PREHOOK: Output: default@lineitem
-POSTHOOK: query: CREATE TABLE lineitem (L_ORDERKEY INT,
- L_PARTKEY INT,
- L_SUPPKEY INT,
- L_LINENUMBER INT,
- L_QUANTITY DOUBLE,
- L_EXTENDEDPRICE DOUBLE,
- L_DISCOUNT DOUBLE,
- L_TAX DOUBLE,
- L_RETURNFLAG STRING,
- L_LINESTATUS STRING,
- l_shipdate STRING,
- L_COMMITDATE STRING,
- L_RECEIPTDATE STRING,
- L_SHIPINSTRUCT STRING,
- L_SHIPMODE STRING,
- L_COMMENT STRING)
-ROW FORMAT DELIMITED
-FIELDS TERMINATED BY '|'
-POSTHOOK: type: CREATETABLE
-POSTHOOK: Output: database:default
-POSTHOOK: Output: default@lineitem
-PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/lineitem.txt' OVERWRITE INTO TABLE lineitem
-PREHOOK: type: LOAD
-#### A masked pattern was here ####
-PREHOOK: Output: default@lineitem
-POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/lineitem.txt' OVERWRITE INTO TABLE lineitem
-POSTHOOK: type: LOAD
-#### A masked pattern was here ####
-POSTHOOK: Output: default@lineitem
-PREHOOK: query: -- non agg, non corr
+-- non agg, non corr
explain
select *
from src
where src.key in (select key from src s1 where s1.key > '9')
PREHOOK: type: QUERY
-POSTHOOK: query: -- non agg, non corr
+POSTHOOK: query: -- SORT_QUERY_RESULTS
+
+-- non agg, non corr
explain
select *
from src
@@ -349,25 +253,25 @@ STAGE PLANS:
Map Operator Tree:
TableScan
alias: part
- Statistics: Num rows: 30 Data size: 3173 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
key expressions: p_mfgr (type: string), p_size (type: int)
sort order: ++
Map-reduce partition columns: p_mfgr (type: string)
- Statistics: Num rows: 30 Data size: 3173 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
value expressions: p_mfgr (type: string), p_size (type: int)
Reduce Operator Tree:
Extract
- Statistics: Num rows: 30 Data size: 3173 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
PTF Operator
- Statistics: Num rows: 30 Data size: 3173 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
Filter Operator
predicate: (_wcol0 <= 2) (type: boolean)
- Statistics: Num rows: 10 Data size: 1057 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 8 Data size: 968 Basic stats: COMPLETE Column stats: NONE
Select Operator
expressions: _col5 (type: int)
outputColumnNames: _col0
- Statistics: Num rows: 10 Data size: 1057 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 8 Data size: 968 Basic stats: COMPLETE Column stats: NONE
Group By Operator
aggregations: avg(_col0)
mode: hash
@@ -424,15 +328,15 @@ STAGE PLANS:
Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE
TableScan
alias: part
- Statistics: Num rows: 30 Data size: 3173 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
Filter Operator
predicate: UDFToDouble(p_size) is not null (type: boolean)
- Statistics: Num rows: 15 Data size: 1586 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 13 Data size: 1573 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
key expressions: UDFToDouble(p_size) (type: double)
sort order: +
Map-reduce partition columns: UDFToDouble(p_size) (type: double)
- Statistics: Num rows: 15 Data size: 1586 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 13 Data size: 1573 Basic stats: COMPLETE Column stats: NONE
value expressions: p_name (type: string), p_size (type: int)
Reduce Operator Tree:
Join Operator
@@ -442,14 +346,14 @@ STAGE PLANS:
0 {VALUE._col1} {VALUE._col5}
1
outputColumnNames: _col1, _col5
- Statistics: Num rows: 16 Data size: 1744 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 14 Data size: 1730 Basic stats: COMPLETE Column stats: NONE
Select Operator
expressions: _col1 (type: string), _col5 (type: int)
outputColumnNames: _col0, _col1
- Statistics: Num rows: 16 Data size: 1744 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 14 Data size: 1730 Basic stats: COMPLETE Column stats: NONE
File Output Operator
compressed: false
- Statistics: Num rows: 16 Data size: 1744 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 14 Data size: 1730 Basic stats: COMPLETE Column stats: NONE
table:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
@@ -513,31 +417,31 @@ STAGE PLANS:
Map Operator Tree:
TableScan
alias: part
- Statistics: Num rows: 30 Data size: 3173 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
key expressions: p_mfgr (type: string), p_size (type: int)
sort order: ++
Map-reduce partition columns: p_mfgr (type: string)
- Statistics: Num rows: 30 Data size: 3173 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
value expressions: p_mfgr (type: string), p_size (type: int)
Reduce Operator Tree:
Extract
- Statistics: Num rows: 30 Data size: 3173 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
PTF Operator
- Statistics: Num rows: 30 Data size: 3173 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
Filter Operator
predicate: ((_wcol0 <= 2) and _col2 is not null) (type: boolean)
- Statistics: Num rows: 5 Data size: 528 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 4 Data size: 484 Basic stats: COMPLETE Column stats: NONE
Select Operator
expressions: _col2 (type: string), _col5 (type: int)
outputColumnNames: _col0, _col1
- Statistics: Num rows: 5 Data size: 528 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 4 Data size: 484 Basic stats: COMPLETE Column stats: NONE
Group By Operator
aggregations: min(_col1)
keys: _col0 (type: string)
mode: hash
outputColumnNames: _col0, _col1
- Statistics: Num rows: 5 Data size: 528 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 4 Data size: 484 Basic stats: COMPLETE Column stats: NONE
File Output Operator
compressed: false
table:
@@ -553,7 +457,7 @@ STAGE PLANS:
key expressions: _col0 (type: string)
sort order: +
Map-reduce partition columns: _col0 (type: string)
- Statistics: Num rows: 5 Data size: 528 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 4 Data size: 484 Basic stats: COMPLETE Column stats: NONE
value expressions: _col1 (type: int)
Reduce Operator Tree:
Group By Operator
@@ -561,19 +465,19 @@ STAGE PLANS:
keys: KEY._col0 (type: string)
mode: mergepartial
outputColumnNames: _col0, _col1
- Statistics: Num rows: 2 Data size: 211 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 2 Data size: 242 Basic stats: COMPLETE Column stats: NONE
Filter Operator
predicate: _col1 is not null (type: boolean)
- Statistics: Num rows: 2 Data size: 211 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 2 Data size: 242 Basic stats: COMPLETE Column stats: NONE
Select Operator
expressions: _col1 (type: int), _col0 (type: string)
outputColumnNames: _col0, _col1
- Statistics: Num rows: 2 Data size: 211 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 2 Data size: 242 Basic stats: COMPLETE Column stats: NONE
Group By Operator
keys: _col0 (type: int), _col1 (type: string)
mode: hash
outputColumnNames: _col0, _col1
- Statistics: Num rows: 2 Data size: 211 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 2 Data size: 242 Basic stats: COMPLETE Column stats: NONE
File Output Operator
compressed: false
table:
@@ -586,22 +490,22 @@ STAGE PLANS:
Map Operator Tree:
TableScan
alias: b
- Statistics: Num rows: 15 Data size: 3173 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
Filter Operator
predicate: (p_size is not null and p_mfgr is not null) (type: boolean)
- Statistics: Num rows: 4 Data size: 846 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 7 Data size: 847 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
key expressions: p_size (type: int), p_mfgr (type: string)
sort order: ++
Map-reduce partition columns: p_size (type: int), p_mfgr (type: string)
- Statistics: Num rows: 4 Data size: 846 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 7 Data size: 847 Basic stats: COMPLETE Column stats: NONE
value expressions: p_name (type: string)
TableScan
Reduce Output Operator
key expressions: _col0 (type: int), _col1 (type: string)
sort order: ++
Map-reduce partition columns: _col0 (type: int), _col1 (type: string)
- Statistics: Num rows: 2 Data size: 211 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 2 Data size: 242 Basic stats: COMPLETE Column stats: NONE
Reduce Operator Tree:
Join Operator
condition map:
@@ -610,14 +514,14 @@ STAGE PLANS:
0 {VALUE._col1} {KEY.reducesinkkey1} {KEY.reducesinkkey0}
1
outputColumnNames: _col1, _col2, _col5
- Statistics: Num rows: 4 Data size: 930 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 7 Data size: 931 Basic stats: COMPLETE Column stats: NONE
Select Operator
expressions: _col2 (type: string), _col1 (type: string), _col5 (type: int)
outputColumnNames: _col0, _col1, _col2
- Statistics: Num rows: 4 Data size: 930 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 7 Data size: 931 Basic stats: COMPLETE Column stats: NONE
File Output Operator
compressed: false
- Statistics: Num rows: 4 Data size: 930 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 7 Data size: 931 Basic stats: COMPLETE Column stats: NONE
table:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
@@ -850,34 +754,34 @@ STAGE PLANS:
Map Operator Tree:
TableScan
alias: lineitem
- Statistics: Num rows: 3024 Data size: 12099 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 100 Data size: 11999 Basic stats: COMPLETE Column stats: NONE
Filter Operator
predicate: l_partkey is not null (type: boolean)
- Statistics: Num rows: 1512 Data size: 6049 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 50 Data size: 5999 Basic stats: COMPLETE Column stats: NONE
Select Operator
expressions: l_partkey (type: int)
outputColumnNames: l_partkey
- Statistics: Num rows: 1512 Data size: 6049 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 50 Data size: 5999 Basic stats: COMPLETE Column stats: NONE
Group By Operator
keys: l_partkey (type: int)
mode: hash
outputColumnNames: _col0
- Statistics: Num rows: 1512 Data size: 6049 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 50 Data size: 5999 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
key expressions: _col0 (type: int)
sort order: +
Map-reduce partition columns: _col0 (type: int)
- Statistics: Num rows: 1512 Data size: 6049 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 50 Data size: 5999 Basic stats: COMPLETE Column stats: NONE
Reduce Operator Tree:
Group By Operator
keys: KEY._col0 (type: int)
mode: mergepartial
outputColumnNames: _col0
- Statistics: Num rows: 756 Data size: 3024 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 25 Data size: 2999 Basic stats: COMPLETE Column stats: NONE
Select Operator
expressions: _col0 (type: int)
outputColumnNames: _col0
- Statistics: Num rows: 756 Data size: 3024 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 25 Data size: 2999 Basic stats: COMPLETE Column stats: NONE
File Output Operator
compressed: false
table:
@@ -890,22 +794,22 @@ STAGE PLANS:
Map Operator Tree:
TableScan
alias: li
- Statistics: Num rows: 756 Data size: 12099 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 100 Data size: 11999 Basic stats: COMPLETE Column stats: NONE
Filter Operator
predicate: ((l_partkey is not null and l_orderkey is not null) and (l_linenumber = 1)) (type: boolean)
- Statistics: Num rows: 94 Data size: 1504 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 12 Data size: 1439 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
key expressions: l_partkey (type: int)
sort order: +
Map-reduce partition columns: l_partkey (type: int)
- Statistics: Num rows: 94 Data size: 1504 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 12 Data size: 1439 Basic stats: COMPLETE Column stats: NONE
value expressions: l_orderkey (type: int), l_suppkey (type: int)
TableScan
Reduce Output Operator
key expressions: _col0 (type: int)
sort order: +
Map-reduce partition columns: _col0 (type: int)
- Statistics: Num rows: 756 Data size: 3024 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 25 Data size: 2999 Basic stats: COMPLETE Column stats: NONE
Reduce Operator Tree:
Join Operator
condition map:
@@ -914,7 +818,7 @@ STAGE PLANS:
0 {KEY.reducesinkkey0}
1 {VALUE._col0} {VALUE._col1}
outputColumnNames: _col0, _col1, _col3
- Statistics: Num rows: 831 Data size: 3326 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 27 Data size: 3298 Basic stats: COMPLETE Column stats: NONE
File Output Operator
compressed: false
table:
@@ -930,28 +834,28 @@ STAGE PLANS:
key expressions: _col1 (type: int)
sort order: +
Map-reduce partition columns: _col1 (type: int)
- Statistics: Num rows: 831 Data size: 3326 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 27 Data size: 3298 Basic stats: COMPLETE Column stats: NONE
value expressions: _col0 (type: int), _col3 (type: int)
TableScan
alias: lineitem
- Statistics: Num rows: 1728 Data size: 12099 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 100 Data size: 11999 Basic stats: COMPLETE Column stats: NONE
Filter Operator
predicate: ((l_shipmode = 'AIR') and l_orderkey is not null) (type: boolean)
- Statistics: Num rows: 432 Data size: 3024 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 25 Data size: 2999 Basic stats: COMPLETE Column stats: NONE
Select Operator
expressions: l_orderkey (type: int)
outputColumnNames: _col0
- Statistics: Num rows: 432 Data size: 3024 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 25 Data size: 2999 Basic stats: COMPLETE Column stats: NONE
Group By Operator
keys: _col0 (type: int)
mode: hash
outputColumnNames: _col0
- Statistics: Num rows: 432 Data size: 3024 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 25 Data size: 2999 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
key expressions: _col0 (type: int)
sort order: +
Map-reduce partition columns: _col0 (type: int)
- Statistics: Num rows: 432 Data size: 3024 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 25 Data size: 2999 Basic stats: COMPLETE Column stats: NONE
Reduce Operator Tree:
Join Operator
condition map:
@@ -960,14 +864,14 @@ STAGE PLANS:
0 {VALUE._col0} {VALUE._col2}
1
outputColumnNames: _col0, _col3
- Statistics: Num rows: 914 Data size: 3658 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 29 Data size: 3627 Basic stats: COMPLETE Column stats: NONE
Select Operator
expressions: _col0 (type: int), _col3 (type: int)
outputColumnNames: _col0, _col1
- Statistics: Num rows: 914 Data size: 3658 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 29 Data size: 3627 Basic stats: COMPLETE Column stats: NONE
File Output Operator
compressed: false
- Statistics: Num rows: 914 Data size: 3658 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 29 Data size: 3627 Basic stats: COMPLETE Column stats: NONE
table:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
Modified: hive/branches/spark/ql/src/test/results/clientpositive/subquery_in_explain_rewrite.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/subquery_in_explain_rewrite.q.out?rev=1637444&r1=1637443&r2=1637444&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/subquery_in_explain_rewrite.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/subquery_in_explain_rewrite.q.out Fri Nov 7 20:41:34 2014
@@ -1,83 +1,3 @@
-PREHOOK: query: DROP TABLE part
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: DROP TABLE part
-POSTHOOK: type: DROPTABLE
-PREHOOK: query: -- data setup
-CREATE TABLE part(
- p_partkey INT,
- p_name STRING,
- p_mfgr STRING,
- p_brand STRING,
- p_type STRING,
- p_size INT,
- p_container STRING,
- p_retailprice DOUBLE,
- p_comment STRING
-)
-PREHOOK: type: CREATETABLE
-PREHOOK: Output: database:default
-PREHOOK: Output: default@part
-POSTHOOK: query: -- data setup
-CREATE TABLE part(
- p_partkey INT,
- p_name STRING,
- p_mfgr STRING,
- p_brand STRING,
- p_type STRING,
- p_size INT,
- p_container STRING,
- p_retailprice DOUBLE,
- p_comment STRING
-)
-POSTHOOK: type: CREATETABLE
-POSTHOOK: Output: database:default
-POSTHOOK: Output: default@part
-PREHOOK: query: DROP TABLE lineitem
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: DROP TABLE lineitem
-POSTHOOK: type: DROPTABLE
-PREHOOK: query: CREATE TABLE lineitem (L_ORDERKEY INT,
- L_PARTKEY INT,
- L_SUPPKEY INT,
- L_LINENUMBER INT,
- L_QUANTITY DOUBLE,
- L_EXTENDEDPRICE DOUBLE,
- L_DISCOUNT DOUBLE,
- L_TAX DOUBLE,
- L_RETURNFLAG STRING,
- L_LINESTATUS STRING,
- l_shipdate STRING,
- L_COMMITDATE STRING,
- L_RECEIPTDATE STRING,
- L_SHIPINSTRUCT STRING,
- L_SHIPMODE STRING,
- L_COMMENT STRING)
-ROW FORMAT DELIMITED
-FIELDS TERMINATED BY '|'
-PREHOOK: type: CREATETABLE
-PREHOOK: Output: database:default
-PREHOOK: Output: default@lineitem
-POSTHOOK: query: CREATE TABLE lineitem (L_ORDERKEY INT,
- L_PARTKEY INT,
- L_SUPPKEY INT,
- L_LINENUMBER INT,
- L_QUANTITY DOUBLE,
- L_EXTENDEDPRICE DOUBLE,
- L_DISCOUNT DOUBLE,
- L_TAX DOUBLE,
- L_RETURNFLAG STRING,
- L_LINESTATUS STRING,
- l_shipdate STRING,
- L_COMMITDATE STRING,
- L_RECEIPTDATE STRING,
- L_SHIPINSTRUCT STRING,
- L_SHIPMODE STRING,
- L_COMMENT STRING)
-ROW FORMAT DELIMITED
-FIELDS TERMINATED BY '|'
-POSTHOOK: type: CREATETABLE
-POSTHOOK: Output: database:default
-POSTHOOK: Output: default@lineitem
PREHOOK: query: -- non agg, non corr
explain rewrite
select *
Modified: hive/branches/spark/ql/src/test/results/clientpositive/subquery_in_having.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/subquery_in_having.q.out?rev=1637444&r1=1637443&r2=1637444&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/subquery_in_having.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/subquery_in_having.q.out Fri Nov 7 20:41:34 2014
@@ -1,7 +1,14 @@
PREHOOK: query: -- SORT_QUERY_RESULTS
-- data setup
-CREATE TABLE part(
+DROP TABLE IF EXISTS part_subq
+PREHOOK: type: DROPTABLE
+POSTHOOK: query: -- SORT_QUERY_RESULTS
+
+-- data setup
+DROP TABLE IF EXISTS part_subq
+POSTHOOK: type: DROPTABLE
+PREHOOK: query: CREATE TABLE part_subq(
p_partkey INT,
p_name STRING,
p_mfgr STRING,
@@ -14,11 +21,8 @@ CREATE TABLE part(
)
PREHOOK: type: CREATETABLE
PREHOOK: Output: database:default
-PREHOOK: Output: default@part
-POSTHOOK: query: -- SORT_QUERY_RESULTS
-
--- data setup
-CREATE TABLE part(
+PREHOOK: Output: default@part_subq
+POSTHOOK: query: CREATE TABLE part_subq(
p_partkey INT,
p_name STRING,
p_mfgr STRING,
@@ -31,15 +35,15 @@ CREATE TABLE part(
)
POSTHOOK: type: CREATETABLE
POSTHOOK: Output: database:default
-POSTHOOK: Output: default@part
-PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/part_tiny.txt' overwrite into table part
+POSTHOOK: Output: default@part_subq
+PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/part_tiny.txt' overwrite into table part_subq
PREHOOK: type: LOAD
#### A masked pattern was here ####
-PREHOOK: Output: default@part
-POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/part_tiny.txt' overwrite into table part
+PREHOOK: Output: default@part_subq
+POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/part_tiny.txt' overwrite into table part_subq
POSTHOOK: type: LOAD
#### A masked pattern was here ####
-POSTHOOK: Output: default@part
+POSTHOOK: Output: default@part_subq
PREHOOK: query: -- non agg, non corr
explain
select key, count(*)
@@ -413,11 +417,11 @@ STAGE PLANS:
PREHOOK: query: -- agg, non corr
explain
select p_mfgr, avg(p_size)
-from part b
+from part_subq b
group by b.p_mfgr
having b.p_mfgr in
(select p_mfgr
- from part
+ from part_subq
group by p_mfgr
having max(p_size) - min(p_size) < 20
)
@@ -425,11 +429,11 @@ PREHOOK: type: QUERY
POSTHOOK: query: -- agg, non corr
explain
select p_mfgr, avg(p_size)
-from part b
+from part_subq b
group by b.p_mfgr
having b.p_mfgr in
(select p_mfgr
- from part
+ from part_subq
group by p_mfgr
having max(p_size) - min(p_size) < 20
)
@@ -445,7 +449,7 @@ STAGE PLANS:
Map Reduce
Map Operator Tree:
TableScan
- alias: b
+ alias: part_subq
Statistics: Num rows: 30 Data size: 3173 Basic stats: COMPLETE Column stats: NONE
Filter Operator
predicate: p_mfgr is not null (type: boolean)
@@ -455,30 +459,42 @@ STAGE PLANS:
outputColumnNames: p_mfgr, p_size
Statistics: Num rows: 15 Data size: 1586 Basic stats: COMPLETE Column stats: NONE
Group By Operator
- aggregations: avg(p_size)
+ aggregations: max(p_size), min(p_size)
keys: p_mfgr (type: string)
mode: hash
- outputColumnNames: _col0, _col1
+ outputColumnNames: _col0, _col1, _col2
Statistics: Num rows: 15 Data size: 1586 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
key expressions: _col0 (type: string)
sort order: +
Map-reduce partition columns: _col0 (type: string)
Statistics: Num rows: 15 Data size: 1586 Basic stats: COMPLETE Column stats: NONE
- value expressions: _col1 (type: struct<count:bigint,sum:double,input:int>)
+ value expressions: _col1 (type: int), _col2 (type: int)
Reduce Operator Tree:
Group By Operator
- aggregations: avg(VALUE._col0)
+ aggregations: max(VALUE._col0), min(VALUE._col1)
keys: KEY._col0 (type: string)
mode: mergepartial
- outputColumnNames: _col0, _col1
- Statistics: Num rows: 15 Data size: 1586 Basic stats: COMPLETE Column stats: NONE
- File Output Operator
- compressed: false
- table:
- input format: org.apache.hadoop.mapred.SequenceFileInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
- serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe
+ outputColumnNames: _col0, _col1, _col2
+ Statistics: Num rows: 7 Data size: 740 Basic stats: COMPLETE Column stats: NONE
+ Filter Operator
+ predicate: ((_col1 - _col2) < 20) (type: boolean)
+ Statistics: Num rows: 2 Data size: 211 Basic stats: COMPLETE Column stats: NONE
+ Select Operator
+ expressions: _col0 (type: string)
+ outputColumnNames: _col0
+ Statistics: Num rows: 2 Data size: 211 Basic stats: COMPLETE Column stats: NONE
+ Group By Operator
+ keys: _col0 (type: string)
+ mode: hash
+ outputColumnNames: _col0
+ Statistics: Num rows: 2 Data size: 211 Basic stats: COMPLETE Column stats: NONE
+ File Output Operator
+ compressed: false
+ table:
+ input format: org.apache.hadoop.mapred.SequenceFileInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe
Stage: Stage-2
Map Reduce
@@ -488,14 +504,14 @@ STAGE PLANS:
key expressions: _col0 (type: string)
sort order: +
Map-reduce partition columns: _col0 (type: string)
- Statistics: Num rows: 15 Data size: 1586 Basic stats: COMPLETE Column stats: NONE
- value expressions: _col1 (type: double)
+ Statistics: Num rows: 2 Data size: 211 Basic stats: COMPLETE Column stats: NONE
TableScan
Reduce Output Operator
key expressions: _col0 (type: string)
sort order: +
Map-reduce partition columns: _col0 (type: string)
- Statistics: Num rows: 2 Data size: 211 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 15 Data size: 1586 Basic stats: COMPLETE Column stats: NONE
+ value expressions: _col1 (type: double)
Reduce Operator Tree:
Join Operator
condition map:
@@ -521,7 +537,7 @@ STAGE PLANS:
Map Reduce
Map Operator Tree:
TableScan
- alias: part
+ alias: b
Statistics: Num rows: 30 Data size: 3173 Basic stats: COMPLETE Column stats: NONE
Filter Operator
predicate: p_mfgr is not null (type: boolean)
@@ -531,42 +547,30 @@ STAGE PLANS:
outputColumnNames: p_mfgr, p_size
Statistics: Num rows: 15 Data size: 1586 Basic stats: COMPLETE Column stats: NONE
Group By Operator
- aggregations: max(p_size), min(p_size)
+ aggregations: avg(p_size)
keys: p_mfgr (type: string)
mode: hash
- outputColumnNames: _col0, _col1, _col2
+ outputColumnNames: _col0, _col1
Statistics: Num rows: 15 Data size: 1586 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
key expressions: _col0 (type: string)
sort order: +
Map-reduce partition columns: _col0 (type: string)
Statistics: Num rows: 15 Data size: 1586 Basic stats: COMPLETE Column stats: NONE
- value expressions: _col1 (type: int), _col2 (type: int)
+ value expressions: _col1 (type: struct<count:bigint,sum:double,input:int>)
Reduce Operator Tree:
Group By Operator
- aggregations: max(VALUE._col0), min(VALUE._col1)
+ aggregations: avg(VALUE._col0)
keys: KEY._col0 (type: string)
mode: mergepartial
- outputColumnNames: _col0, _col1, _col2
- Statistics: Num rows: 7 Data size: 740 Basic stats: COMPLETE Column stats: NONE
- Filter Operator
- predicate: ((_col1 - _col2) < 20) (type: boolean)
- Statistics: Num rows: 2 Data size: 211 Basic stats: COMPLETE Column stats: NONE
- Select Operator
- expressions: _col0 (type: string)
- outputColumnNames: _col0
- Statistics: Num rows: 2 Data size: 211 Basic stats: COMPLETE Column stats: NONE
- Group By Operator
- keys: _col0 (type: string)
- mode: hash
- outputColumnNames: _col0
- Statistics: Num rows: 2 Data size: 211 Basic stats: COMPLETE Column stats: NONE
- File Output Operator
- compressed: false
- table:
- input format: org.apache.hadoop.mapred.SequenceFileInputFormat
- output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
- serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe
+ outputColumnNames: _col0, _col1
+ Statistics: Num rows: 15 Data size: 1586 Basic stats: COMPLETE Column stats: NONE
+ File Output Operator
+ compressed: false
+ table:
+ input format: org.apache.hadoop.mapred.SequenceFileInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe
Stage: Stage-0
Fetch Operator
@@ -577,11 +581,11 @@ STAGE PLANS:
PREHOOK: query: -- agg, non corr
explain
select p_mfgr, avg(p_size)
-from part b
+from part_subq b
group by b.p_mfgr
having b.p_mfgr in
(select p_mfgr
- from part
+ from part_subq
group by p_mfgr
having max(p_size) - min(p_size) < 20
)
@@ -589,11 +593,11 @@ PREHOOK: type: QUERY
POSTHOOK: query: -- agg, non corr
explain
select p_mfgr, avg(p_size)
-from part b
+from part_subq b
group by b.p_mfgr
having b.p_mfgr in
(select p_mfgr
- from part
+ from part_subq
group by p_mfgr
having max(p_size) - min(p_size) < 20
)
@@ -607,7 +611,7 @@ STAGE PLANS:
Map Reduce
Map Operator Tree:
TableScan
- alias: b
+ alias: part_subq
Statistics: Num rows: 30 Data size: 3173 Basic stats: COMPLETE Column stats: NONE
Filter Operator
predicate: p_mfgr is not null (type: boolean)
@@ -617,19 +621,19 @@ STAGE PLANS:
outputColumnNames: p_mfgr, p_size
Statistics: Num rows: 15 Data size: 1586 Basic stats: COMPLETE Column stats: NONE
Group By Operator
- aggregations: avg(p_size)
+ aggregations: max(p_size), min(p_size)
keys: p_mfgr (type: string)
mode: hash
- outputColumnNames: _col0, _col1
+ outputColumnNames: _col0, _col1, _col2
Statistics: Num rows: 15 Data size: 1586 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
key expressions: _col0 (type: string)
sort order: +
Map-reduce partition columns: _col0 (type: string)
Statistics: Num rows: 15 Data size: 1586 Basic stats: COMPLETE Column stats: NONE
- value expressions: _col1 (type: struct<count:bigint,sum:double,input:int>)
+ value expressions: _col1 (type: int), _col2 (type: int)
TableScan
- alias: part
+ alias: b
Statistics: Num rows: 30 Data size: 3173 Basic stats: COMPLETE Column stats: NONE
Filter Operator
predicate: p_mfgr is not null (type: boolean)
@@ -639,17 +643,17 @@ STAGE PLANS:
outputColumnNames: p_mfgr, p_size
Statistics: Num rows: 15 Data size: 1586 Basic stats: COMPLETE Column stats: NONE
Group By Operator
- aggregations: max(p_size), min(p_size)
+ aggregations: avg(p_size)
keys: p_mfgr (type: string)
mode: hash
- outputColumnNames: _col0, _col1, _col2
+ outputColumnNames: _col0, _col1
Statistics: Num rows: 15 Data size: 1586 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
key expressions: _col0 (type: string)
sort order: +
Map-reduce partition columns: _col0 (type: string)
Statistics: Num rows: 15 Data size: 1586 Basic stats: COMPLETE Column stats: NONE
- value expressions: _col1 (type: int), _col2 (type: int)
+ value expressions: _col1 (type: struct<count:bigint,sum:double,input:int>)
Reduce Operator Tree:
Demux Operator
Statistics: Num rows: 30 Data size: 3172 Basic stats: COMPLETE Column stats: NONE
@@ -1233,18 +1237,18 @@ STAGE PLANS:
PREHOOK: query: -- non agg, non corr, windowing
explain
select p_mfgr, p_name, avg(p_size)
-from part
+from part_subq
group by p_mfgr, p_name
having p_name in
- (select first_value(p_name) over(partition by p_mfgr order by p_size) from part)
+ (select first_value(p_name) over(partition by p_mfgr order by p_size) from part_subq)
PREHOOK: type: QUERY
POSTHOOK: query: -- non agg, non corr, windowing
explain
select p_mfgr, p_name, avg(p_size)
-from part
+from part_subq
group by p_mfgr, p_name
having p_name in
- (select first_value(p_name) over(partition by p_mfgr order by p_size) from part)
+ (select first_value(p_name) over(partition by p_mfgr order by p_size) from part_subq)
POSTHOOK: type: QUERY
STAGE DEPENDENCIES:
Stage-1 is a root stage
@@ -1260,7 +1264,7 @@ STAGE PLANS:
Map Reduce
Map Operator Tree:
TableScan
- alias: part
+ alias: part_subq
Statistics: Num rows: 15 Data size: 3173 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
key expressions: p_mfgr (type: string), p_size (type: int)
@@ -1379,7 +1383,7 @@ STAGE PLANS:
Map Reduce
Map Operator Tree:
TableScan
- alias: part
+ alias: part_subq
Statistics: Num rows: 15 Data size: 3173 Basic stats: COMPLETE Column stats: NONE
Filter Operator
predicate: p_name is not null (type: boolean)
@@ -1420,3 +1424,11 @@ STAGE PLANS:
Processor Tree:
ListSink
+PREHOOK: query: DROP TABLE part_subq
+PREHOOK: type: DROPTABLE
+PREHOOK: Input: default@part_subq
+PREHOOK: Output: default@part_subq
+POSTHOOK: query: DROP TABLE part_subq
+POSTHOOK: type: DROPTABLE
+POSTHOOK: Input: default@part_subq
+POSTHOOK: Output: default@part_subq