You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hive.apache.org by br...@apache.org on 2014/11/07 21:42:45 UTC
svn commit: r1637445 [4/7] -
/hive/branches/spark/ql/src/test/results/clientpositive/spark/
Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/ptf_decimal.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/ptf_decimal.q.out?rev=1637445&r1=1637444&r2=1637445&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/spark/ptf_decimal.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/spark/ptf_decimal.q.out Fri Nov 7 20:42:43 2014
@@ -1,45 +1,3 @@
-PREHOOK: query: DROP TABLE IF EXISTS part
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: DROP TABLE IF EXISTS part
-POSTHOOK: type: DROPTABLE
-PREHOOK: query: -- data setup
-CREATE TABLE part(
- p_partkey INT,
- p_name STRING,
- p_mfgr STRING,
- p_brand STRING,
- p_type STRING,
- p_size INT,
- p_container STRING,
- p_retailprice DECIMAL(6,2),
- p_comment STRING
-)
-PREHOOK: type: CREATETABLE
-PREHOOK: Output: database:default
-PREHOOK: Output: default@part
-POSTHOOK: query: -- data setup
-CREATE TABLE part(
- p_partkey INT,
- p_name STRING,
- p_mfgr STRING,
- p_brand STRING,
- p_type STRING,
- p_size INT,
- p_container STRING,
- p_retailprice DECIMAL(6,2),
- p_comment STRING
-)
-POSTHOOK: type: CREATETABLE
-POSTHOOK: Output: database:default
-POSTHOOK: Output: default@part
-PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/part_tiny.txt' overwrite into table part
-PREHOOK: type: LOAD
-#### A masked pattern was here ####
-PREHOOK: Output: default@part
-POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/part_tiny.txt' overwrite into table part
-POSTHOOK: type: LOAD
-#### A masked pattern was here ####
-POSTHOOK: Output: default@part
PREHOOK: query: -- 1. aggregate functions with decimal type
select p_mfgr, p_retailprice,
Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/ptf_general_queries.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/ptf_general_queries.q.out?rev=1637445&r1=1637444&r2=1637445&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/spark/ptf_general_queries.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/spark/ptf_general_queries.q.out Fri Nov 7 20:42:43 2014
@@ -1,45 +1,3 @@
-PREHOOK: query: DROP TABLE part
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: DROP TABLE part
-POSTHOOK: type: DROPTABLE
-PREHOOK: query: -- data setup
-CREATE TABLE part(
- p_partkey INT,
- p_name STRING,
- p_mfgr STRING,
- p_brand STRING,
- p_type STRING,
- p_size INT,
- p_container STRING,
- p_retailprice DOUBLE,
- p_comment STRING
-)
-PREHOOK: type: CREATETABLE
-PREHOOK: Output: database:default
-PREHOOK: Output: default@part
-POSTHOOK: query: -- data setup
-CREATE TABLE part(
- p_partkey INT,
- p_name STRING,
- p_mfgr STRING,
- p_brand STRING,
- p_type STRING,
- p_size INT,
- p_container STRING,
- p_retailprice DOUBLE,
- p_comment STRING
-)
-POSTHOOK: type: CREATETABLE
-POSTHOOK: Output: database:default
-POSTHOOK: Output: default@part
-PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/part_tiny.txt' overwrite into table part
-PREHOOK: type: LOAD
-#### A masked pattern was here ####
-PREHOOK: Output: default@part
-POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/part_tiny.txt' overwrite into table part
-POSTHOOK: type: LOAD
-#### A masked pattern was here ####
-POSTHOOK: Output: default@part
PREHOOK: query: -- 1. testNoPTFNoWindowing
select p_mfgr, p_name, p_size
from part
Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/ptf_streaming.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/ptf_streaming.q.out?rev=1637445&r1=1637444&r2=1637445&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/spark/ptf_streaming.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/spark/ptf_streaming.q.out Fri Nov 7 20:42:43 2014
@@ -1,45 +1,3 @@
-PREHOOK: query: DROP TABLE part
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: DROP TABLE part
-POSTHOOK: type: DROPTABLE
-PREHOOK: query: -- data setup
-CREATE TABLE part(
- p_partkey INT,
- p_name STRING,
- p_mfgr STRING,
- p_brand STRING,
- p_type STRING,
- p_size INT,
- p_container STRING,
- p_retailprice DOUBLE,
- p_comment STRING
-)
-PREHOOK: type: CREATETABLE
-PREHOOK: Output: database:default
-PREHOOK: Output: default@part
-POSTHOOK: query: -- data setup
-CREATE TABLE part(
- p_partkey INT,
- p_name STRING,
- p_mfgr STRING,
- p_brand STRING,
- p_type STRING,
- p_size INT,
- p_container STRING,
- p_retailprice DOUBLE,
- p_comment STRING
-)
-POSTHOOK: type: CREATETABLE
-POSTHOOK: Output: database:default
-POSTHOOK: Output: default@part
-PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/part_tiny.txt' overwrite into table part
-PREHOOK: type: LOAD
-#### A masked pattern was here ####
-PREHOOK: Output: default@part
-POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/part_tiny.txt' overwrite into table part
-POSTHOOK: type: LOAD
-#### A masked pattern was here ####
-POSTHOOK: Output: default@part
PREHOOK: query: create temporary function noopstreaming as 'org.apache.hadoop.hive.ql.udf.ptf.NoopStreaming$NoopStreamingResolver'
PREHOOK: type: CREATEFUNCTION
PREHOOK: Output: noopstreaming
Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/subquery_in.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/subquery_in.q.out?rev=1637445&r1=1637444&r2=1637445&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/spark/subquery_in.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/spark/subquery_in.q.out Fri Nov 7 20:42:43 2014
@@ -1,110 +1,14 @@
PREHOOK: query: -- SORT_QUERY_RESULTS
-DROP TABLE part
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: -- SORT_QUERY_RESULTS
-
-DROP TABLE part
-POSTHOOK: type: DROPTABLE
-PREHOOK: query: -- data setup
-CREATE TABLE part(
- p_partkey INT,
- p_name STRING,
- p_mfgr STRING,
- p_brand STRING,
- p_type STRING,
- p_size INT,
- p_container STRING,
- p_retailprice DOUBLE,
- p_comment STRING
-)
-PREHOOK: type: CREATETABLE
-PREHOOK: Output: database:default
-PREHOOK: Output: default@part
-POSTHOOK: query: -- data setup
-CREATE TABLE part(
- p_partkey INT,
- p_name STRING,
- p_mfgr STRING,
- p_brand STRING,
- p_type STRING,
- p_size INT,
- p_container STRING,
- p_retailprice DOUBLE,
- p_comment STRING
-)
-POSTHOOK: type: CREATETABLE
-POSTHOOK: Output: database:default
-POSTHOOK: Output: default@part
-PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/part_tiny.txt' overwrite into table part
-PREHOOK: type: LOAD
-#### A masked pattern was here ####
-PREHOOK: Output: default@part
-POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/part_tiny.txt' overwrite into table part
-POSTHOOK: type: LOAD
-#### A masked pattern was here ####
-POSTHOOK: Output: default@part
-PREHOOK: query: DROP TABLE lineitem
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: DROP TABLE lineitem
-POSTHOOK: type: DROPTABLE
-PREHOOK: query: CREATE TABLE lineitem (L_ORDERKEY INT,
- L_PARTKEY INT,
- L_SUPPKEY INT,
- L_LINENUMBER INT,
- L_QUANTITY DOUBLE,
- L_EXTENDEDPRICE DOUBLE,
- L_DISCOUNT DOUBLE,
- L_TAX DOUBLE,
- L_RETURNFLAG STRING,
- L_LINESTATUS STRING,
- l_shipdate STRING,
- L_COMMITDATE STRING,
- L_RECEIPTDATE STRING,
- L_SHIPINSTRUCT STRING,
- L_SHIPMODE STRING,
- L_COMMENT STRING)
-ROW FORMAT DELIMITED
-FIELDS TERMINATED BY '|'
-PREHOOK: type: CREATETABLE
-PREHOOK: Output: database:default
-PREHOOK: Output: default@lineitem
-POSTHOOK: query: CREATE TABLE lineitem (L_ORDERKEY INT,
- L_PARTKEY INT,
- L_SUPPKEY INT,
- L_LINENUMBER INT,
- L_QUANTITY DOUBLE,
- L_EXTENDEDPRICE DOUBLE,
- L_DISCOUNT DOUBLE,
- L_TAX DOUBLE,
- L_RETURNFLAG STRING,
- L_LINESTATUS STRING,
- l_shipdate STRING,
- L_COMMITDATE STRING,
- L_RECEIPTDATE STRING,
- L_SHIPINSTRUCT STRING,
- L_SHIPMODE STRING,
- L_COMMENT STRING)
-ROW FORMAT DELIMITED
-FIELDS TERMINATED BY '|'
-POSTHOOK: type: CREATETABLE
-POSTHOOK: Output: database:default
-POSTHOOK: Output: default@lineitem
-PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/lineitem.txt' OVERWRITE INTO TABLE lineitem
-PREHOOK: type: LOAD
-#### A masked pattern was here ####
-PREHOOK: Output: default@lineitem
-POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/lineitem.txt' OVERWRITE INTO TABLE lineitem
-POSTHOOK: type: LOAD
-#### A masked pattern was here ####
-POSTHOOK: Output: default@lineitem
-PREHOOK: query: -- non agg, non corr
+-- non agg, non corr
explain
select *
from src
where src.key in (select key from src s1 where s1.key > '9')
PREHOOK: type: QUERY
-POSTHOOK: query: -- non agg, non corr
+POSTHOOK: query: -- SORT_QUERY_RESULTS
+
+-- non agg, non corr
explain
select *
from src
@@ -370,40 +274,40 @@ STAGE PLANS:
Map Operator Tree:
TableScan
alias: part
- Statistics: Num rows: 30 Data size: 3173 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
key expressions: p_mfgr (type: string), p_size (type: int)
sort order: ++
Map-reduce partition columns: p_mfgr (type: string)
- Statistics: Num rows: 30 Data size: 3173 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
value expressions: p_mfgr (type: string), p_size (type: int)
Map 5
Map Operator Tree:
TableScan
alias: part
- Statistics: Num rows: 30 Data size: 3173 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
Filter Operator
predicate: UDFToDouble(p_size) is not null (type: boolean)
- Statistics: Num rows: 15 Data size: 1586 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 13 Data size: 1573 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
key expressions: UDFToDouble(p_size) (type: double)
sort order: +
Map-reduce partition columns: UDFToDouble(p_size) (type: double)
- Statistics: Num rows: 15 Data size: 1586 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 13 Data size: 1573 Basic stats: COMPLETE Column stats: NONE
value expressions: p_name (type: string), p_size (type: int)
Reducer 2
Reduce Operator Tree:
Extract
- Statistics: Num rows: 30 Data size: 3173 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
PTF Operator
- Statistics: Num rows: 30 Data size: 3173 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
Filter Operator
predicate: (_wcol0 <= 2) (type: boolean)
- Statistics: Num rows: 10 Data size: 1057 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 8 Data size: 968 Basic stats: COMPLETE Column stats: NONE
Select Operator
expressions: _col5 (type: int)
outputColumnNames: _col0
- Statistics: Num rows: 10 Data size: 1057 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 8 Data size: 968 Basic stats: COMPLETE Column stats: NONE
Group By Operator
aggregations: avg(_col0)
mode: hash
@@ -446,14 +350,14 @@ STAGE PLANS:
0 {VALUE._col1} {VALUE._col5}
1
outputColumnNames: _col1, _col5
- Statistics: Num rows: 16 Data size: 1744 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 14 Data size: 1730 Basic stats: COMPLETE Column stats: NONE
Select Operator
expressions: _col1 (type: string), _col5 (type: int)
outputColumnNames: _col0, _col1
- Statistics: Num rows: 16 Data size: 1744 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 14 Data size: 1730 Basic stats: COMPLETE Column stats: NONE
File Output Operator
compressed: false
- Statistics: Num rows: 16 Data size: 1744 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 14 Data size: 1730 Basic stats: COMPLETE Column stats: NONE
table:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
@@ -522,26 +426,26 @@ STAGE PLANS:
Map Operator Tree:
TableScan
alias: b
- Statistics: Num rows: 15 Data size: 3173 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
Filter Operator
predicate: (p_size is not null and p_mfgr is not null) (type: boolean)
- Statistics: Num rows: 4 Data size: 846 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 7 Data size: 847 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
key expressions: p_size (type: int), p_mfgr (type: string)
sort order: ++
Map-reduce partition columns: p_size (type: int), p_mfgr (type: string)
- Statistics: Num rows: 4 Data size: 846 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 7 Data size: 847 Basic stats: COMPLETE Column stats: NONE
value expressions: p_name (type: string)
Map 3
Map Operator Tree:
TableScan
alias: part
- Statistics: Num rows: 30 Data size: 3173 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
key expressions: p_mfgr (type: string), p_size (type: int)
sort order: ++
Map-reduce partition columns: p_mfgr (type: string)
- Statistics: Num rows: 30 Data size: 3173 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
value expressions: p_mfgr (type: string), p_size (type: int)
Reducer 2
Reduce Operator Tree:
@@ -552,14 +456,14 @@ STAGE PLANS:
0 {VALUE._col1} {KEY.reducesinkkey1} {KEY.reducesinkkey0}
1
outputColumnNames: _col1, _col2, _col5
- Statistics: Num rows: 4 Data size: 930 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 7 Data size: 931 Basic stats: COMPLETE Column stats: NONE
Select Operator
expressions: _col2 (type: string), _col1 (type: string), _col5 (type: int)
outputColumnNames: _col0, _col1, _col2
- Statistics: Num rows: 4 Data size: 930 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 7 Data size: 931 Basic stats: COMPLETE Column stats: NONE
File Output Operator
compressed: false
- Statistics: Num rows: 4 Data size: 930 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 7 Data size: 931 Basic stats: COMPLETE Column stats: NONE
table:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
@@ -567,27 +471,27 @@ STAGE PLANS:
Reducer 4
Reduce Operator Tree:
Extract
- Statistics: Num rows: 30 Data size: 3173 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
PTF Operator
- Statistics: Num rows: 30 Data size: 3173 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE
Filter Operator
predicate: ((_wcol0 <= 2) and _col2 is not null) (type: boolean)
- Statistics: Num rows: 5 Data size: 528 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 4 Data size: 484 Basic stats: COMPLETE Column stats: NONE
Select Operator
expressions: _col2 (type: string), _col5 (type: int)
outputColumnNames: _col0, _col1
- Statistics: Num rows: 5 Data size: 528 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 4 Data size: 484 Basic stats: COMPLETE Column stats: NONE
Group By Operator
aggregations: min(_col1)
keys: _col0 (type: string)
mode: hash
outputColumnNames: _col0, _col1
- Statistics: Num rows: 5 Data size: 528 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 4 Data size: 484 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
key expressions: _col0 (type: string)
sort order: +
Map-reduce partition columns: _col0 (type: string)
- Statistics: Num rows: 5 Data size: 528 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 4 Data size: 484 Basic stats: COMPLETE Column stats: NONE
value expressions: _col1 (type: int)
Reducer 5
Reduce Operator Tree:
@@ -596,24 +500,24 @@ STAGE PLANS:
keys: KEY._col0 (type: string)
mode: mergepartial
outputColumnNames: _col0, _col1
- Statistics: Num rows: 2 Data size: 211 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 2 Data size: 242 Basic stats: COMPLETE Column stats: NONE
Filter Operator
predicate: _col1 is not null (type: boolean)
- Statistics: Num rows: 2 Data size: 211 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 2 Data size: 242 Basic stats: COMPLETE Column stats: NONE
Select Operator
expressions: _col1 (type: int), _col0 (type: string)
outputColumnNames: _col0, _col1
- Statistics: Num rows: 2 Data size: 211 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 2 Data size: 242 Basic stats: COMPLETE Column stats: NONE
Group By Operator
keys: _col0 (type: int), _col1 (type: string)
mode: hash
outputColumnNames: _col0, _col1
- Statistics: Num rows: 2 Data size: 211 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 2 Data size: 242 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
key expressions: _col0 (type: int), _col1 (type: string)
sort order: ++
Map-reduce partition columns: _col0 (type: int), _col1 (type: string)
- Statistics: Num rows: 2 Data size: 211 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 2 Data size: 242 Basic stats: COMPLETE Column stats: NONE
Stage: Stage-0
Fetch Operator
@@ -845,60 +749,60 @@ STAGE PLANS:
Map Operator Tree:
TableScan
alias: li
- Statistics: Num rows: 756 Data size: 12099 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 100 Data size: 11999 Basic stats: COMPLETE Column stats: NONE
Filter Operator
predicate: ((l_partkey is not null and l_orderkey is not null) and (l_linenumber = 1)) (type: boolean)
- Statistics: Num rows: 94 Data size: 1504 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 12 Data size: 1439 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
key expressions: l_partkey (type: int)
sort order: +
Map-reduce partition columns: l_partkey (type: int)
- Statistics: Num rows: 94 Data size: 1504 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 12 Data size: 1439 Basic stats: COMPLETE Column stats: NONE
value expressions: l_orderkey (type: int), l_suppkey (type: int)
Map 4
Map Operator Tree:
TableScan
alias: lineitem
- Statistics: Num rows: 3024 Data size: 12099 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 100 Data size: 11999 Basic stats: COMPLETE Column stats: NONE
Filter Operator
predicate: l_partkey is not null (type: boolean)
- Statistics: Num rows: 1512 Data size: 6049 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 50 Data size: 5999 Basic stats: COMPLETE Column stats: NONE
Select Operator
expressions: l_partkey (type: int)
outputColumnNames: l_partkey
- Statistics: Num rows: 1512 Data size: 6049 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 50 Data size: 5999 Basic stats: COMPLETE Column stats: NONE
Group By Operator
keys: l_partkey (type: int)
mode: hash
outputColumnNames: _col0
- Statistics: Num rows: 1512 Data size: 6049 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 50 Data size: 5999 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
key expressions: _col0 (type: int)
sort order: +
Map-reduce partition columns: _col0 (type: int)
- Statistics: Num rows: 1512 Data size: 6049 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 50 Data size: 5999 Basic stats: COMPLETE Column stats: NONE
Map 6
Map Operator Tree:
TableScan
alias: lineitem
- Statistics: Num rows: 1728 Data size: 12099 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 100 Data size: 11999 Basic stats: COMPLETE Column stats: NONE
Filter Operator
predicate: ((l_shipmode = 'AIR') and l_orderkey is not null) (type: boolean)
- Statistics: Num rows: 432 Data size: 3024 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 25 Data size: 2999 Basic stats: COMPLETE Column stats: NONE
Select Operator
expressions: l_orderkey (type: int)
outputColumnNames: _col0
- Statistics: Num rows: 432 Data size: 3024 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 25 Data size: 2999 Basic stats: COMPLETE Column stats: NONE
Group By Operator
keys: _col0 (type: int)
mode: hash
outputColumnNames: _col0
- Statistics: Num rows: 432 Data size: 3024 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 25 Data size: 2999 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
key expressions: _col0 (type: int)
sort order: +
Map-reduce partition columns: _col0 (type: int)
- Statistics: Num rows: 432 Data size: 3024 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 25 Data size: 2999 Basic stats: COMPLETE Column stats: NONE
Reducer 2
Reduce Operator Tree:
Join Operator
@@ -908,12 +812,12 @@ STAGE PLANS:
0 {KEY.reducesinkkey0}
1 {VALUE._col0} {VALUE._col1}
outputColumnNames: _col0, _col1, _col3
- Statistics: Num rows: 831 Data size: 3326 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 27 Data size: 3298 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
key expressions: _col1 (type: int)
sort order: +
Map-reduce partition columns: _col1 (type: int)
- Statistics: Num rows: 831 Data size: 3326 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 27 Data size: 3298 Basic stats: COMPLETE Column stats: NONE
value expressions: _col0 (type: int), _col3 (type: int)
Reducer 3
Reduce Operator Tree:
@@ -924,14 +828,14 @@ STAGE PLANS:
0 {VALUE._col0} {VALUE._col2}
1
outputColumnNames: _col0, _col3
- Statistics: Num rows: 914 Data size: 3658 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 29 Data size: 3627 Basic stats: COMPLETE Column stats: NONE
Select Operator
expressions: _col0 (type: int), _col3 (type: int)
outputColumnNames: _col0, _col1
- Statistics: Num rows: 914 Data size: 3658 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 29 Data size: 3627 Basic stats: COMPLETE Column stats: NONE
File Output Operator
compressed: false
- Statistics: Num rows: 914 Data size: 3658 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 29 Data size: 3627 Basic stats: COMPLETE Column stats: NONE
table:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
@@ -942,16 +846,16 @@ STAGE PLANS:
keys: KEY._col0 (type: int)
mode: mergepartial
outputColumnNames: _col0
- Statistics: Num rows: 756 Data size: 3024 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 25 Data size: 2999 Basic stats: COMPLETE Column stats: NONE
Select Operator
expressions: _col0 (type: int)
outputColumnNames: _col0
- Statistics: Num rows: 756 Data size: 3024 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 25 Data size: 2999 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
key expressions: _col0 (type: int)
sort order: +
Map-reduce partition columns: _col0 (type: int)
- Statistics: Num rows: 756 Data size: 3024 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 25 Data size: 2999 Basic stats: COMPLETE Column stats: NONE
Stage: Stage-0
Fetch Operator
Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/temp_table.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/temp_table.q.out?rev=1637445&r1=1637444&r2=1637445&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/spark/temp_table.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/spark/temp_table.q.out Fri Nov 7 20:42:43 2014
@@ -355,9 +355,15 @@ alltypesorc
bar
bay
baz
+cbo_t1
+cbo_t2
+cbo_t3
foo
+lineitem
+part
src
src1
+src_cbo
src_json
src_sequencefile
src_thrift
Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/union11.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/union11.q.out?rev=1637445&r1=1637444&r2=1637445&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/spark/union11.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/spark/union11.q.out Fri Nov 7 20:42:43 2014
@@ -110,14 +110,14 @@ STAGE PLANS:
keys: KEY._col0 (type: string)
mode: mergepartial
outputColumnNames: _col0, _col1
- Statistics: Num rows: 1 Data size: 100 Basic stats: COMPLETE Column stats: COMPLETE
+ Statistics: Num rows: 1 Data size: 96 Basic stats: COMPLETE Column stats: COMPLETE
Select Operator
expressions: _col0 (type: string), _col1 (type: bigint)
outputColumnNames: _col0, _col1
- Statistics: Num rows: 1 Data size: 100 Basic stats: COMPLETE Column stats: COMPLETE
+ Statistics: Num rows: 1 Data size: 96 Basic stats: COMPLETE Column stats: COMPLETE
File Output Operator
compressed: false
- Statistics: Num rows: 1 Data size: 100 Basic stats: COMPLETE Column stats: COMPLETE
+ Statistics: Num rows: 1 Data size: 96 Basic stats: COMPLETE Column stats: COMPLETE
table:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
@@ -191,6 +191,6 @@ POSTHOOK: query: select unionsrc.key, co
POSTHOOK: type: QUERY
POSTHOOK: Input: default@src
#### A masked pattern was here ####
-tst3 1
-tst1 1
tst2 1
+tst1 1
+tst3 1
Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/union14.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/union14.q.out?rev=1637445&r1=1637444&r2=1637445&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/spark/union14.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/spark/union14.q.out Fri Nov 7 20:42:43 2014
@@ -72,14 +72,14 @@ STAGE PLANS:
keys: KEY._col0 (type: string)
mode: mergepartial
outputColumnNames: _col0, _col1
- Statistics: Num rows: 1 Data size: 100 Basic stats: COMPLETE Column stats: COMPLETE
+ Statistics: Num rows: 1 Data size: 96 Basic stats: COMPLETE Column stats: COMPLETE
Select Operator
expressions: _col0 (type: string), _col1 (type: bigint)
outputColumnNames: _col0, _col1
- Statistics: Num rows: 1 Data size: 100 Basic stats: COMPLETE Column stats: COMPLETE
+ Statistics: Num rows: 1 Data size: 96 Basic stats: COMPLETE Column stats: COMPLETE
File Output Operator
compressed: false
- Statistics: Num rows: 1 Data size: 100 Basic stats: COMPLETE Column stats: COMPLETE
+ Statistics: Num rows: 1 Data size: 96 Basic stats: COMPLETE Column stats: COMPLETE
table:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
@@ -131,20 +131,20 @@ POSTHOOK: type: QUERY
POSTHOOK: Input: default@src
POSTHOOK: Input: default@src1
#### A masked pattern was here ####
-278 1
-273 1
-128 1
255 1
-tst1 1
+213 1
238 1
-146 1
150 1
- 10
-406 1
-369 1
-213 1
311 1
-224 1
+146 1
+369 1
66 1
+ 10
+278 1
401 1
+406 1
+tst1 1
+128 1
98 1
+273 1
+224 1
Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/union15.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/union15.q.out?rev=1637445&r1=1637444&r2=1637445&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/spark/union15.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/spark/union15.q.out Fri Nov 7 20:42:43 2014
@@ -116,14 +116,14 @@ STAGE PLANS:
keys: KEY._col0 (type: string)
mode: mergepartial
outputColumnNames: _col0, _col1
- Statistics: Num rows: 1 Data size: 100 Basic stats: COMPLETE Column stats: PARTIAL
+ Statistics: Num rows: 1 Data size: 96 Basic stats: COMPLETE Column stats: PARTIAL
Select Operator
expressions: _col0 (type: string), _col1 (type: bigint)
outputColumnNames: _col0, _col1
- Statistics: Num rows: 1 Data size: 100 Basic stats: COMPLETE Column stats: PARTIAL
+ Statistics: Num rows: 1 Data size: 96 Basic stats: COMPLETE Column stats: PARTIAL
File Output Operator
compressed: false
- Statistics: Num rows: 1 Data size: 100 Basic stats: COMPLETE Column stats: PARTIAL
+ Statistics: Num rows: 1 Data size: 96 Basic stats: COMPLETE Column stats: PARTIAL
table:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/union19.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/union19.q.out?rev=1637445&r1=1637444&r2=1637445&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/spark/union19.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/spark/union19.q.out Fri Nov 7 20:42:43 2014
@@ -134,14 +134,14 @@ STAGE PLANS:
keys: KEY._col0 (type: string)
mode: mergepartial
outputColumnNames: _col0, _col1
- Statistics: Num rows: 1 Data size: 100 Basic stats: COMPLETE Column stats: PARTIAL
+ Statistics: Num rows: 1 Data size: 96 Basic stats: COMPLETE Column stats: PARTIAL
Select Operator
expressions: _col0 (type: string), _col1 (type: bigint)
outputColumnNames: _col0, _col1
- Statistics: Num rows: 1 Data size: 100 Basic stats: COMPLETE Column stats: PARTIAL
+ Statistics: Num rows: 1 Data size: 96 Basic stats: COMPLETE Column stats: PARTIAL
File Output Operator
compressed: false
- Statistics: Num rows: 1 Data size: 100 Basic stats: COMPLETE Column stats: PARTIAL
+ Statistics: Num rows: 1 Data size: 96 Basic stats: COMPLETE Column stats: PARTIAL
table:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/union5.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/union5.q.out?rev=1637445&r1=1637444&r2=1637445&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/spark/union5.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/spark/union5.q.out Fri Nov 7 20:42:43 2014
@@ -89,14 +89,14 @@ STAGE PLANS:
keys: KEY._col0 (type: string)
mode: mergepartial
outputColumnNames: _col0, _col1
- Statistics: Num rows: 1 Data size: 100 Basic stats: COMPLETE Column stats: COMPLETE
+ Statistics: Num rows: 1 Data size: 96 Basic stats: COMPLETE Column stats: COMPLETE
Select Operator
expressions: _col0 (type: string), _col1 (type: bigint)
outputColumnNames: _col0, _col1
- Statistics: Num rows: 1 Data size: 100 Basic stats: COMPLETE Column stats: COMPLETE
+ Statistics: Num rows: 1 Data size: 96 Basic stats: COMPLETE Column stats: COMPLETE
File Output Operator
compressed: false
- Statistics: Num rows: 1 Data size: 100 Basic stats: COMPLETE Column stats: COMPLETE
+ Statistics: Num rows: 1 Data size: 96 Basic stats: COMPLETE Column stats: COMPLETE
table:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
@@ -144,5 +144,5 @@ POSTHOOK: query: select unionsrc.key, co
POSTHOOK: type: QUERY
POSTHOOK: Input: default@src
#### A masked pattern was here ####
-tst1 1
tst2 1
+tst1 1
Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/union7.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/union7.q.out?rev=1637445&r1=1637444&r2=1637445&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/spark/union7.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/spark/union7.q.out Fri Nov 7 20:42:43 2014
@@ -92,14 +92,14 @@ STAGE PLANS:
keys: KEY._col0 (type: string)
mode: mergepartial
outputColumnNames: _col0, _col1
- Statistics: Num rows: 1 Data size: 100 Basic stats: COMPLETE Column stats: PARTIAL
+ Statistics: Num rows: 1 Data size: 96 Basic stats: COMPLETE Column stats: PARTIAL
Select Operator
expressions: _col0 (type: string), _col1 (type: bigint)
outputColumnNames: _col0, _col1
- Statistics: Num rows: 1 Data size: 100 Basic stats: COMPLETE Column stats: PARTIAL
+ Statistics: Num rows: 1 Data size: 96 Basic stats: COMPLETE Column stats: PARTIAL
File Output Operator
compressed: false
- Statistics: Num rows: 1 Data size: 100 Basic stats: COMPLETE Column stats: PARTIAL
+ Statistics: Num rows: 1 Data size: 96 Basic stats: COMPLETE Column stats: PARTIAL
table:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/vector_between_in.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/vector_between_in.q.out?rev=1637445&r1=1637444&r2=1637445&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/spark/vector_between_in.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/spark/vector_between_in.q.out Fri Nov 7 20:42:43 2014
@@ -718,15 +718,15 @@ POSTHOOK: Input: default@decimal_date_te
14.9324324324
19.1135135135
20.3081081081
-22.1000000000
+22.1
24.4891891892
33.4486486486
34.6432432432
40.0189189189
42.4081081081
43.0054054054
-44.2000000000
-44.2000000000
+44.2
+44.2
44.7972972973
45.9918918919
PREHOOK: query: SELECT COUNT(*) FROM decimal_date_test WHERE cdecimal1 NOT BETWEEN -2000 AND 4390.1351351351
Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/vector_decimal_aggregate.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/vector_decimal_aggregate.q.out?rev=1637445&r1=1637444&r2=1637445&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/spark/vector_decimal_aggregate.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/spark/vector_decimal_aggregate.q.out Fri Nov 7 20:42:43 2014
@@ -113,14 +113,14 @@ POSTHOOK: query: SELECT cint,
POSTHOOK: type: QUERY
POSTHOOK: Input: default@decimal_vgby
#### A masked pattern was here ####
-762 2 5831542.2692483780 1531.2194054054 5833073.4886537834 2 6984454.21109769200000 1833.9456923076925 6986288.15678999969250
-253665376 1024 9767.0054054054 -9779.5486486487 -347484.0818378374 1024 11697.96923076923100 -11712.99230769231000 -416182.64030769233089
--3728 6 5831542.2692483780 -3367.6517567568 5817556.0411483778 6 6984454.21109769200000 -4033.445769230769 6967702.86724384584710
--563 2 -515.6210729730 -3367.6517567568 -3883.2728297298 2 -617.56077692307690 -4033.445769230769 -4651.00654615384590
-6981 3 5831542.269248378 -515.6210729730 5830511.0271024320 3 6984454.211097692 -617.56077692307690 6983219.08954384584620
-626923679 1024 9723.4027027027 -9778.9513513514 10541.0525297287 1024 11645.74615384615400 -11712.27692307692300 12625.04759999997746
-528534767 1024 5831542.2692483780 -9777.1594594595 11646372.8607481068 1024 6984454.21109769200000 -11710.13076923077100 13948892.79980307629003
-NULL 3072 9318.4351351351 -4298.1513513514 5018444.1081079808 3072 11160.71538461538500 -5147.90769230769300 6010604.30769230735360
+762 2 5831542.269248378 1531.2194054054 5833073.4886537834 2 6984454.211097692 1833.9456923076925 6986288.1567899996925
+253665376 1024 9767.0054054054 -9779.5486486487 -347484.0818378374 1024 11697.969230769231 -11712.99230769231 -416182.64030769233089
+-3728 6 5831542.269248378 -3367.6517567568 5817556.0411483778 6 6984454.211097692 -4033.445769230769 6967702.8672438458471
+-563 2 -515.6210729730 -3367.6517567568 -3883.2728297298 2 -617.5607769230769 -4033.445769230769 -4651.0065461538459
+6981 3 5831542.269248378 -515.6210729730 5830511.0271024320 3 6984454.211097692 -617.5607769230769 6983219.0895438458462
+626923679 1024 9723.4027027027 -9778.9513513514 10541.0525297287 1024 11645.746153846154 -11712.276923076923 12625.04759999997746
+528534767 1024 5831542.269248378 -9777.1594594595 11646372.8607481068 1024 6984454.211097692 -11710.130769230771 13948892.79980307629003
+NULL 3072 9318.4351351351 -4298.1513513514 5018444.1081079808 3072 11160.715384615385 -5147.907692307693 6010604.3076923073536
PREHOOK: query: -- Now add the others...
EXPLAIN SELECT cint,
COUNT(cdecimal1), MAX(cdecimal1), MIN(cdecimal1), SUM(cdecimal1), AVG(cdecimal1), STDDEV_POP(cdecimal1), STDDEV_SAMP(cdecimal1),
@@ -217,11 +217,11 @@ POSTHOOK: query: SELECT cint,
POSTHOOK: type: QUERY
POSTHOOK: Input: default@decimal_vgby
#### A masked pattern was here ####
-762 2 5831542.2692483780 1531.2194054054 5833073.4886537834 2916536.7443268917 2915005.5249214866 4122440.3477364695 2 6984454.21109769200000 1833.9456923076925 6986288.15678999969250 3493144.07839499984625 3491310.1327026924 4937458.140118758
-253665376 1024 9767.0054054054 -9779.5486486487 -347484.0818378374 -339.33992366976309 5708.9563478862 5711.745967572779 1024 11697.96923076923100 -11712.99230769231000 -416182.64030769233089 -406.428359675480791885 6837.632716002934 6840.973851172274
--3728 6 5831542.2692483780 -3367.6517567568 5817556.0411483778 969592.67352472963333 2174330.2092403853 2381859.406131774 6 6984454.21109769200000 -4033.445769230769 6967702.86724384584710 1161283.811207307641183333 2604201.2704476737 2852759.5602156054
--563 2 -515.6210729730 -3367.6517567568 -3883.2728297298 -1941.6364148649 1426.0153418918999 2016.6902366556308 2 -617.56077692307690 -4033.445769230769 -4651.00654615384590 -2325.50327307692295 1707.9424961538462 2415.395441814127
-6981 3 5831542.269248378 -515.6210729730 5830511.0271024320 1943503.67570081066667 2749258.455012492 3367140.1929065133 3 6984454.211097692 -617.56077692307690 6983219.08954384584620 2327739.696514615282066667 3292794.4113115156 4032833.0678006653
-626923679 1024 9723.4027027027 -9778.9513513514 10541.0525297287 10.29399661106318 5742.09145323734 5744.897264034267 1024 11645.74615384615400 -11712.27692307692300 12625.04759999997746 12.329148046874977988 6877.318722794877 6880.679250101603
-528534767 1024 5831542.2692483780 -9777.1594594595 11646372.8607481068 11373.41099682432305 257528.92988206653 257654.7686043977 1024 6984454.21109769200000 -11710.13076923077100 13948892.79980307629003 13621.965624807691689482 308443.1074570801 308593.82484083984
-NULL 3072 9318.4351351351 -4298.1513513514 5018444.1081079808 1633.60810810806667 5695.483082135364 5696.4103077145055 3072 11160.71538461538500 -5147.90769230769300 6010604.30769230735360 1956.576923076922966667 6821.495748565159 6822.606289190924
+762 2 5831542.269248378 1531.2194054054 5833073.4886537834 2916536.7443268917 2915005.5249214866 4122440.3477364695 2 6984454.211097692 1833.9456923076925 6986288.1567899996925 3493144.07839499984625 3491310.1327026924 4937458.140118758
+253665376 1024 9767.0054054054 -9779.5486486487 -347484.0818378374 -339.33992366976309 5708.9563478862 5711.745967572779 1024 11697.969230769231 -11712.99230769231 -416182.64030769233089 -406.428359675480791885 6837.632716002934 6840.973851172274
+-3728 6 5831542.269248378 -3367.6517567568 5817556.0411483778 969592.67352472963333 2174330.2092403853 2381859.406131774 6 6984454.211097692 -4033.445769230769 6967702.8672438458471 1161283.811207307641183333 2604201.2704476737 2852759.5602156054
+-563 2 -515.6210729730 -3367.6517567568 -3883.2728297298 -1941.6364148649 1426.0153418918999 2016.6902366556308 2 -617.5607769230769 -4033.445769230769 -4651.0065461538459 -2325.50327307692295 1707.9424961538462 2415.395441814127
+6981 3 5831542.269248378 -515.6210729730 5830511.0271024320 1943503.67570081066667 2749258.455012492 3367140.1929065133 3 6984454.211097692 -617.5607769230769 6983219.0895438458462 2327739.696514615282066667 3292794.4113115156 4032833.0678006653
+626923679 1024 9723.4027027027 -9778.9513513514 10541.0525297287 10.29399661106318 5742.09145323734 5744.897264034267 1024 11645.746153846154 -11712.276923076923 12625.04759999997746 12.329148046874977988 6877.318722794877 6880.679250101603
+528534767 1024 5831542.269248378 -9777.1594594595 11646372.8607481068 11373.41099682432305 257528.92988206653 257654.7686043977 1024 6984454.211097692 -11710.130769230771 13948892.79980307629003 13621.965624807691689482 308443.1074570801 308593.82484083984
+NULL 3072 9318.4351351351 -4298.1513513514 5018444.1081079808 1633.60810810806667 5695.483082135364 5696.4103077145055 3072 11160.715384615385 -5147.907692307693 6010604.3076923073536 1956.576923076922966667 6821.495748565159 6822.606289190924
Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/vector_mapjoin_reduce.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/vector_mapjoin_reduce.q.out?rev=1637445&r1=1637444&r2=1637445&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/spark/vector_mapjoin_reduce.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientpositive/spark/vector_mapjoin_reduce.q.out Fri Nov 7 20:42:43 2014
@@ -1,104 +1,6 @@
PREHOOK: query: -- SORT_QUERY_RESULTS
-DROP TABLE part
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: -- SORT_QUERY_RESULTS
-
-DROP TABLE part
-POSTHOOK: type: DROPTABLE
-PREHOOK: query: -- data setup
-CREATE TABLE part(
- p_partkey INT,
- p_name STRING,
- p_mfgr STRING,
- p_brand STRING,
- p_type STRING,
- p_size INT,
- p_container STRING,
- p_retailprice DOUBLE,
- p_comment STRING
-)
-PREHOOK: type: CREATETABLE
-PREHOOK: Output: database:default
-PREHOOK: Output: default@part
-POSTHOOK: query: -- data setup
-CREATE TABLE part(
- p_partkey INT,
- p_name STRING,
- p_mfgr STRING,
- p_brand STRING,
- p_type STRING,
- p_size INT,
- p_container STRING,
- p_retailprice DOUBLE,
- p_comment STRING
-)
-POSTHOOK: type: CREATETABLE
-POSTHOOK: Output: database:default
-POSTHOOK: Output: default@part
-PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/part_tiny.txt' overwrite into table part
-PREHOOK: type: LOAD
-#### A masked pattern was here ####
-PREHOOK: Output: default@part
-POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/part_tiny.txt' overwrite into table part
-POSTHOOK: type: LOAD
-#### A masked pattern was here ####
-POSTHOOK: Output: default@part
-PREHOOK: query: DROP TABLE lineitem
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: DROP TABLE lineitem
-POSTHOOK: type: DROPTABLE
-PREHOOK: query: CREATE TABLE lineitem (L_ORDERKEY INT,
- L_PARTKEY INT,
- L_SUPPKEY INT,
- L_LINENUMBER INT,
- L_QUANTITY DOUBLE,
- L_EXTENDEDPRICE DOUBLE,
- L_DISCOUNT DOUBLE,
- L_TAX DOUBLE,
- L_RETURNFLAG STRING,
- L_LINESTATUS STRING,
- l_shipdate STRING,
- L_COMMITDATE STRING,
- L_RECEIPTDATE STRING,
- L_SHIPINSTRUCT STRING,
- L_SHIPMODE STRING,
- L_COMMENT STRING)
-ROW FORMAT DELIMITED
-FIELDS TERMINATED BY '|'
-PREHOOK: type: CREATETABLE
-PREHOOK: Output: database:default
-PREHOOK: Output: default@lineitem
-POSTHOOK: query: CREATE TABLE lineitem (L_ORDERKEY INT,
- L_PARTKEY INT,
- L_SUPPKEY INT,
- L_LINENUMBER INT,
- L_QUANTITY DOUBLE,
- L_EXTENDEDPRICE DOUBLE,
- L_DISCOUNT DOUBLE,
- L_TAX DOUBLE,
- L_RETURNFLAG STRING,
- L_LINESTATUS STRING,
- l_shipdate STRING,
- L_COMMITDATE STRING,
- L_RECEIPTDATE STRING,
- L_SHIPINSTRUCT STRING,
- L_SHIPMODE STRING,
- L_COMMENT STRING)
-ROW FORMAT DELIMITED
-FIELDS TERMINATED BY '|'
-POSTHOOK: type: CREATETABLE
-POSTHOOK: Output: database:default
-POSTHOOK: Output: default@lineitem
-PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/lineitem.txt' OVERWRITE INTO TABLE lineitem
-PREHOOK: type: LOAD
-#### A masked pattern was here ####
-PREHOOK: Output: default@lineitem
-POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/lineitem.txt' OVERWRITE INTO TABLE lineitem
-POSTHOOK: type: LOAD
-#### A masked pattern was here ####
-POSTHOOK: Output: default@lineitem
-PREHOOK: query: -- Verify HIVE-8097 with a query that has a Vectorized MapJoin in the Reducer.
+-- Verify HIVE-8097 with a query that has a Vectorized MapJoin in the Reducer.
-- Query copied from subquery_in.q
-- non agg, non corr, with join in Parent Query
@@ -108,7 +10,9 @@ from (select distinct l_partkey as p_par
where li.l_linenumber = 1 and
li.l_orderkey in (select l_orderkey from lineitem where l_shipmode = 'AIR')
PREHOOK: type: QUERY
-POSTHOOK: query: -- Verify HIVE-8097 with a query that has a Vectorized MapJoin in the Reducer.
+POSTHOOK: query: -- SORT_QUERY_RESULTS
+
+-- Verify HIVE-8097 with a query that has a Vectorized MapJoin in the Reducer.
-- Query copied from subquery_in.q
-- non agg, non corr, with join in Parent Query
@@ -135,60 +39,60 @@ STAGE PLANS:
Map Operator Tree:
TableScan
alias: li
- Statistics: Num rows: 756 Data size: 12099 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 100 Data size: 11999 Basic stats: COMPLETE Column stats: NONE
Filter Operator
predicate: ((l_partkey is not null and l_orderkey is not null) and (l_linenumber = 1)) (type: boolean)
- Statistics: Num rows: 94 Data size: 1504 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 12 Data size: 1439 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
key expressions: l_partkey (type: int)
sort order: +
Map-reduce partition columns: l_partkey (type: int)
- Statistics: Num rows: 94 Data size: 1504 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 12 Data size: 1439 Basic stats: COMPLETE Column stats: NONE
value expressions: l_orderkey (type: int), l_suppkey (type: int)
Map 4
Map Operator Tree:
TableScan
alias: lineitem
- Statistics: Num rows: 3024 Data size: 12099 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 100 Data size: 11999 Basic stats: COMPLETE Column stats: NONE
Filter Operator
predicate: l_partkey is not null (type: boolean)
- Statistics: Num rows: 1512 Data size: 6049 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 50 Data size: 5999 Basic stats: COMPLETE Column stats: NONE
Select Operator
expressions: l_partkey (type: int)
outputColumnNames: l_partkey
- Statistics: Num rows: 1512 Data size: 6049 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 50 Data size: 5999 Basic stats: COMPLETE Column stats: NONE
Group By Operator
keys: l_partkey (type: int)
mode: hash
outputColumnNames: _col0
- Statistics: Num rows: 1512 Data size: 6049 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 50 Data size: 5999 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
key expressions: _col0 (type: int)
sort order: +
Map-reduce partition columns: _col0 (type: int)
- Statistics: Num rows: 1512 Data size: 6049 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 50 Data size: 5999 Basic stats: COMPLETE Column stats: NONE
Map 6
Map Operator Tree:
TableScan
alias: lineitem
- Statistics: Num rows: 1728 Data size: 12099 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 100 Data size: 11999 Basic stats: COMPLETE Column stats: NONE
Filter Operator
predicate: ((l_shipmode = 'AIR') and l_orderkey is not null) (type: boolean)
- Statistics: Num rows: 432 Data size: 3024 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 25 Data size: 2999 Basic stats: COMPLETE Column stats: NONE
Select Operator
expressions: l_orderkey (type: int)
outputColumnNames: _col0
- Statistics: Num rows: 432 Data size: 3024 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 25 Data size: 2999 Basic stats: COMPLETE Column stats: NONE
Group By Operator
keys: _col0 (type: int)
mode: hash
outputColumnNames: _col0
- Statistics: Num rows: 432 Data size: 3024 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 25 Data size: 2999 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
key expressions: _col0 (type: int)
sort order: +
Map-reduce partition columns: _col0 (type: int)
- Statistics: Num rows: 432 Data size: 3024 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 25 Data size: 2999 Basic stats: COMPLETE Column stats: NONE
Reducer 2
Reduce Operator Tree:
Join Operator
@@ -198,12 +102,12 @@ STAGE PLANS:
0 {KEY.reducesinkkey0}
1 {VALUE._col0} {VALUE._col1}
outputColumnNames: _col0, _col1, _col3
- Statistics: Num rows: 831 Data size: 3326 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 27 Data size: 3298 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
key expressions: _col1 (type: int)
sort order: +
Map-reduce partition columns: _col1 (type: int)
- Statistics: Num rows: 831 Data size: 3326 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 27 Data size: 3298 Basic stats: COMPLETE Column stats: NONE
value expressions: _col0 (type: int), _col3 (type: int)
Reducer 3
Reduce Operator Tree:
@@ -214,14 +118,14 @@ STAGE PLANS:
0 {VALUE._col0} {VALUE._col2}
1
outputColumnNames: _col0, _col3
- Statistics: Num rows: 914 Data size: 3658 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 29 Data size: 3627 Basic stats: COMPLETE Column stats: NONE
Select Operator
expressions: _col0 (type: int), _col3 (type: int)
outputColumnNames: _col0, _col1
- Statistics: Num rows: 914 Data size: 3658 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 29 Data size: 3627 Basic stats: COMPLETE Column stats: NONE
File Output Operator
compressed: false
- Statistics: Num rows: 914 Data size: 3658 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 29 Data size: 3627 Basic stats: COMPLETE Column stats: NONE
table:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
@@ -232,16 +136,16 @@ STAGE PLANS:
keys: KEY._col0 (type: int)
mode: mergepartial
outputColumnNames: _col0
- Statistics: Num rows: 756 Data size: 3024 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 25 Data size: 2999 Basic stats: COMPLETE Column stats: NONE
Select Operator
expressions: _col0 (type: int)
outputColumnNames: _col0
- Statistics: Num rows: 756 Data size: 3024 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 25 Data size: 2999 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
key expressions: _col0 (type: int)
sort order: +
Map-reduce partition columns: _col0 (type: int)
- Statistics: Num rows: 756 Data size: 3024 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 25 Data size: 2999 Basic stats: COMPLETE Column stats: NONE
Execution mode: vectorized
Stage: Stage-0
@@ -305,60 +209,60 @@ STAGE PLANS:
Map Operator Tree:
TableScan
alias: li
- Statistics: Num rows: 756 Data size: 12099 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 100 Data size: 11999 Basic stats: COMPLETE Column stats: NONE
Filter Operator
predicate: (((l_partkey is not null and l_orderkey is not null) and l_linenumber is not null) and (l_linenumber = 1)) (type: boolean)
- Statistics: Num rows: 47 Data size: 752 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 6 Data size: 719 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
key expressions: l_partkey (type: int)
sort order: +
Map-reduce partition columns: l_partkey (type: int)
- Statistics: Num rows: 47 Data size: 752 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 6 Data size: 719 Basic stats: COMPLETE Column stats: NONE
value expressions: l_orderkey (type: int), l_suppkey (type: int)
Map 4
Map Operator Tree:
TableScan
alias: lineitem
- Statistics: Num rows: 3024 Data size: 12099 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 100 Data size: 11999 Basic stats: COMPLETE Column stats: NONE
Filter Operator
predicate: l_partkey is not null (type: boolean)
- Statistics: Num rows: 1512 Data size: 6049 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 50 Data size: 5999 Basic stats: COMPLETE Column stats: NONE
Select Operator
expressions: l_partkey (type: int)
outputColumnNames: l_partkey
- Statistics: Num rows: 1512 Data size: 6049 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 50 Data size: 5999 Basic stats: COMPLETE Column stats: NONE
Group By Operator
keys: l_partkey (type: int)
mode: hash
outputColumnNames: _col0
- Statistics: Num rows: 1512 Data size: 6049 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 50 Data size: 5999 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
key expressions: _col0 (type: int)
sort order: +
Map-reduce partition columns: _col0 (type: int)
- Statistics: Num rows: 1512 Data size: 6049 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 50 Data size: 5999 Basic stats: COMPLETE Column stats: NONE
Map 6
Map Operator Tree:
TableScan
alias: lineitem
- Statistics: Num rows: 1099 Data size: 12099 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 100 Data size: 11999 Basic stats: COMPLETE Column stats: NONE
Filter Operator
- predicate: (((l_shipmode = 'AIR') and l_orderkey is not null) and l_linenumber is not null) (type: boolean)
- Statistics: Num rows: 138 Data size: 1519 Basic stats: COMPLETE Column stats: NONE
+ predicate: ((((l_shipmode = 'AIR') and l_orderkey is not null) and l_linenumber is not null) and (l_linenumber = 1)) (type: boolean)
+ Statistics: Num rows: 6 Data size: 719 Basic stats: COMPLETE Column stats: NONE
Select Operator
- expressions: l_orderkey (type: int), l_linenumber (type: int)
+ expressions: l_orderkey (type: int), 1 (type: int)
outputColumnNames: _col0, _col1
- Statistics: Num rows: 138 Data size: 1519 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 6 Data size: 719 Basic stats: COMPLETE Column stats: NONE
Group By Operator
keys: _col0 (type: int), _col1 (type: int)
mode: hash
outputColumnNames: _col0, _col1
- Statistics: Num rows: 138 Data size: 1519 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 6 Data size: 719 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
key expressions: _col0 (type: int), _col1 (type: int)
sort order: ++
Map-reduce partition columns: _col0 (type: int), _col1 (type: int)
- Statistics: Num rows: 138 Data size: 1519 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 6 Data size: 719 Basic stats: COMPLETE Column stats: NONE
Reducer 2
Reduce Operator Tree:
Join Operator
@@ -368,12 +272,12 @@ STAGE PLANS:
0 {KEY.reducesinkkey0}
1 {VALUE._col0} {VALUE._col1}
outputColumnNames: _col0, _col1, _col3
- Statistics: Num rows: 831 Data size: 3326 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 27 Data size: 3298 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
key expressions: _col1 (type: int), 1 (type: int)
sort order: ++
Map-reduce partition columns: _col1 (type: int)
- Statistics: Num rows: 831 Data size: 3326 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 27 Data size: 3298 Basic stats: COMPLETE Column stats: NONE
value expressions: _col0 (type: int), _col3 (type: int)
Reducer 3
Reduce Operator Tree:
@@ -384,14 +288,14 @@ STAGE PLANS:
0 {VALUE._col0} {VALUE._col2}
1
outputColumnNames: _col0, _col3
- Statistics: Num rows: 914 Data size: 3658 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 29 Data size: 3627 Basic stats: COMPLETE Column stats: NONE
Select Operator
expressions: _col0 (type: int), _col3 (type: int)
outputColumnNames: _col0, _col1
- Statistics: Num rows: 914 Data size: 3658 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 29 Data size: 3627 Basic stats: COMPLETE Column stats: NONE
File Output Operator
compressed: false
- Statistics: Num rows: 914 Data size: 3658 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 29 Data size: 3627 Basic stats: COMPLETE Column stats: NONE
table:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
@@ -402,16 +306,16 @@ STAGE PLANS:
keys: KEY._col0 (type: int)
mode: mergepartial
outputColumnNames: _col0
- Statistics: Num rows: 756 Data size: 3024 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 25 Data size: 2999 Basic stats: COMPLETE Column stats: NONE
Select Operator
expressions: _col0 (type: int)
outputColumnNames: _col0
- Statistics: Num rows: 756 Data size: 3024 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 25 Data size: 2999 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
key expressions: _col0 (type: int)
sort order: +
Map-reduce partition columns: _col0 (type: int)
- Statistics: Num rows: 756 Data size: 3024 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 25 Data size: 2999 Basic stats: COMPLETE Column stats: NONE
Execution mode: vectorized
Stage: Stage-0