You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hive.apache.org by kg...@apache.org on 2019/01/23 16:39:12 UTC
[40/51] [partial] hive git commit: HIVE-17084: Turn on
hive.stats.fetch.column.stats configuration flag (Zoltan Haindrich reviewed
by Ashutosh Chauhan)
http://git-wip-us.apache.org/repos/asf/hive/blob/268a6e5a/ql/src/test/results/clientpositive/bucketmapjoin_negative3.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/bucketmapjoin_negative3.q.out b/ql/src/test/results/clientpositive/bucketmapjoin_negative3.q.out
index bd6b45f..7b2176a 100644
--- a/ql/src/test/results/clientpositive/bucketmapjoin_negative3.q.out
+++ b/ql/src/test/results/clientpositive/bucketmapjoin_negative3.q.out
@@ -167,12 +167,12 @@ STAGE PLANS:
TableScan
alias: r
filterExpr: (key is not null and value is not null) (type: boolean)
- Statistics: Num rows: 1 Data size: 42000 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE
GatherStats: false
Filter Operator
isSamplingPred: false
predicate: (key is not null and value is not null) (type: boolean)
- Statistics: Num rows: 1 Data size: 42000 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE
HashTable Sink Operator
keys:
0 key (type: string), value (type: string)
@@ -192,12 +192,12 @@ STAGE PLANS:
TableScan
alias: l
filterExpr: (key is not null and value is not null) (type: boolean)
- Statistics: Num rows: 1 Data size: 42000 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE
GatherStats: false
Filter Operator
isSamplingPred: false
predicate: (key is not null and value is not null) (type: boolean)
- Statistics: Num rows: 1 Data size: 42000 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE
Map Join Operator
condition map:
Inner Join 0 to 1
@@ -206,18 +206,18 @@ STAGE PLANS:
1 key (type: string), value (type: string)
outputColumnNames: _col0, _col1, _col5, _col6
Position of Big Table: 0
- Statistics: Num rows: 1 Data size: 46200 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 404 Basic stats: COMPLETE Column stats: NONE
BucketMapJoin: true
Select Operator
expressions: _col0 (type: string), _col1 (type: string), _col5 (type: string), _col6 (type: string)
outputColumnNames: _col0, _col1, _col2, _col3
- Statistics: Num rows: 1 Data size: 46200 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 404 Basic stats: COMPLETE Column stats: NONE
File Output Operator
compressed: false
GlobalTableId: 0
#### A masked pattern was here ####
NumFilesPerFileSink: 1
- Statistics: Num rows: 1 Data size: 46200 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 404 Basic stats: COMPLETE Column stats: NONE
#### A masked pattern was here ####
table:
input format: org.apache.hadoop.mapred.SequenceFileInputFormat
@@ -324,12 +324,12 @@ STAGE PLANS:
TableScan
alias: r
filterExpr: (key is not null and value is not null) (type: boolean)
- Statistics: Num rows: 1 Data size: 42000 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE
GatherStats: false
Filter Operator
isSamplingPred: false
predicate: (key is not null and value is not null) (type: boolean)
- Statistics: Num rows: 1 Data size: 42000 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE
HashTable Sink Operator
keys:
0 key (type: string), value (type: string)
@@ -349,12 +349,12 @@ STAGE PLANS:
TableScan
alias: l
filterExpr: (key is not null and value is not null) (type: boolean)
- Statistics: Num rows: 1 Data size: 42000 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE
GatherStats: false
Filter Operator
isSamplingPred: false
predicate: (key is not null and value is not null) (type: boolean)
- Statistics: Num rows: 1 Data size: 42000 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE
Map Join Operator
condition map:
Inner Join 0 to 1
@@ -363,18 +363,18 @@ STAGE PLANS:
1 key (type: string), value (type: string)
outputColumnNames: _col0, _col1, _col5, _col6
Position of Big Table: 0
- Statistics: Num rows: 1 Data size: 46200 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 404 Basic stats: COMPLETE Column stats: NONE
BucketMapJoin: true
Select Operator
expressions: _col0 (type: string), _col1 (type: string), _col5 (type: string), _col6 (type: string)
outputColumnNames: _col0, _col1, _col2, _col3
- Statistics: Num rows: 1 Data size: 46200 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 404 Basic stats: COMPLETE Column stats: NONE
File Output Operator
compressed: false
GlobalTableId: 0
#### A masked pattern was here ####
NumFilesPerFileSink: 1
- Statistics: Num rows: 1 Data size: 46200 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 404 Basic stats: COMPLETE Column stats: NONE
#### A masked pattern was here ####
table:
input format: org.apache.hadoop.mapred.SequenceFileInputFormat
@@ -481,12 +481,12 @@ STAGE PLANS:
TableScan
alias: r
filterExpr: UDFToDouble(key) is not null (type: boolean)
- Statistics: Num rows: 1 Data size: 42000 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE
GatherStats: false
Filter Operator
isSamplingPred: false
predicate: UDFToDouble(key) is not null (type: boolean)
- Statistics: Num rows: 1 Data size: 42000 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE
HashTable Sink Operator
keys:
0 (key + key) (type: double)
@@ -499,12 +499,12 @@ STAGE PLANS:
TableScan
alias: l
filterExpr: (key + key) is not null (type: boolean)
- Statistics: Num rows: 1 Data size: 42000 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE
GatherStats: false
Filter Operator
isSamplingPred: false
predicate: (key + key) is not null (type: boolean)
- Statistics: Num rows: 1 Data size: 42000 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE
Map Join Operator
condition map:
Inner Join 0 to 1
@@ -513,17 +513,17 @@ STAGE PLANS:
1 UDFToDouble(key) (type: double)
outputColumnNames: _col0, _col1, _col5, _col6
Position of Big Table: 0
- Statistics: Num rows: 1 Data size: 46200 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 404 Basic stats: COMPLETE Column stats: NONE
Select Operator
expressions: _col0 (type: string), _col1 (type: string), _col5 (type: string), _col6 (type: string)
outputColumnNames: _col0, _col1, _col2, _col3
- Statistics: Num rows: 1 Data size: 46200 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 404 Basic stats: COMPLETE Column stats: NONE
File Output Operator
compressed: false
GlobalTableId: 0
#### A masked pattern was here ####
NumFilesPerFileSink: 1
- Statistics: Num rows: 1 Data size: 46200 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 404 Basic stats: COMPLETE Column stats: NONE
#### A masked pattern was here ####
table:
input format: org.apache.hadoop.mapred.SequenceFileInputFormat
@@ -632,12 +632,12 @@ STAGE PLANS:
TableScan
alias: r
filterExpr: (key is not null and value is not null) (type: boolean)
- Statistics: Num rows: 1 Data size: 42000 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE
GatherStats: false
Filter Operator
isSamplingPred: false
predicate: (key is not null and value is not null) (type: boolean)
- Statistics: Num rows: 1 Data size: 42000 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE
HashTable Sink Operator
keys:
0 key (type: string), value (type: string)
@@ -650,12 +650,12 @@ STAGE PLANS:
TableScan
alias: l
filterExpr: (key is not null and value is not null) (type: boolean)
- Statistics: Num rows: 1 Data size: 42000 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE
GatherStats: false
Filter Operator
isSamplingPred: false
predicate: (key is not null and value is not null) (type: boolean)
- Statistics: Num rows: 1 Data size: 42000 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE
Map Join Operator
condition map:
Inner Join 0 to 1
@@ -664,17 +664,17 @@ STAGE PLANS:
1 key (type: string), value (type: string)
outputColumnNames: _col0, _col1, _col5, _col6
Position of Big Table: 0
- Statistics: Num rows: 1 Data size: 46200 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 404 Basic stats: COMPLETE Column stats: NONE
Select Operator
expressions: _col0 (type: string), _col1 (type: string), _col5 (type: string), _col6 (type: string)
outputColumnNames: _col0, _col1, _col2, _col3
- Statistics: Num rows: 1 Data size: 46200 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 404 Basic stats: COMPLETE Column stats: NONE
File Output Operator
compressed: false
GlobalTableId: 0
#### A masked pattern was here ####
NumFilesPerFileSink: 1
- Statistics: Num rows: 1 Data size: 46200 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 404 Basic stats: COMPLETE Column stats: NONE
#### A masked pattern was here ####
table:
input format: org.apache.hadoop.mapred.SequenceFileInputFormat
@@ -783,12 +783,12 @@ STAGE PLANS:
TableScan
alias: r
filterExpr: (key is not null and value is not null) (type: boolean)
- Statistics: Num rows: 1 Data size: 42000 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE
GatherStats: false
Filter Operator
isSamplingPred: false
predicate: (key is not null and value is not null) (type: boolean)
- Statistics: Num rows: 1 Data size: 42000 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE
HashTable Sink Operator
keys:
0 key (type: string), value (type: string)
@@ -801,12 +801,12 @@ STAGE PLANS:
TableScan
alias: l
filterExpr: (key is not null and value is not null) (type: boolean)
- Statistics: Num rows: 1 Data size: 42000 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE
GatherStats: false
Filter Operator
isSamplingPred: false
predicate: (key is not null and value is not null) (type: boolean)
- Statistics: Num rows: 1 Data size: 42000 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE
Map Join Operator
condition map:
Inner Join 0 to 1
@@ -815,17 +815,17 @@ STAGE PLANS:
1 key (type: string), value (type: string)
outputColumnNames: _col0, _col1, _col5, _col6
Position of Big Table: 0
- Statistics: Num rows: 1 Data size: 46200 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 404 Basic stats: COMPLETE Column stats: NONE
Select Operator
expressions: _col0 (type: string), _col1 (type: string), _col5 (type: string), _col6 (type: string)
outputColumnNames: _col0, _col1, _col2, _col3
- Statistics: Num rows: 1 Data size: 46200 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 404 Basic stats: COMPLETE Column stats: NONE
File Output Operator
compressed: false
GlobalTableId: 0
#### A masked pattern was here ####
NumFilesPerFileSink: 1
- Statistics: Num rows: 1 Data size: 46200 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 404 Basic stats: COMPLETE Column stats: NONE
#### A masked pattern was here ####
table:
input format: org.apache.hadoop.mapred.SequenceFileInputFormat
@@ -934,12 +934,12 @@ STAGE PLANS:
TableScan
alias: r
filterExpr: (key is not null and value is not null) (type: boolean)
- Statistics: Num rows: 1 Data size: 42000 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE
GatherStats: false
Filter Operator
isSamplingPred: false
predicate: (key is not null and value is not null) (type: boolean)
- Statistics: Num rows: 1 Data size: 42000 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE
HashTable Sink Operator
keys:
0 key (type: string), value (type: string)
@@ -952,12 +952,12 @@ STAGE PLANS:
TableScan
alias: l
filterExpr: (key is not null and value is not null) (type: boolean)
- Statistics: Num rows: 1 Data size: 42000 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE
GatherStats: false
Filter Operator
isSamplingPred: false
predicate: (key is not null and value is not null) (type: boolean)
- Statistics: Num rows: 1 Data size: 42000 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE
Map Join Operator
condition map:
Inner Join 0 to 1
@@ -966,17 +966,17 @@ STAGE PLANS:
1 key (type: string), value (type: string)
outputColumnNames: _col0, _col1, _col5, _col6
Position of Big Table: 0
- Statistics: Num rows: 1 Data size: 46200 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 404 Basic stats: COMPLETE Column stats: NONE
Select Operator
expressions: _col0 (type: string), _col1 (type: string), _col5 (type: string), _col6 (type: string)
outputColumnNames: _col0, _col1, _col2, _col3
- Statistics: Num rows: 1 Data size: 46200 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 404 Basic stats: COMPLETE Column stats: NONE
File Output Operator
compressed: false
GlobalTableId: 0
#### A masked pattern was here ####
NumFilesPerFileSink: 1
- Statistics: Num rows: 1 Data size: 46200 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 404 Basic stats: COMPLETE Column stats: NONE
#### A masked pattern was here ####
table:
input format: org.apache.hadoop.mapred.SequenceFileInputFormat
@@ -1085,12 +1085,12 @@ STAGE PLANS:
TableScan
alias: r
filterExpr: (key is not null and value is not null) (type: boolean)
- Statistics: Num rows: 1 Data size: 42000 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE
GatherStats: false
Filter Operator
isSamplingPred: false
predicate: (key is not null and value is not null) (type: boolean)
- Statistics: Num rows: 1 Data size: 42000 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE
HashTable Sink Operator
keys:
0 key (type: string), value (type: string)
@@ -1103,12 +1103,12 @@ STAGE PLANS:
TableScan
alias: l
filterExpr: (key is not null and value is not null) (type: boolean)
- Statistics: Num rows: 1 Data size: 42000 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE
GatherStats: false
Filter Operator
isSamplingPred: false
predicate: (key is not null and value is not null) (type: boolean)
- Statistics: Num rows: 1 Data size: 42000 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE
Map Join Operator
condition map:
Inner Join 0 to 1
@@ -1117,17 +1117,17 @@ STAGE PLANS:
1 key (type: string), value (type: string)
outputColumnNames: _col0, _col1, _col5, _col6
Position of Big Table: 0
- Statistics: Num rows: 1 Data size: 46200 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 404 Basic stats: COMPLETE Column stats: NONE
Select Operator
expressions: _col0 (type: string), _col1 (type: string), _col5 (type: string), _col6 (type: string)
outputColumnNames: _col0, _col1, _col2, _col3
- Statistics: Num rows: 1 Data size: 46200 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 404 Basic stats: COMPLETE Column stats: NONE
File Output Operator
compressed: false
GlobalTableId: 0
#### A masked pattern was here ####
NumFilesPerFileSink: 1
- Statistics: Num rows: 1 Data size: 46200 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 404 Basic stats: COMPLETE Column stats: NONE
#### A masked pattern was here ####
table:
input format: org.apache.hadoop.mapred.SequenceFileInputFormat
@@ -1236,12 +1236,12 @@ STAGE PLANS:
TableScan
alias: r
filterExpr: (key is not null and value is not null) (type: boolean)
- Statistics: Num rows: 1 Data size: 42000 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE
GatherStats: false
Filter Operator
isSamplingPred: false
predicate: (key is not null and value is not null) (type: boolean)
- Statistics: Num rows: 1 Data size: 42000 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE
HashTable Sink Operator
keys:
0 key (type: string), value (type: string)
@@ -1254,12 +1254,12 @@ STAGE PLANS:
TableScan
alias: l
filterExpr: (key is not null and value is not null) (type: boolean)
- Statistics: Num rows: 1 Data size: 42000 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE
GatherStats: false
Filter Operator
isSamplingPred: false
predicate: (key is not null and value is not null) (type: boolean)
- Statistics: Num rows: 1 Data size: 42000 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE
Map Join Operator
condition map:
Inner Join 0 to 1
@@ -1268,17 +1268,17 @@ STAGE PLANS:
1 key (type: string), value (type: string)
outputColumnNames: _col0, _col1, _col5, _col6
Position of Big Table: 0
- Statistics: Num rows: 1 Data size: 46200 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 404 Basic stats: COMPLETE Column stats: NONE
Select Operator
expressions: _col0 (type: string), _col1 (type: string), _col5 (type: string), _col6 (type: string)
outputColumnNames: _col0, _col1, _col2, _col3
- Statistics: Num rows: 1 Data size: 46200 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 404 Basic stats: COMPLETE Column stats: NONE
File Output Operator
compressed: false
GlobalTableId: 0
#### A masked pattern was here ####
NumFilesPerFileSink: 1
- Statistics: Num rows: 1 Data size: 46200 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 404 Basic stats: COMPLETE Column stats: NONE
#### A masked pattern was here ####
table:
input format: org.apache.hadoop.mapred.SequenceFileInputFormat
@@ -1387,12 +1387,12 @@ STAGE PLANS:
TableScan
alias: r
filterExpr: (key is not null and value is not null) (type: boolean)
- Statistics: Num rows: 1 Data size: 42000 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE
GatherStats: false
Filter Operator
isSamplingPred: false
predicate: (key is not null and value is not null) (type: boolean)
- Statistics: Num rows: 1 Data size: 42000 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE
HashTable Sink Operator
keys:
0 key (type: string), value (type: string)
@@ -1405,12 +1405,12 @@ STAGE PLANS:
TableScan
alias: l
filterExpr: (key is not null and value is not null) (type: boolean)
- Statistics: Num rows: 1 Data size: 42000 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE
GatherStats: false
Filter Operator
isSamplingPred: false
predicate: (key is not null and value is not null) (type: boolean)
- Statistics: Num rows: 1 Data size: 42000 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE
Map Join Operator
condition map:
Inner Join 0 to 1
@@ -1419,17 +1419,17 @@ STAGE PLANS:
1 key (type: string), value (type: string)
outputColumnNames: _col0, _col1, _col5, _col6
Position of Big Table: 0
- Statistics: Num rows: 1 Data size: 46200 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 404 Basic stats: COMPLETE Column stats: NONE
Select Operator
expressions: _col0 (type: string), _col1 (type: string), _col5 (type: string), _col6 (type: string)
outputColumnNames: _col0, _col1, _col2, _col3
- Statistics: Num rows: 1 Data size: 46200 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 404 Basic stats: COMPLETE Column stats: NONE
File Output Operator
compressed: false
GlobalTableId: 0
#### A masked pattern was here ####
NumFilesPerFileSink: 1
- Statistics: Num rows: 1 Data size: 46200 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 404 Basic stats: COMPLETE Column stats: NONE
#### A masked pattern was here ####
table:
input format: org.apache.hadoop.mapred.SequenceFileInputFormat
http://git-wip-us.apache.org/repos/asf/hive/blob/268a6e5a/ql/src/test/results/clientpositive/bucketsortoptimize_insert_1.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/bucketsortoptimize_insert_1.q.out b/ql/src/test/results/clientpositive/bucketsortoptimize_insert_1.q.out
index a410806..46fd587 100644
--- a/ql/src/test/results/clientpositive/bucketsortoptimize_insert_1.q.out
+++ b/ql/src/test/results/clientpositive/bucketsortoptimize_insert_1.q.out
@@ -62,14 +62,14 @@ STAGE PLANS:
TableScan
alias: a
filterExpr: (ds = '1') (type: boolean)
- Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE
Select Operator
expressions: key (type: int), value (type: string)
outputColumnNames: _col0, _col1
- Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE
File Output Operator
compressed: false
- Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE
table:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
@@ -179,14 +179,14 @@ STAGE PLANS:
TableScan
alias: a
filterExpr: (ds = '1') (type: boolean)
- Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE
Select Operator
expressions: key (type: int), value (type: string)
outputColumnNames: _col0, _col1
- Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE
File Output Operator
compressed: false
- Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE
table:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
@@ -296,14 +296,14 @@ STAGE PLANS:
TableScan
alias: a
filterExpr: (ds = '1') (type: boolean)
- Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE
Select Operator
expressions: key (type: int), concat(value, value) (type: string)
outputColumnNames: _col0, _col1
- Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 500 Data size: 94000 Basic stats: COMPLETE Column stats: COMPLETE
File Output Operator
compressed: false
- Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 500 Data size: 94000 Basic stats: COMPLETE Column stats: COMPLETE
table:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
@@ -363,25 +363,25 @@ STAGE PLANS:
TableScan
alias: a
filterExpr: (ds = '1') (type: boolean)
- Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE
Select Operator
expressions: (key + key) (type: int), value (type: string)
outputColumnNames: _col0, _col1
- Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE
Reduce Output Operator
key expressions: _col0 (type: int)
sort order: +
Map-reduce partition columns: _col0 (type: int)
- Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE
value expressions: _col1 (type: string)
Reduce Operator Tree:
Select Operator
expressions: KEY.reducesinkkey0 (type: int), VALUE._col0 (type: string)
outputColumnNames: _col0, _col1
- Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE
File Output Operator
compressed: false
- Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE
table:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
@@ -390,13 +390,13 @@ STAGE PLANS:
Select Operator
expressions: _col0 (type: int), _col1 (type: string), '1' (type: string)
outputColumnNames: key, value, ds
- Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 500 Data size: 90000 Basic stats: COMPLETE Column stats: COMPLETE
Group By Operator
aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll')
keys: ds (type: string)
mode: hash
outputColumnNames: _col0, _col1, _col2
- Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 949 Basic stats: COMPLETE Column stats: COMPLETE
File Output Operator
compressed: false
table:
@@ -432,7 +432,7 @@ STAGE PLANS:
key expressions: _col0 (type: string)
sort order: +
Map-reduce partition columns: _col0 (type: string)
- Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 949 Basic stats: COMPLETE Column stats: COMPLETE
value expressions: _col1 (type: struct<columntype:string,min:bigint,max:bigint,countnulls:bigint,bitvector:binary>), _col2 (type: struct<columntype:string,maxlength:bigint,sumlength:bigint,count:bigint,countnulls:bigint,bitvector:binary>)
Reduce Operator Tree:
Group By Operator
@@ -440,14 +440,14 @@ STAGE PLANS:
keys: KEY._col0 (type: string)
mode: mergepartial
outputColumnNames: _col0, _col1, _col2
- Statistics: Num rows: 250 Data size: 2656 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 965 Basic stats: COMPLETE Column stats: COMPLETE
Select Operator
expressions: _col1 (type: struct<columntype:string,min:bigint,max:bigint,countnulls:bigint,numdistinctvalues:bigint,ndvbitvector:binary>), _col2 (type: struct<columntype:string,maxlength:bigint,avglength:double,countnulls:bigint,numdistinctvalues:bigint,ndvbitvector:binary>), _col0 (type: string)
outputColumnNames: _col0, _col1, _col2
- Statistics: Num rows: 250 Data size: 2656 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 965 Basic stats: COMPLETE Column stats: COMPLETE
File Output Operator
compressed: false
- Statistics: Num rows: 250 Data size: 2656 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 965 Basic stats: COMPLETE Column stats: COMPLETE
table:
input format: org.apache.hadoop.mapred.SequenceFileInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
@@ -485,14 +485,14 @@ STAGE PLANS:
TableScan
alias: a
filterExpr: (ds = '1') (type: boolean)
- Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE
Select Operator
expressions: key (type: int), concat(value, value) (type: string)
outputColumnNames: _col0, _col1
- Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 500 Data size: 94000 Basic stats: COMPLETE Column stats: COMPLETE
File Output Operator
compressed: false
- Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 500 Data size: 94000 Basic stats: COMPLETE Column stats: COMPLETE
table:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
http://git-wip-us.apache.org/repos/asf/hive/blob/268a6e5a/ql/src/test/results/clientpositive/bucketsortoptimize_insert_3.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/bucketsortoptimize_insert_3.q.out b/ql/src/test/results/clientpositive/bucketsortoptimize_insert_3.q.out
index bb7d1dd..a4f414c 100644
--- a/ql/src/test/results/clientpositive/bucketsortoptimize_insert_3.q.out
+++ b/ql/src/test/results/clientpositive/bucketsortoptimize_insert_3.q.out
@@ -58,14 +58,14 @@ STAGE PLANS:
TableScan
alias: a
filterExpr: (ds = '1') (type: boolean)
- Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE
Select Operator
expressions: value (type: string), key (type: int)
outputColumnNames: _col0, _col1
- Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE
File Output Operator
compressed: false
- Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE
table:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
@@ -178,25 +178,25 @@ STAGE PLANS:
TableScan
alias: a
filterExpr: (ds = '1') (type: boolean)
- Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE
Select Operator
expressions: key (type: int), value (type: string)
outputColumnNames: _col0, _col1
- Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE
Reduce Output Operator
key expressions: UDFToInteger(_col1) (type: int)
sort order: +
Map-reduce partition columns: UDFToInteger(_col1) (type: int)
- Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE
value expressions: _col0 (type: int), _col1 (type: string)
Reduce Operator Tree:
Select Operator
expressions: CAST( VALUE._col0 AS STRING) (type: string), UDFToInteger(VALUE._col1) (type: int)
outputColumnNames: _col0, _col1
- Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 500 Data size: 94000 Basic stats: COMPLETE Column stats: COMPLETE
File Output Operator
compressed: false
- Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 500 Data size: 94000 Basic stats: COMPLETE Column stats: COMPLETE
table:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
@@ -205,13 +205,13 @@ STAGE PLANS:
Select Operator
expressions: _col0 (type: string), _col1 (type: int), '1' (type: string)
outputColumnNames: value, key, ds
- Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 500 Data size: 136500 Basic stats: COMPLETE Column stats: COMPLETE
Group By Operator
aggregations: compute_stats(value, 'hll'), compute_stats(key, 'hll')
keys: ds (type: string)
mode: hash
outputColumnNames: _col0, _col1, _col2
- Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 949 Basic stats: COMPLETE Column stats: COMPLETE
File Output Operator
compressed: false
table:
@@ -247,7 +247,7 @@ STAGE PLANS:
key expressions: _col0 (type: string)
sort order: +
Map-reduce partition columns: _col0 (type: string)
- Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 949 Basic stats: COMPLETE Column stats: COMPLETE
value expressions: _col1 (type: struct<columntype:string,maxlength:bigint,sumlength:bigint,count:bigint,countnulls:bigint,bitvector:binary>), _col2 (type: struct<columntype:string,min:bigint,max:bigint,countnulls:bigint,bitvector:binary>)
Reduce Operator Tree:
Group By Operator
@@ -255,14 +255,14 @@ STAGE PLANS:
keys: KEY._col0 (type: string)
mode: mergepartial
outputColumnNames: _col0, _col1, _col2
- Statistics: Num rows: 250 Data size: 2656 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 965 Basic stats: COMPLETE Column stats: COMPLETE
Select Operator
expressions: _col1 (type: struct<columntype:string,maxlength:bigint,avglength:double,countnulls:bigint,numdistinctvalues:bigint,ndvbitvector:binary>), _col2 (type: struct<columntype:string,min:bigint,max:bigint,countnulls:bigint,numdistinctvalues:bigint,ndvbitvector:binary>), _col0 (type: string)
outputColumnNames: _col0, _col1, _col2
- Statistics: Num rows: 250 Data size: 2656 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 965 Basic stats: COMPLETE Column stats: COMPLETE
File Output Operator
compressed: false
- Statistics: Num rows: 250 Data size: 2656 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 965 Basic stats: COMPLETE Column stats: COMPLETE
table:
input format: org.apache.hadoop.mapred.SequenceFileInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
http://git-wip-us.apache.org/repos/asf/hive/blob/268a6e5a/ql/src/test/results/clientpositive/bucketsortoptimize_insert_4.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/bucketsortoptimize_insert_4.q.out b/ql/src/test/results/clientpositive/bucketsortoptimize_insert_4.q.out
index 3889804..f0552b6 100644
--- a/ql/src/test/results/clientpositive/bucketsortoptimize_insert_4.q.out
+++ b/ql/src/test/results/clientpositive/bucketsortoptimize_insert_4.q.out
@@ -87,14 +87,14 @@ STAGE PLANS:
TableScan
alias: a
filterExpr: ((ds = '1') and key is not null) (type: boolean)
- Statistics: Num rows: 10 Data size: 70 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 10 Data size: 930 Basic stats: COMPLETE Column stats: COMPLETE
Filter Operator
predicate: key is not null (type: boolean)
- Statistics: Num rows: 10 Data size: 70 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 10 Data size: 930 Basic stats: COMPLETE Column stats: COMPLETE
Select Operator
expressions: key (type: int), value (type: string)
outputColumnNames: _col0, _col1
- Statistics: Num rows: 10 Data size: 70 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 10 Data size: 930 Basic stats: COMPLETE Column stats: COMPLETE
Sorted Merge Bucket Map Join Operator
condition map:
Inner Join 0 to 1
@@ -299,14 +299,14 @@ STAGE PLANS:
TableScan
alias: a
filterExpr: ((ds = '1') and key is not null) (type: boolean)
- Statistics: Num rows: 10 Data size: 70 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 10 Data size: 930 Basic stats: COMPLETE Column stats: COMPLETE
Filter Operator
predicate: key is not null (type: boolean)
- Statistics: Num rows: 10 Data size: 70 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 10 Data size: 930 Basic stats: COMPLETE Column stats: COMPLETE
Select Operator
expressions: key (type: int), value (type: string)
outputColumnNames: _col0, _col1
- Statistics: Num rows: 10 Data size: 70 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 10 Data size: 930 Basic stats: COMPLETE Column stats: COMPLETE
Sorted Merge Bucket Map Join Operator
condition map:
Inner Join 0 to 1
http://git-wip-us.apache.org/repos/asf/hive/blob/268a6e5a/ql/src/test/results/clientpositive/bucketsortoptimize_insert_5.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/bucketsortoptimize_insert_5.q.out b/ql/src/test/results/clientpositive/bucketsortoptimize_insert_5.q.out
index 6baa706..f141fee 100644
--- a/ql/src/test/results/clientpositive/bucketsortoptimize_insert_5.q.out
+++ b/ql/src/test/results/clientpositive/bucketsortoptimize_insert_5.q.out
@@ -87,14 +87,14 @@ STAGE PLANS:
TableScan
alias: a
filterExpr: ((ds = '1') and key is not null) (type: boolean)
- Statistics: Num rows: 10 Data size: 70 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 10 Data size: 930 Basic stats: COMPLETE Column stats: COMPLETE
Filter Operator
predicate: key is not null (type: boolean)
- Statistics: Num rows: 10 Data size: 70 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 10 Data size: 930 Basic stats: COMPLETE Column stats: COMPLETE
Select Operator
expressions: key (type: int), value (type: string)
outputColumnNames: _col0, _col1
- Statistics: Num rows: 10 Data size: 70 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 10 Data size: 930 Basic stats: COMPLETE Column stats: COMPLETE
Sorted Merge Bucket Map Join Operator
condition map:
Inner Join 0 to 1
@@ -286,14 +286,14 @@ STAGE PLANS:
TableScan
alias: test_table1_n8
filterExpr: ((ds = '1') and key is not null) (type: boolean)
- Statistics: Num rows: 10 Data size: 70 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 10 Data size: 930 Basic stats: COMPLETE Column stats: COMPLETE
Filter Operator
predicate: key is not null (type: boolean)
- Statistics: Num rows: 10 Data size: 70 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 10 Data size: 930 Basic stats: COMPLETE Column stats: COMPLETE
Select Operator
expressions: key (type: int), value (type: string)
outputColumnNames: _col0, _col1
- Statistics: Num rows: 10 Data size: 70 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 10 Data size: 930 Basic stats: COMPLETE Column stats: COMPLETE
Sorted Merge Bucket Map Join Operator
condition map:
Inner Join 0 to 1
http://git-wip-us.apache.org/repos/asf/hive/blob/268a6e5a/ql/src/test/results/clientpositive/bucketsortoptimize_insert_8.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/bucketsortoptimize_insert_8.q.out b/ql/src/test/results/clientpositive/bucketsortoptimize_insert_8.q.out
index 5cd668f..f3081fd 100644
--- a/ql/src/test/results/clientpositive/bucketsortoptimize_insert_8.q.out
+++ b/ql/src/test/results/clientpositive/bucketsortoptimize_insert_8.q.out
@@ -87,14 +87,14 @@ STAGE PLANS:
TableScan
alias: a
filterExpr: ((ds = '1') and key is not null) (type: boolean)
- Statistics: Num rows: 10 Data size: 70 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 10 Data size: 930 Basic stats: COMPLETE Column stats: COMPLETE
Filter Operator
predicate: key is not null (type: boolean)
- Statistics: Num rows: 10 Data size: 70 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 10 Data size: 930 Basic stats: COMPLETE Column stats: COMPLETE
Select Operator
expressions: key (type: int), value (type: string)
outputColumnNames: _col0, _col1
- Statistics: Num rows: 10 Data size: 70 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 10 Data size: 930 Basic stats: COMPLETE Column stats: COMPLETE
Sorted Merge Bucket Map Join Operator
condition map:
Inner Join 0 to 1
@@ -281,14 +281,14 @@ STAGE PLANS:
TableScan
alias: a
filterExpr: ((ds = '1') and key is not null) (type: boolean)
- Statistics: Num rows: 10 Data size: 70 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 10 Data size: 930 Basic stats: COMPLETE Column stats: COMPLETE
Filter Operator
predicate: key is not null (type: boolean)
- Statistics: Num rows: 10 Data size: 70 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 10 Data size: 930 Basic stats: COMPLETE Column stats: COMPLETE
Select Operator
expressions: key (type: int), value (type: string)
outputColumnNames: _col0, _col1
- Statistics: Num rows: 10 Data size: 70 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 10 Data size: 930 Basic stats: COMPLETE Column stats: COMPLETE
Sorted Merge Bucket Map Join Operator
condition map:
Inner Join 0 to 1
http://git-wip-us.apache.org/repos/asf/hive/blob/268a6e5a/ql/src/test/results/clientpositive/case_sensitivity.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/case_sensitivity.q.out b/ql/src/test/results/clientpositive/case_sensitivity.q.out
index cd16098..6df0b1b 100644
--- a/ql/src/test/results/clientpositive/case_sensitivity.q.out
+++ b/ql/src/test/results/clientpositive/case_sensitivity.q.out
@@ -35,17 +35,17 @@ STAGE PLANS:
TableScan
alias: src_thrift
filterExpr: (lint[0] > 0) (type: boolean)
- Statistics: Num rows: 11 Data size: 30700 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 11 Data size: 29480 Basic stats: COMPLETE Column stats: NONE
Filter Operator
predicate: (lint[0] > 0) (type: boolean)
- Statistics: Num rows: 3 Data size: 8372 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 3 Data size: 8039 Basic stats: COMPLETE Column stats: NONE
Select Operator
expressions: lint[1] (type: int), lintstring[0].mystring (type: string)
outputColumnNames: _col0, _col1
- Statistics: Num rows: 3 Data size: 8372 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 3 Data size: 8039 Basic stats: COMPLETE Column stats: NONE
File Output Operator
compressed: false
- Statistics: Num rows: 3 Data size: 8372 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 3 Data size: 8039 Basic stats: COMPLETE Column stats: NONE
table:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
@@ -54,25 +54,25 @@ STAGE PLANS:
Select Operator
expressions: _col0 (type: int), _col1 (type: string)
outputColumnNames: key, value
- Statistics: Num rows: 3 Data size: 8372 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 3 Data size: 8039 Basic stats: COMPLETE Column stats: NONE
Group By Operator
aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll')
mode: hash
outputColumnNames: _col0, _col1
- Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 3544 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
sort order:
- Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 3544 Basic stats: COMPLETE Column stats: NONE
value expressions: _col0 (type: struct<columntype:string,min:bigint,max:bigint,countnulls:bigint,bitvector:binary>), _col1 (type: struct<columntype:string,maxlength:bigint,sumlength:bigint,count:bigint,countnulls:bigint,bitvector:binary>)
Reduce Operator Tree:
Group By Operator
aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1)
mode: mergepartial
outputColumnNames: _col0, _col1
- Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 3560 Basic stats: COMPLETE Column stats: NONE
File Output Operator
compressed: false
- Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 3560 Basic stats: COMPLETE Column stats: NONE
table:
input format: org.apache.hadoop.mapred.SequenceFileInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
http://git-wip-us.apache.org/repos/asf/hive/blob/268a6e5a/ql/src/test/results/clientpositive/cast1.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/cast1.q.out b/ql/src/test/results/clientpositive/cast1.q.out
index 5f6e00c..1828773 100644
--- a/ql/src/test/results/clientpositive/cast1.q.out
+++ b/ql/src/test/results/clientpositive/cast1.q.out
@@ -33,17 +33,17 @@ STAGE PLANS:
TableScan
alias: src
filterExpr: (UDFToDouble(key) = 86.0D) (type: boolean)
- Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE
Filter Operator
predicate: (UDFToDouble(key) = 86.0D) (type: boolean)
- Statistics: Num rows: 250 Data size: 2656 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 250 Data size: 21750 Basic stats: COMPLETE Column stats: COMPLETE
Select Operator
expressions: 5 (type: int), 5.0D (type: double), 5.0D (type: double), 5.0D (type: double), 5 (type: int), 'TRUE' (type: string), 1 (type: int)
outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6
- Statistics: Num rows: 250 Data size: 2656 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 250 Data size: 31000 Basic stats: COMPLETE Column stats: COMPLETE
File Output Operator
compressed: false
- Statistics: Num rows: 250 Data size: 2656 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 250 Data size: 31000 Basic stats: COMPLETE Column stats: COMPLETE
table:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
@@ -52,25 +52,25 @@ STAGE PLANS:
Select Operator
expressions: _col0 (type: int), _col1 (type: double), _col2 (type: double), _col3 (type: double), _col4 (type: int), _col5 (type: string), _col6 (type: int)
outputColumnNames: c1, c2, c3, c4, c5, c6, c7
- Statistics: Num rows: 250 Data size: 2656 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 250 Data size: 31000 Basic stats: COMPLETE Column stats: COMPLETE
Group By Operator
aggregations: compute_stats(c1, 'hll'), compute_stats(c2, 'hll'), compute_stats(c3, 'hll'), compute_stats(c4, 'hll'), compute_stats(c5, 'hll'), compute_stats(c6, 'hll'), compute_stats(c7, 'hll')
mode: hash
outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6
- Statistics: Num rows: 1 Data size: 2984 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 2984 Basic stats: COMPLETE Column stats: COMPLETE
Reduce Output Operator
sort order:
- Statistics: Num rows: 1 Data size: 2984 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 2984 Basic stats: COMPLETE Column stats: COMPLETE
value expressions: _col0 (type: struct<columntype:string,min:bigint,max:bigint,countnulls:bigint,bitvector:binary>), _col1 (type: struct<columntype:string,min:double,max:double,countnulls:bigint,bitvector:binary>), _col2 (type: struct<columntype:string,min:double,max:double,countnulls:bigint,bitvector:binary>), _col3 (type: struct<columntype:string,min:double,max:double,countnulls:bigint,bitvector:binary>), _col4 (type: struct<columntype:string,min:bigint,max:bigint,countnulls:bigint,bitvector:binary>), _col5 (type: struct<columntype:string,maxlength:bigint,sumlength:bigint,count:bigint,countnulls:bigint,bitvector:binary>), _col6 (type: struct<columntype:string,min:bigint,max:bigint,countnulls:bigint,bitvector:binary>)
Reduce Operator Tree:
Group By Operator
aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2), compute_stats(VALUE._col3), compute_stats(VALUE._col4), compute_stats(VALUE._col5), compute_stats(VALUE._col6)
mode: mergepartial
outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6
- Statistics: Num rows: 1 Data size: 3080 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 3080 Basic stats: COMPLETE Column stats: COMPLETE
File Output Operator
compressed: false
- Statistics: Num rows: 1 Data size: 3080 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 3080 Basic stats: COMPLETE Column stats: COMPLETE
table:
input format: org.apache.hadoop.mapred.SequenceFileInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
http://git-wip-us.apache.org/repos/asf/hive/blob/268a6e5a/ql/src/test/results/clientpositive/cast_on_constant.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/cast_on_constant.q.out b/ql/src/test/results/clientpositive/cast_on_constant.q.out
index 1541304..0279796 100644
--- a/ql/src/test/results/clientpositive/cast_on_constant.q.out
+++ b/ql/src/test/results/clientpositive/cast_on_constant.q.out
@@ -25,17 +25,17 @@ STAGE PLANS:
TableScan
alias: t1_n138
filterExpr: (ts_field = TIMESTAMP'2016-01-23 00:00:00') (type: boolean)
- Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL Column stats: NONE
+ Statistics: Num rows: 1 Data size: 96 Basic stats: COMPLETE Column stats: NONE
Filter Operator
predicate: (ts_field = TIMESTAMP'2016-01-23 00:00:00') (type: boolean)
- Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL Column stats: NONE
+ Statistics: Num rows: 1 Data size: 96 Basic stats: COMPLETE Column stats: NONE
Select Operator
expressions: TIMESTAMP'2016-01-23 00:00:00' (type: timestamp), date_field (type: date)
outputColumnNames: _col0, _col1
- Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL Column stats: NONE
+ Statistics: Num rows: 1 Data size: 96 Basic stats: COMPLETE Column stats: NONE
File Output Operator
compressed: false
- Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL Column stats: NONE
+ Statistics: Num rows: 1 Data size: 96 Basic stats: COMPLETE Column stats: NONE
table:
input format: org.apache.hadoop.mapred.SequenceFileInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
@@ -67,17 +67,17 @@ STAGE PLANS:
TableScan
alias: t1_n138
filterExpr: (date_field = DATE'2016-01-23') (type: boolean)
- Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL Column stats: NONE
+ Statistics: Num rows: 1 Data size: 96 Basic stats: COMPLETE Column stats: NONE
Filter Operator
predicate: (date_field = DATE'2016-01-23') (type: boolean)
- Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL Column stats: NONE
+ Statistics: Num rows: 1 Data size: 96 Basic stats: COMPLETE Column stats: NONE
Select Operator
expressions: ts_field (type: timestamp), DATE'2016-01-23' (type: date)
outputColumnNames: _col0, _col1
- Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL Column stats: NONE
+ Statistics: Num rows: 1 Data size: 96 Basic stats: COMPLETE Column stats: NONE
File Output Operator
compressed: false
- Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL Column stats: NONE
+ Statistics: Num rows: 1 Data size: 96 Basic stats: COMPLETE Column stats: NONE
table:
input format: org.apache.hadoop.mapred.SequenceFileInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
@@ -109,17 +109,17 @@ STAGE PLANS:
TableScan
alias: t1_n138
filterExpr: (ts_field = TIMESTAMP'2016-01-23 00:00:00') (type: boolean)
- Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL Column stats: NONE
+ Statistics: Num rows: 1 Data size: 96 Basic stats: COMPLETE Column stats: NONE
Filter Operator
predicate: (ts_field = TIMESTAMP'2016-01-23 00:00:00') (type: boolean)
- Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL Column stats: NONE
+ Statistics: Num rows: 1 Data size: 96 Basic stats: COMPLETE Column stats: NONE
Select Operator
expressions: TIMESTAMP'2016-01-23 00:00:00' (type: timestamp), date_field (type: date)
outputColumnNames: _col0, _col1
- Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL Column stats: NONE
+ Statistics: Num rows: 1 Data size: 96 Basic stats: COMPLETE Column stats: NONE
File Output Operator
compressed: false
- Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL Column stats: NONE
+ Statistics: Num rows: 1 Data size: 96 Basic stats: COMPLETE Column stats: NONE
table:
input format: org.apache.hadoop.mapred.SequenceFileInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
@@ -151,17 +151,17 @@ STAGE PLANS:
TableScan
alias: t1_n138
filterExpr: (date_field = DATE'2016-01-23') (type: boolean)
- Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL Column stats: NONE
+ Statistics: Num rows: 1 Data size: 96 Basic stats: COMPLETE Column stats: NONE
Filter Operator
predicate: (date_field = DATE'2016-01-23') (type: boolean)
- Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL Column stats: NONE
+ Statistics: Num rows: 1 Data size: 96 Basic stats: COMPLETE Column stats: NONE
Select Operator
expressions: ts_field (type: timestamp), DATE'2016-01-23' (type: date)
outputColumnNames: _col0, _col1
- Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL Column stats: NONE
+ Statistics: Num rows: 1 Data size: 96 Basic stats: COMPLETE Column stats: NONE
File Output Operator
compressed: false
- Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL Column stats: NONE
+ Statistics: Num rows: 1 Data size: 96 Basic stats: COMPLETE Column stats: NONE
table:
input format: org.apache.hadoop.mapred.SequenceFileInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
@@ -193,17 +193,17 @@ STAGE PLANS:
TableScan
alias: t1_n138
filterExpr: (CAST( date_field AS TIMESTAMP) = ts_field) (type: boolean)
- Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL Column stats: NONE
+ Statistics: Num rows: 1 Data size: 96 Basic stats: COMPLETE Column stats: NONE
Filter Operator
predicate: (CAST( date_field AS TIMESTAMP) = ts_field) (type: boolean)
- Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL Column stats: NONE
+ Statistics: Num rows: 1 Data size: 96 Basic stats: COMPLETE Column stats: NONE
Select Operator
expressions: ts_field (type: timestamp), date_field (type: date)
outputColumnNames: _col0, _col1
- Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL Column stats: NONE
+ Statistics: Num rows: 1 Data size: 96 Basic stats: COMPLETE Column stats: NONE
File Output Operator
compressed: false
- Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL Column stats: NONE
+ Statistics: Num rows: 1 Data size: 96 Basic stats: COMPLETE Column stats: NONE
table:
input format: org.apache.hadoop.mapred.SequenceFileInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat