You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hive.apache.org by px...@apache.org on 2016/05/09 17:41:57 UTC
[15/21] hive git commit: HIVE-13341: Stats state is not captured
correctly: differentiate load table and create table (Pengcheng Xiong,
reviewed by Ashutosh Chauhan)
http://git-wip-us.apache.org/repos/asf/hive/blob/244ce09c/ql/src/test/results/clientpositive/list_bucket_dml_8.q.java1.7.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/list_bucket_dml_8.q.java1.7.out b/ql/src/test/results/clientpositive/list_bucket_dml_8.q.java1.7.out
index d40a693..de1305f 100644
--- a/ql/src/test/results/clientpositive/list_bucket_dml_8.q.java1.7.out
+++ b/ql/src/test/results/clientpositive/list_bucket_dml_8.q.java1.7.out
@@ -439,6 +439,8 @@ Table: list_bucketing_dynamic_part
#### A masked pattern was here ####
Partition Parameters:
numFiles 3
+ numRows 984
+ rawDataSize 9488
totalSize 10586
#### A masked pattern was here ####
@@ -555,8 +557,10 @@ STAGE PLANS:
#### A masked pattern was here ####
name default.list_bucketing_dynamic_part
numFiles 3
+ numRows 984
partition_columns ds/hr
partition_columns.types string:string
+ rawDataSize 9488
serialization.ddl struct list_bucketing_dynamic_part { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe
@@ -585,16 +589,16 @@ STAGE PLANS:
Processor Tree:
TableScan
alias: list_bucketing_dynamic_part
- Statistics: Num rows: 16 Data size: 136 Basic stats: PARTIAL Column stats: NONE
+ Statistics: Num rows: 1000 Data size: 9624 Basic stats: COMPLETE Column stats: NONE
GatherStats: false
Filter Operator
isSamplingPred: false
predicate: ((key = '484') and (value = 'val_484')) (type: boolean)
- Statistics: Num rows: 4 Data size: 34 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 250 Data size: 2406 Basic stats: COMPLETE Column stats: NONE
Select Operator
expressions: '484' (type: string), 'val_484' (type: string), ds (type: string), hr (type: string)
outputColumnNames: _col0, _col1, _col2, _col3
- Statistics: Num rows: 4 Data size: 34 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 250 Data size: 2406 Basic stats: COMPLETE Column stats: NONE
ListSink
PREHOOK: query: select * from list_bucketing_dynamic_part where key = '484' and value = 'val_484'
http://git-wip-us.apache.org/repos/asf/hive/blob/244ce09c/ql/src/test/results/clientpositive/mapjoin_memcheck.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/mapjoin_memcheck.q.out b/ql/src/test/results/clientpositive/mapjoin_memcheck.q.out
index 8f5492f..65d896b 100644
--- a/ql/src/test/results/clientpositive/mapjoin_memcheck.q.out
+++ b/ql/src/test/results/clientpositive/mapjoin_memcheck.q.out
@@ -44,14 +44,14 @@ STAGE PLANS:
$hdt$_0:src1
TableScan
alias: src1
- Statistics: Num rows: 10 Data size: 70 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 80 Basic stats: COMPLETE Column stats: NONE
Filter Operator
predicate: key is not null (type: boolean)
- Statistics: Num rows: 10 Data size: 70 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 80 Basic stats: COMPLETE Column stats: NONE
Select Operator
expressions: key (type: string), value (type: string)
outputColumnNames: _col0, _col1
- Statistics: Num rows: 10 Data size: 70 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 80 Basic stats: COMPLETE Column stats: NONE
HashTable Sink Operator
keys:
0 _col0 (type: string)
@@ -62,14 +62,14 @@ STAGE PLANS:
Map Operator Tree:
TableScan
alias: src1
- Statistics: Num rows: 10 Data size: 70 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 80 Basic stats: COMPLETE Column stats: NONE
Filter Operator
predicate: key is not null (type: boolean)
- Statistics: Num rows: 10 Data size: 70 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 80 Basic stats: COMPLETE Column stats: NONE
Select Operator
expressions: key (type: string), value (type: string)
outputColumnNames: _col0, _col1
- Statistics: Num rows: 10 Data size: 70 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 80 Basic stats: COMPLETE Column stats: NONE
Map Join Operator
condition map:
Inner Join 0 to 1
@@ -77,10 +77,10 @@ STAGE PLANS:
0 _col0 (type: string)
1 _col0 (type: string)
outputColumnNames: _col0, _col1, _col2, _col3
- Statistics: Num rows: 11 Data size: 77 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 88 Basic stats: COMPLETE Column stats: NONE
File Output Operator
compressed: false
- Statistics: Num rows: 11 Data size: 77 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 88 Basic stats: COMPLETE Column stats: NONE
table:
input format: org.apache.hadoop.mapred.SequenceFileInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
http://git-wip-us.apache.org/repos/asf/hive/blob/244ce09c/ql/src/test/results/clientpositive/metadataonly1.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/metadataonly1.q.out b/ql/src/test/results/clientpositive/metadataonly1.q.out
index 0281407..4b4c57c 100644
--- a/ql/src/test/results/clientpositive/metadataonly1.q.out
+++ b/ql/src/test/results/clientpositive/metadataonly1.q.out
@@ -132,17 +132,22 @@ STAGE PLANS:
partition values:
ds 1
properties:
+ COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
bucket_count -1
columns a,b
columns.comments
columns.types int:double
#### A masked pattern was here ####
name default.test1
+ numFiles 0
+ numRows 0
partition_columns ds
partition_columns.types string
+ rawDataSize 0
serialization.ddl struct test1 { i32 a, double b}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.NullStructSerDe
+ totalSize 0
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.NullStructSerDe
@@ -256,17 +261,22 @@ STAGE PLANS:
partition values:
ds 1
properties:
+ COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
bucket_count -1
columns a,b
columns.comments
columns.types int:double
#### A masked pattern was here ####
name default.test1
+ numFiles 0
+ numRows 0
partition_columns ds
partition_columns.types string
+ rawDataSize 0
serialization.ddl struct test1 { i32 a, double b}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.NullStructSerDe
+ totalSize 0
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.NullStructSerDe
@@ -380,17 +390,22 @@ STAGE PLANS:
partition values:
ds 1
properties:
+ COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
bucket_count -1
columns a,b
columns.comments
columns.types int:double
#### A masked pattern was here ####
name default.test1
+ numFiles 0
+ numRows 0
partition_columns ds
partition_columns.types string
+ rawDataSize 0
serialization.ddl struct test1 { i32 a, double b}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ totalSize 0
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -515,17 +530,22 @@ STAGE PLANS:
partition values:
ds 1
properties:
+ COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
bucket_count -1
columns a,b
columns.comments
columns.types int:double
#### A masked pattern was here ####
name default.test1
+ numFiles 0
+ numRows 0
partition_columns ds
partition_columns.types string
+ rawDataSize 0
serialization.ddl struct test1 { i32 a, double b}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.NullStructSerDe
+ totalSize 0
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.NullStructSerDe
@@ -554,17 +574,22 @@ STAGE PLANS:
partition values:
ds 2
properties:
+ COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
bucket_count -1
columns a,b
columns.comments
columns.types int:double
#### A masked pattern was here ####
name default.test1
+ numFiles 0
+ numRows 0
partition_columns ds
partition_columns.types string
+ rawDataSize 0
serialization.ddl struct test1 { i32 a, double b}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.NullStructSerDe
+ totalSize 0
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.NullStructSerDe
@@ -678,17 +703,22 @@ STAGE PLANS:
partition values:
ds 1
properties:
+ COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
bucket_count -1
columns a,b
columns.comments
columns.types int:double
#### A masked pattern was here ####
name default.test1
+ numFiles 0
+ numRows 0
partition_columns ds
partition_columns.types string
+ rawDataSize 0
serialization.ddl struct test1 { i32 a, double b}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ totalSize 0
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -718,17 +748,22 @@ STAGE PLANS:
partition values:
ds 2
properties:
+ COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
bucket_count -1
columns a,b
columns.comments
columns.types int:double
#### A masked pattern was here ####
name default.test1
+ numFiles 0
+ numRows 0
partition_columns ds
partition_columns.types string
+ rawDataSize 0
serialization.ddl struct test1 { i32 a, double b}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ totalSize 0
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -948,17 +983,22 @@ STAGE PLANS:
ds 1
hr 1
properties:
+ COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
bucket_count -1
columns a,b
columns.comments
columns.types int:double
#### A masked pattern was here ####
name default.test2
+ numFiles 0
+ numRows 0
partition_columns ds/hr
partition_columns.types string:string
+ rawDataSize 0
serialization.ddl struct test2 { i32 a, double b}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.NullStructSerDe
+ totalSize 0
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.NullStructSerDe
@@ -988,17 +1028,22 @@ STAGE PLANS:
ds 1
hr 2
properties:
+ COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
bucket_count -1
columns a,b
columns.comments
columns.types int:double
#### A masked pattern was here ####
name default.test2
+ numFiles 0
+ numRows 0
partition_columns ds/hr
partition_columns.types string:string
+ rawDataSize 0
serialization.ddl struct test2 { i32 a, double b}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.NullStructSerDe
+ totalSize 0
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.NullStructSerDe
@@ -1028,17 +1073,22 @@ STAGE PLANS:
ds 1
hr 3
properties:
+ COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
bucket_count -1
columns a,b
columns.comments
columns.types int:double
#### A masked pattern was here ####
name default.test2
+ numFiles 0
+ numRows 0
partition_columns ds/hr
partition_columns.types string:string
+ rawDataSize 0
serialization.ddl struct test2 { i32 a, double b}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.NullStructSerDe
+ totalSize 0
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.NullStructSerDe
@@ -1163,17 +1213,22 @@ STAGE PLANS:
ds 1
hr 1
properties:
+ COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
bucket_count -1
columns a,b
columns.comments
columns.types int:double
#### A masked pattern was here ####
name default.test2
+ numFiles 0
+ numRows 0
partition_columns ds/hr
partition_columns.types string:string
+ rawDataSize 0
serialization.ddl struct test2 { i32 a, double b}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ totalSize 0
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -1204,17 +1259,22 @@ STAGE PLANS:
ds 1
hr 2
properties:
+ COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
bucket_count -1
columns a,b
columns.comments
columns.types int:double
#### A masked pattern was here ####
name default.test2
+ numFiles 0
+ numRows 0
partition_columns ds/hr
partition_columns.types string:string
+ rawDataSize 0
serialization.ddl struct test2 { i32 a, double b}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ totalSize 0
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -1245,17 +1305,22 @@ STAGE PLANS:
ds 1
hr 3
properties:
+ COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
bucket_count -1
columns a,b
columns.comments
columns.types int:double
#### A masked pattern was here ####
name default.test2
+ numFiles 0
+ numRows 0
partition_columns ds/hr
partition_columns.types string:string
+ rawDataSize 0
serialization.ddl struct test2 { i32 a, double b}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ totalSize 0
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -1375,17 +1440,22 @@ STAGE PLANS:
partition values:
ds 1
properties:
+ COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
bucket_count -1
columns a,b
columns.comments
columns.types int:double
#### A masked pattern was here ####
name default.test1
+ numFiles 0
+ numRows 0
partition_columns ds
partition_columns.types string
+ rawDataSize 0
serialization.ddl struct test1 { i32 a, double b}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.NullStructSerDe
+ totalSize 0
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.NullStructSerDe
@@ -1414,17 +1484,22 @@ STAGE PLANS:
partition values:
ds 2
properties:
+ COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
bucket_count -1
columns a,b
columns.comments
columns.types int:double
#### A masked pattern was here ####
name default.test1
+ numFiles 0
+ numRows 0
partition_columns ds
partition_columns.types string
+ rawDataSize 0
serialization.ddl struct test1 { i32 a, double b}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.NullStructSerDe
+ totalSize 0
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.NullStructSerDe
@@ -1598,17 +1673,22 @@ STAGE PLANS:
ds 01:10:10
hr 01
properties:
+ COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
bucket_count -1
columns a,b
columns.comments
columns.types int:double
#### A masked pattern was here ####
name default.test2
+ numFiles 0
+ numRows 0
partition_columns ds/hr
partition_columns.types string:string
+ rawDataSize 0
serialization.ddl struct test2 { i32 a, double b}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.NullStructSerDe
+ totalSize 0
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.NullStructSerDe
@@ -1638,17 +1718,22 @@ STAGE PLANS:
ds 01:10:20
hr 02
properties:
+ COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
bucket_count -1
columns a,b
columns.comments
columns.types int:double
#### A masked pattern was here ####
name default.test2
+ numFiles 0
+ numRows 0
partition_columns ds/hr
partition_columns.types string:string
+ rawDataSize 0
serialization.ddl struct test2 { i32 a, double b}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.NullStructSerDe
+ totalSize 0
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.NullStructSerDe
@@ -1678,17 +1763,22 @@ STAGE PLANS:
ds 1
hr 1
properties:
+ COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
bucket_count -1
columns a,b
columns.comments
columns.types int:double
#### A masked pattern was here ####
name default.test2
+ numFiles 0
+ numRows 0
partition_columns ds/hr
partition_columns.types string:string
+ rawDataSize 0
serialization.ddl struct test2 { i32 a, double b}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.NullStructSerDe
+ totalSize 0
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.NullStructSerDe
@@ -1718,17 +1808,22 @@ STAGE PLANS:
ds 1
hr 2
properties:
+ COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
bucket_count -1
columns a,b
columns.comments
columns.types int:double
#### A masked pattern was here ####
name default.test2
+ numFiles 0
+ numRows 0
partition_columns ds/hr
partition_columns.types string:string
+ rawDataSize 0
serialization.ddl struct test2 { i32 a, double b}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.NullStructSerDe
+ totalSize 0
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.NullStructSerDe
@@ -1758,17 +1853,22 @@ STAGE PLANS:
ds 1
hr 3
properties:
+ COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
bucket_count -1
columns a,b
columns.comments
columns.types int:double
#### A masked pattern was here ####
name default.test2
+ numFiles 0
+ numRows 0
partition_columns ds/hr
partition_columns.types string:string
+ rawDataSize 0
serialization.ddl struct test2 { i32 a, double b}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.NullStructSerDe
+ totalSize 0
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.NullStructSerDe
http://git-wip-us.apache.org/repos/asf/hive/blob/244ce09c/ql/src/test/results/clientpositive/nullformat.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/nullformat.q.out b/ql/src/test/results/clientpositive/nullformat.q.out
index af91470..2a80359 100644
--- a/ql/src/test/results/clientpositive/nullformat.q.out
+++ b/ql/src/test/results/clientpositive/nullformat.q.out
@@ -93,6 +93,11 @@ OUTPUTFORMAT
LOCATION
#### A masked pattern was here ####
TBLPROPERTIES (
+ 'COLUMN_STATS_ACCURATE'='{\"BASIC_STATS\":\"true\"}',
+ 'numFiles'='0',
+ 'numRows'='0',
+ 'rawDataSize'='0',
+ 'totalSize'='0',
#### A masked pattern was here ####
PREHOOK: query: -- load null data from another table and verify that the null is stored in the expected format
INSERT OVERWRITE TABLE null_tab1 SELECT a,b FROM base_tab
http://git-wip-us.apache.org/repos/asf/hive/blob/244ce09c/ql/src/test/results/clientpositive/orc_create.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/orc_create.q.out b/ql/src/test/results/clientpositive/orc_create.q.out
index 3013fda..20c3fce 100644
--- a/ql/src/test/results/clientpositive/orc_create.q.out
+++ b/ql/src/test/results/clientpositive/orc_create.q.out
@@ -70,6 +70,11 @@ Retention: 0
#### A masked pattern was here ####
Table Type: MANAGED_TABLE
Table Parameters:
+ COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\"}
+ numFiles 0
+ numRows 0
+ rawDataSize 0
+ totalSize 0
#### A masked pattern was here ####
# Storage Information
@@ -319,6 +324,11 @@ Retention: 0
#### A masked pattern was here ####
Table Type: MANAGED_TABLE
Table Parameters:
+ COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\"}
+ numFiles 0
+ numRows 0
+ rawDataSize 0
+ totalSize 0
#### A masked pattern was here ####
# Storage Information
http://git-wip-us.apache.org/repos/asf/hive/blob/244ce09c/ql/src/test/results/clientpositive/orc_llap.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/orc_llap.q.out b/ql/src/test/results/clientpositive/orc_llap.q.out
index 6fc73b7..bae69bb 100644
--- a/ql/src/test/results/clientpositive/orc_llap.q.out
+++ b/ql/src/test/results/clientpositive/orc_llap.q.out
@@ -719,17 +719,17 @@ STAGE PLANS:
TableScan
alias: orc_llap
filterExpr: ((cint > 10) and cbigint is not null) (type: boolean)
- Statistics: Num rows: 98779 Data size: 1580469 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 245760 Data size: 58159880 Basic stats: COMPLETE Column stats: NONE
Filter Operator
predicate: ((cint > 10) and cbigint is not null) (type: boolean)
- Statistics: Num rows: 32926 Data size: 526817 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 81920 Data size: 19386626 Basic stats: COMPLETE Column stats: NONE
Select Operator
expressions: cint (type: int), csmallint (type: smallint), cbigint (type: bigint)
outputColumnNames: _col0, _col1, _col2
- Statistics: Num rows: 32926 Data size: 526817 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 81920 Data size: 19386626 Basic stats: COMPLETE Column stats: NONE
File Output Operator
compressed: false
- Statistics: Num rows: 32926 Data size: 526817 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 81920 Data size: 19386626 Basic stats: COMPLETE Column stats: NONE
table:
input format: org.apache.hadoop.mapred.SequenceFileInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
@@ -792,17 +792,17 @@ STAGE PLANS:
TableScan
alias: orc_llap
filterExpr: ((cint > 10) and cbigint is not null) (type: boolean)
- Statistics: Num rows: 4938 Data size: 1580469 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 245760 Data size: 58159880 Basic stats: COMPLETE Column stats: NONE
Filter Operator
predicate: ((cint > 10) and cbigint is not null) (type: boolean)
- Statistics: Num rows: 1646 Data size: 526823 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 81920 Data size: 19386626 Basic stats: COMPLETE Column stats: NONE
Select Operator
expressions: ctinyint (type: tinyint), csmallint (type: smallint), cint (type: int), cbigint (type: bigint), cfloat (type: float), cdouble (type: double), cstring1 (type: string), cstring2 (type: string), ctimestamp1 (type: timestamp), ctimestamp2 (type: timestamp), cboolean1 (type: boolean), cboolean2 (type: boolean)
outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11
- Statistics: Num rows: 1646 Data size: 526823 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 81920 Data size: 19386626 Basic stats: COMPLETE Column stats: NONE
File Output Operator
compressed: false
- Statistics: Num rows: 1646 Data size: 526823 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 81920 Data size: 19386626 Basic stats: COMPLETE Column stats: NONE
table:
input format: org.apache.hadoop.mapred.SequenceFileInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
@@ -874,17 +874,17 @@ STAGE PLANS:
TableScan
alias: orc_llap
filterExpr: ((cint > 5) and (cint < 10)) (type: boolean)
- Statistics: Num rows: 15196 Data size: 1580469 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 245760 Data size: 58159880 Basic stats: COMPLETE Column stats: NONE
Filter Operator
predicate: ((cint > 5) and (cint < 10)) (type: boolean)
- Statistics: Num rows: 1688 Data size: 175561 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 27306 Data size: 6462051 Basic stats: COMPLETE Column stats: NONE
Select Operator
expressions: cstring2 (type: string)
outputColumnNames: _col0
- Statistics: Num rows: 1688 Data size: 175561 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 27306 Data size: 6462051 Basic stats: COMPLETE Column stats: NONE
File Output Operator
compressed: false
- Statistics: Num rows: 1688 Data size: 175561 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 27306 Data size: 6462051 Basic stats: COMPLETE Column stats: NONE
table:
input format: org.apache.hadoop.mapred.SequenceFileInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
@@ -944,22 +944,22 @@ STAGE PLANS:
Map Operator Tree:
TableScan
alias: orc_llap
- Statistics: Num rows: 7902 Data size: 1580469 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 245760 Data size: 58159880 Basic stats: COMPLETE Column stats: NONE
Select Operator
expressions: cstring1 (type: string), cstring2 (type: string)
outputColumnNames: cstring1, cstring2
- Statistics: Num rows: 7902 Data size: 1580469 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 245760 Data size: 58159880 Basic stats: COMPLETE Column stats: NONE
Group By Operator
aggregations: count()
keys: cstring1 (type: string), cstring2 (type: string)
mode: hash
outputColumnNames: _col0, _col1, _col2
- Statistics: Num rows: 7902 Data size: 1580469 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 245760 Data size: 58159880 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
key expressions: _col0 (type: string), _col1 (type: string)
sort order: ++
Map-reduce partition columns: _col0 (type: string), _col1 (type: string)
- Statistics: Num rows: 7902 Data size: 1580469 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 245760 Data size: 58159880 Basic stats: COMPLETE Column stats: NONE
value expressions: _col2 (type: bigint)
Execution mode: vectorized
LLAP IO: all inputs
@@ -969,10 +969,10 @@ STAGE PLANS:
keys: KEY._col0 (type: string), KEY._col1 (type: string)
mode: mergepartial
outputColumnNames: _col0, _col1, _col2
- Statistics: Num rows: 3951 Data size: 790234 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 122880 Data size: 29079940 Basic stats: COMPLETE Column stats: NONE
File Output Operator
compressed: false
- Statistics: Num rows: 3951 Data size: 790234 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 122880 Data size: 29079940 Basic stats: COMPLETE Column stats: NONE
table:
input format: org.apache.hadoop.mapred.SequenceFileInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
@@ -1039,14 +1039,14 @@ STAGE PLANS:
TableScan
alias: o1
filterExpr: (csmallint is not null and cbigint is not null) (type: boolean)
- Statistics: Num rows: 14111 Data size: 1580469 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 245760 Data size: 58159880 Basic stats: COMPLETE Column stats: NONE
Filter Operator
predicate: (csmallint is not null and cbigint is not null) (type: boolean)
- Statistics: Num rows: 14111 Data size: 1580469 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 245760 Data size: 58159880 Basic stats: COMPLETE Column stats: NONE
Select Operator
expressions: csmallint (type: smallint), cstring1 (type: string)
outputColumnNames: _col0, _col2
- Statistics: Num rows: 14111 Data size: 1580469 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 245760 Data size: 58159880 Basic stats: COMPLETE Column stats: NONE
HashTable Sink Operator
keys:
0 _col0 (type: smallint)
@@ -1058,14 +1058,14 @@ STAGE PLANS:
TableScan
alias: o1
filterExpr: (csmallint is not null and cbigint is not null) (type: boolean)
- Statistics: Num rows: 14111 Data size: 1580469 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 245760 Data size: 58159880 Basic stats: COMPLETE Column stats: NONE
Filter Operator
predicate: (csmallint is not null and cbigint is not null) (type: boolean)
- Statistics: Num rows: 14111 Data size: 1580469 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 245760 Data size: 58159880 Basic stats: COMPLETE Column stats: NONE
Select Operator
expressions: csmallint (type: smallint), cstring2 (type: string)
outputColumnNames: _col0, _col2
- Statistics: Num rows: 14111 Data size: 1580469 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 245760 Data size: 58159880 Basic stats: COMPLETE Column stats: NONE
Map Join Operator
condition map:
Inner Join 0 to 1
@@ -1073,14 +1073,14 @@ STAGE PLANS:
0 _col0 (type: smallint)
1 _col0 (type: smallint)
outputColumnNames: _col2, _col5
- Statistics: Num rows: 15522 Data size: 1738515 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 270336 Data size: 63975869 Basic stats: COMPLETE Column stats: NONE
Select Operator
expressions: _col2 (type: string), _col5 (type: string)
outputColumnNames: _col0, _col1
- Statistics: Num rows: 15522 Data size: 1738515 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 270336 Data size: 63975869 Basic stats: COMPLETE Column stats: NONE
File Output Operator
compressed: false
- Statistics: Num rows: 15522 Data size: 1738515 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 270336 Data size: 63975869 Basic stats: COMPLETE Column stats: NONE
table:
input format: org.apache.hadoop.mapred.SequenceFileInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
http://git-wip-us.apache.org/repos/asf/hive/blob/244ce09c/ql/src/test/results/clientpositive/orc_predicate_pushdown.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/orc_predicate_pushdown.q.out b/ql/src/test/results/clientpositive/orc_predicate_pushdown.q.out
index 7b361b7..38321e9 100644
--- a/ql/src/test/results/clientpositive/orc_predicate_pushdown.q.out
+++ b/ql/src/test/results/clientpositive/orc_predicate_pushdown.q.out
@@ -135,11 +135,11 @@ STAGE PLANS:
Map Operator Tree:
TableScan
alias: orc_pred
- Statistics: Num rows: 1049 Data size: 311170 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 6037 Data size: 24150 Basic stats: COMPLETE Column stats: NONE
Select Operator
expressions: hash(t) (type: int)
outputColumnNames: _col0
- Statistics: Num rows: 1049 Data size: 311170 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 6037 Data size: 24150 Basic stats: COMPLETE Column stats: NONE
Group By Operator
aggregations: sum(_col0)
mode: hash
@@ -183,11 +183,11 @@ STAGE PLANS:
Map Operator Tree:
TableScan
alias: orc_pred
- Statistics: Num rows: 1049 Data size: 311170 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 6037 Data size: 24150 Basic stats: COMPLETE Column stats: NONE
Select Operator
expressions: hash(t) (type: int)
outputColumnNames: _col0
- Statistics: Num rows: 1049 Data size: 311170 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 6037 Data size: 24150 Basic stats: COMPLETE Column stats: NONE
Group By Operator
aggregations: sum(_col0)
mode: hash
@@ -311,14 +311,14 @@ STAGE PLANS:
Map Operator Tree:
TableScan
alias: orc_pred
- Statistics: Num rows: 1049 Data size: 311170 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 6037 Data size: 24150 Basic stats: COMPLETE Column stats: NONE
Filter Operator
predicate: ((t < 0) and (UDFToInteger(t) > -2)) (type: boolean)
- Statistics: Num rows: 116 Data size: 34409 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 670 Data size: 2680 Basic stats: COMPLETE Column stats: NONE
Select Operator
expressions: hash(t) (type: int)
outputColumnNames: _col0
- Statistics: Num rows: 116 Data size: 34409 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 670 Data size: 2680 Basic stats: COMPLETE Column stats: NONE
Group By Operator
aggregations: sum(_col0)
mode: hash
@@ -369,14 +369,14 @@ STAGE PLANS:
TableScan
alias: orc_pred
filterExpr: ((t < 0) and (UDFToInteger(t) > -2)) (type: boolean)
- Statistics: Num rows: 1049 Data size: 311170 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 6037 Data size: 24150 Basic stats: COMPLETE Column stats: NONE
Filter Operator
predicate: ((t < 0) and (UDFToInteger(t) > -2)) (type: boolean)
- Statistics: Num rows: 116 Data size: 34409 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 670 Data size: 2680 Basic stats: COMPLETE Column stats: NONE
Select Operator
expressions: hash(t) (type: int)
outputColumnNames: _col0
- Statistics: Num rows: 116 Data size: 34409 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 670 Data size: 2680 Basic stats: COMPLETE Column stats: NONE
Group By Operator
aggregations: sum(_col0)
mode: hash
@@ -458,17 +458,17 @@ STAGE PLANS:
Map Operator Tree:
TableScan
alias: orc_pred
- Statistics: Num rows: 1049 Data size: 311170 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 232 Data size: 24150 Basic stats: COMPLETE Column stats: NONE
Filter Operator
predicate: ((t = -1) and s is not null and (s like 'bob%')) (type: boolean)
- Statistics: Num rows: 262 Data size: 77718 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 58 Data size: 6037 Basic stats: COMPLETE Column stats: NONE
Select Operator
expressions: -1 (type: tinyint), s (type: string)
outputColumnNames: _col0, _col1
- Statistics: Num rows: 262 Data size: 77718 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 58 Data size: 6037 Basic stats: COMPLETE Column stats: NONE
File Output Operator
compressed: false
- Statistics: Num rows: 262 Data size: 77718 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 58 Data size: 6037 Basic stats: COMPLETE Column stats: NONE
table:
input format: org.apache.hadoop.mapred.SequenceFileInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
@@ -501,17 +501,17 @@ STAGE PLANS:
TableScan
alias: orc_pred
filterExpr: ((t = -1) and s is not null and (s like 'bob%')) (type: boolean)
- Statistics: Num rows: 1049 Data size: 311170 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 232 Data size: 24150 Basic stats: COMPLETE Column stats: NONE
Filter Operator
predicate: ((t = -1) and s is not null and (s like 'bob%')) (type: boolean)
- Statistics: Num rows: 262 Data size: 77718 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 58 Data size: 6037 Basic stats: COMPLETE Column stats: NONE
Select Operator
expressions: -1 (type: tinyint), s (type: string)
outputColumnNames: _col0, _col1
- Statistics: Num rows: 262 Data size: 77718 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 58 Data size: 6037 Basic stats: COMPLETE Column stats: NONE
File Output Operator
compressed: false
- Statistics: Num rows: 262 Data size: 77718 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 58 Data size: 6037 Basic stats: COMPLETE Column stats: NONE
table:
input format: org.apache.hadoop.mapred.SequenceFileInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
@@ -589,26 +589,26 @@ STAGE PLANS:
Map Operator Tree:
TableScan
alias: orc_pred
- Statistics: Num rows: 1049 Data size: 311170 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 232 Data size: 24150 Basic stats: COMPLETE Column stats: NONE
Filter Operator
predicate: (s is not null and (s like 'bob%') and (not (t) IN (-1, -2, -3)) and t BETWEEN 25 AND 30) (type: boolean)
- Statistics: Num rows: 131 Data size: 38859 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 29 Data size: 3018 Basic stats: COMPLETE Column stats: NONE
Select Operator
expressions: t (type: tinyint), s (type: string)
outputColumnNames: _col0, _col1
- Statistics: Num rows: 131 Data size: 38859 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 29 Data size: 3018 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
key expressions: _col0 (type: tinyint), _col1 (type: string)
sort order: ++
- Statistics: Num rows: 131 Data size: 38859 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 29 Data size: 3018 Basic stats: COMPLETE Column stats: NONE
Reduce Operator Tree:
Select Operator
expressions: KEY.reducesinkkey0 (type: tinyint), KEY.reducesinkkey1 (type: string)
outputColumnNames: _col0, _col1
- Statistics: Num rows: 131 Data size: 38859 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 29 Data size: 3018 Basic stats: COMPLETE Column stats: NONE
File Output Operator
compressed: false
- Statistics: Num rows: 131 Data size: 38859 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 29 Data size: 3018 Basic stats: COMPLETE Column stats: NONE
table:
input format: org.apache.hadoop.mapred.SequenceFileInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
@@ -645,26 +645,26 @@ STAGE PLANS:
TableScan
alias: orc_pred
filterExpr: (s is not null and (s like 'bob%') and (not (t) IN (-1, -2, -3)) and t BETWEEN 25 AND 30) (type: boolean)
- Statistics: Num rows: 1049 Data size: 311170 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 232 Data size: 24150 Basic stats: COMPLETE Column stats: NONE
Filter Operator
predicate: (s is not null and (s like 'bob%') and (not (t) IN (-1, -2, -3)) and t BETWEEN 25 AND 30) (type: boolean)
- Statistics: Num rows: 131 Data size: 38859 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 29 Data size: 3018 Basic stats: COMPLETE Column stats: NONE
Select Operator
expressions: t (type: tinyint), s (type: string)
outputColumnNames: _col0, _col1
- Statistics: Num rows: 131 Data size: 38859 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 29 Data size: 3018 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
key expressions: _col0 (type: tinyint), _col1 (type: string)
sort order: ++
- Statistics: Num rows: 131 Data size: 38859 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 29 Data size: 3018 Basic stats: COMPLETE Column stats: NONE
Reduce Operator Tree:
Select Operator
expressions: KEY.reducesinkkey0 (type: tinyint), KEY.reducesinkkey1 (type: string)
outputColumnNames: _col0, _col1
- Statistics: Num rows: 131 Data size: 38859 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 29 Data size: 3018 Basic stats: COMPLETE Column stats: NONE
File Output Operator
compressed: false
- Statistics: Num rows: 131 Data size: 38859 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 29 Data size: 3018 Basic stats: COMPLETE Column stats: NONE
table:
input format: org.apache.hadoop.mapred.SequenceFileInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
@@ -766,31 +766,31 @@ STAGE PLANS:
Map Operator Tree:
TableScan
alias: orc_pred
- Statistics: Num rows: 1049 Data size: 311170 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 208 Data size: 24150 Basic stats: COMPLETE Column stats: NONE
Filter Operator
predicate: ((d >= 10.0) and (d < 12.0) and (s like '%son') and (t > 0) and si BETWEEN 300 AND 400 and (not (s like '%car%'))) (type: boolean)
- Statistics: Num rows: 5 Data size: 1483 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: NONE
Select Operator
expressions: t (type: tinyint), si (type: smallint), d (type: double), s (type: string)
outputColumnNames: _col0, _col1, _col2, _col3
- Statistics: Num rows: 5 Data size: 1483 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
key expressions: _col3 (type: string)
sort order: -
- Statistics: Num rows: 5 Data size: 1483 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: NONE
TopN Hash Memory Usage: 0.1
value expressions: _col0 (type: tinyint), _col1 (type: smallint), _col2 (type: double)
Reduce Operator Tree:
Select Operator
expressions: VALUE._col0 (type: tinyint), VALUE._col1 (type: smallint), VALUE._col2 (type: double), KEY.reducesinkkey0 (type: string)
outputColumnNames: _col0, _col1, _col2, _col3
- Statistics: Num rows: 5 Data size: 1483 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: NONE
Limit
Number of rows: 3
- Statistics: Num rows: 3 Data size: 888 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: NONE
File Output Operator
compressed: false
- Statistics: Num rows: 3 Data size: 888 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: NONE
table:
input format: org.apache.hadoop.mapred.SequenceFileInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
@@ -835,31 +835,31 @@ STAGE PLANS:
TableScan
alias: orc_pred
filterExpr: ((d >= 10.0) and (d < 12.0) and (s like '%son') and (t > 0) and si BETWEEN 300 AND 400 and (not (s like '%car%'))) (type: boolean)
- Statistics: Num rows: 1049 Data size: 311170 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 208 Data size: 24150 Basic stats: COMPLETE Column stats: NONE
Filter Operator
predicate: ((d >= 10.0) and (d < 12.0) and (s like '%son') and (t > 0) and si BETWEEN 300 AND 400 and (not (s like '%car%'))) (type: boolean)
- Statistics: Num rows: 5 Data size: 1483 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: NONE
Select Operator
expressions: t (type: tinyint), si (type: smallint), d (type: double), s (type: string)
outputColumnNames: _col0, _col1, _col2, _col3
- Statistics: Num rows: 5 Data size: 1483 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
key expressions: _col3 (type: string)
sort order: -
- Statistics: Num rows: 5 Data size: 1483 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: NONE
TopN Hash Memory Usage: 0.1
value expressions: _col0 (type: tinyint), _col1 (type: smallint), _col2 (type: double)
Reduce Operator Tree:
Select Operator
expressions: VALUE._col0 (type: tinyint), VALUE._col1 (type: smallint), VALUE._col2 (type: double), KEY.reducesinkkey0 (type: string)
outputColumnNames: _col0, _col1, _col2, _col3
- Statistics: Num rows: 5 Data size: 1483 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: NONE
Limit
Number of rows: 3
- Statistics: Num rows: 3 Data size: 888 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: NONE
File Output Operator
compressed: false
- Statistics: Num rows: 3 Data size: 888 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: NONE
table:
input format: org.apache.hadoop.mapred.SequenceFileInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
@@ -970,28 +970,28 @@ STAGE PLANS:
Map Operator Tree:
TableScan
alias: orc_pred
- Statistics: Num rows: 1049 Data size: 311170 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 208 Data size: 24150 Basic stats: COMPLETE Column stats: NONE
Filter Operator
predicate: ((t > 10) and (t <> 101) and (d >= 10.0) and (d < 12.0) and (s like '%son') and (not (s like '%car%')) and (t > 0) and si BETWEEN 300 AND 400) (type: boolean)
- Statistics: Num rows: 1 Data size: 296 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: NONE
Select Operator
expressions: t (type: tinyint), si (type: smallint), d (type: double), s (type: string)
outputColumnNames: _col0, _col1, _col2, _col3
- Statistics: Num rows: 1 Data size: 296 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
key expressions: _col3 (type: string)
sort order: -
- Statistics: Num rows: 1 Data size: 296 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: NONE
TopN Hash Memory Usage: 0.1
value expressions: _col0 (type: tinyint), _col1 (type: smallint), _col2 (type: double)
Reduce Operator Tree:
Select Operator
expressions: VALUE._col0 (type: tinyint), VALUE._col1 (type: smallint), VALUE._col2 (type: double), KEY.reducesinkkey0 (type: string)
outputColumnNames: _col0, _col1, _col2, _col3
- Statistics: Num rows: 1 Data size: 296 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: NONE
Limit
Number of rows: 3
- Statistics: Num rows: 1 Data size: 296 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: NONE
File Output Operator
compressed: false
table:
@@ -1006,20 +1006,20 @@ STAGE PLANS:
Reduce Output Operator
key expressions: _col3 (type: string)
sort order: -
- Statistics: Num rows: 1 Data size: 296 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: NONE
TopN Hash Memory Usage: 0.1
value expressions: _col0 (type: tinyint), _col1 (type: smallint), _col2 (type: double)
Reduce Operator Tree:
Select Operator
expressions: VALUE._col0 (type: tinyint), VALUE._col1 (type: smallint), VALUE._col2 (type: double), KEY.reducesinkkey0 (type: string)
outputColumnNames: _col0, _col1, _col2, _col3
- Statistics: Num rows: 1 Data size: 296 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: NONE
Limit
Number of rows: 3
- Statistics: Num rows: 1 Data size: 296 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: NONE
File Output Operator
compressed: false
- Statistics: Num rows: 1 Data size: 296 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: NONE
table:
input format: org.apache.hadoop.mapred.SequenceFileInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
@@ -1069,28 +1069,28 @@ STAGE PLANS:
TableScan
alias: orc_pred
filterExpr: ((t > 10) and (t <> 101) and (d >= 10.0) and (d < 12.0) and (s like '%son') and (not (s like '%car%')) and (t > 0) and si BETWEEN 300 AND 400) (type: boolean)
- Statistics: Num rows: 1049 Data size: 311170 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 208 Data size: 24150 Basic stats: COMPLETE Column stats: NONE
Filter Operator
predicate: ((t > 10) and (t <> 101) and (d >= 10.0) and (d < 12.0) and (s like '%son') and (not (s like '%car%')) and (t > 0) and si BETWEEN 300 AND 400) (type: boolean)
- Statistics: Num rows: 1 Data size: 296 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: NONE
Select Operator
expressions: t (type: tinyint), si (type: smallint), d (type: double), s (type: string)
outputColumnNames: _col0, _col1, _col2, _col3
- Statistics: Num rows: 1 Data size: 296 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: NONE
Reduce Output Operator
key expressions: _col3 (type: string)
sort order: -
- Statistics: Num rows: 1 Data size: 296 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: NONE
TopN Hash Memory Usage: 0.1
value expressions: _col0 (type: tinyint), _col1 (type: smallint), _col2 (type: double)
Reduce Operator Tree:
Select Operator
expressions: VALUE._col0 (type: tinyint), VALUE._col1 (type: smallint), VALUE._col2 (type: double), KEY.reducesinkkey0 (type: string)
outputColumnNames: _col0, _col1, _col2, _col3
- Statistics: Num rows: 1 Data size: 296 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: NONE
Limit
Number of rows: 3
- Statistics: Num rows: 1 Data size: 296 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: NONE
File Output Operator
compressed: false
table:
@@ -1105,20 +1105,20 @@ STAGE PLANS:
Reduce Output Operator
key expressions: _col3 (type: string)
sort order: -
- Statistics: Num rows: 1 Data size: 296 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: NONE
TopN Hash Memory Usage: 0.1
value expressions: _col0 (type: tinyint), _col1 (type: smallint), _col2 (type: double)
Reduce Operator Tree:
Select Operator
expressions: VALUE._col0 (type: tinyint), VALUE._col1 (type: smallint), VALUE._col2 (type: double), KEY.reducesinkkey0 (type: string)
outputColumnNames: _col0, _col1, _col2, _col3
- Statistics: Num rows: 1 Data size: 296 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: NONE
Limit
Number of rows: 3
- Statistics: Num rows: 1 Data size: 296 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: NONE
File Output Operator
compressed: false
- Statistics: Num rows: 1 Data size: 296 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: NONE
table:
input format: org.apache.hadoop.mapred.SequenceFileInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
http://git-wip-us.apache.org/repos/asf/hive/blob/244ce09c/ql/src/test/results/clientpositive/parquet_array_null_element.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/parquet_array_null_element.q.out b/ql/src/test/results/clientpositive/parquet_array_null_element.q.out
index 387f01e..75d2d27 100644
--- a/ql/src/test/results/clientpositive/parquet_array_null_element.q.out
+++ b/ql/src/test/results/clientpositive/parquet_array_null_element.q.out
@@ -70,6 +70,11 @@ Retention: 0
#### A masked pattern was here ####
Table Type: MANAGED_TABLE
Table Parameters:
+ COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\"}
+ numFiles 0
+ numRows 0
+ rawDataSize 0
+ totalSize 0
#### A masked pattern was here ####
# Storage Information
http://git-wip-us.apache.org/repos/asf/hive/blob/244ce09c/ql/src/test/results/clientpositive/parquet_create.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/parquet_create.q.out b/ql/src/test/results/clientpositive/parquet_create.q.out
index c6d33ff..cc4b735 100644
--- a/ql/src/test/results/clientpositive/parquet_create.q.out
+++ b/ql/src/test/results/clientpositive/parquet_create.q.out
@@ -73,6 +73,11 @@ Retention: 0
#### A masked pattern was here ####
Table Type: MANAGED_TABLE
Table Parameters:
+ COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\"}
+ numFiles 0
+ numRows 0
+ rawDataSize 0
+ totalSize 0
#### A masked pattern was here ####
# Storage Information
http://git-wip-us.apache.org/repos/asf/hive/blob/244ce09c/ql/src/test/results/clientpositive/parquet_mixed_partition_formats.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/parquet_mixed_partition_formats.q.out b/ql/src/test/results/clientpositive/parquet_mixed_partition_formats.q.out
index d6affd6..e96aa80 100644
--- a/ql/src/test/results/clientpositive/parquet_mixed_partition_formats.q.out
+++ b/ql/src/test/results/clientpositive/parquet_mixed_partition_formats.q.out
@@ -126,6 +126,8 @@ Table: parquet_mixed_partition_formats
#### A masked pattern was here ####
Partition Parameters:
numFiles 1
+ numRows 0
+ rawDataSize 0
totalSize 2521
#### A masked pattern was here ####
@@ -247,6 +249,8 @@ Table: parquet_mixed_partition_formats
#### A masked pattern was here ####
Partition Parameters:
numFiles 1
+ numRows 0
+ rawDataSize 0
totalSize 2521
#### A masked pattern was here ####
http://git-wip-us.apache.org/repos/asf/hive/blob/244ce09c/ql/src/test/results/clientpositive/parquet_serde.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/parquet_serde.q.out b/ql/src/test/results/clientpositive/parquet_serde.q.out
index c1e594a..6d5f0f8 100644
--- a/ql/src/test/results/clientpositive/parquet_serde.q.out
+++ b/ql/src/test/results/clientpositive/parquet_serde.q.out
@@ -72,6 +72,8 @@ Table: parquet_mixed_fileformat
#### A masked pattern was here ####
Partition Parameters:
numFiles 1
+ numRows 0
+ rawDataSize 0
totalSize 36
#### A masked pattern was here ####
@@ -173,6 +175,8 @@ Table: parquet_mixed_fileformat
#### A masked pattern was here ####
Partition Parameters:
numFiles 1
+ numRows 0
+ rawDataSize 0
totalSize 36
#### A masked pattern was here ####
http://git-wip-us.apache.org/repos/asf/hive/blob/244ce09c/ql/src/test/results/clientpositive/part_inherit_tbl_props.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/part_inherit_tbl_props.q.out b/ql/src/test/results/clientpositive/part_inherit_tbl_props.q.out
index 82f4750..876d0db 100644
--- a/ql/src/test/results/clientpositive/part_inherit_tbl_props.q.out
+++ b/ql/src/test/results/clientpositive/part_inherit_tbl_props.q.out
@@ -38,8 +38,13 @@ Database: default
Table: mytbl
#### A masked pattern was here ####
Partition Parameters:
+ COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\"}
a myval
b yourval
+ numFiles 0
+ numRows 0
+ rawDataSize 0
+ totalSize 0
#### A masked pattern was here ####
# Storage Information
http://git-wip-us.apache.org/repos/asf/hive/blob/244ce09c/ql/src/test/results/clientpositive/part_inherit_tbl_props_empty.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/part_inherit_tbl_props_empty.q.out b/ql/src/test/results/clientpositive/part_inherit_tbl_props_empty.q.out
index e4b8003..95c2d6e 100644
--- a/ql/src/test/results/clientpositive/part_inherit_tbl_props_empty.q.out
+++ b/ql/src/test/results/clientpositive/part_inherit_tbl_props_empty.q.out
@@ -34,6 +34,11 @@ Database: default
Table: mytbl
#### A masked pattern was here ####
Partition Parameters:
+ COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\"}
+ numFiles 0
+ numRows 0
+ rawDataSize 0
+ totalSize 0
#### A masked pattern was here ####
# Storage Information
http://git-wip-us.apache.org/repos/asf/hive/blob/244ce09c/ql/src/test/results/clientpositive/part_inherit_tbl_props_with_star.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/part_inherit_tbl_props_with_star.q.out b/ql/src/test/results/clientpositive/part_inherit_tbl_props_with_star.q.out
index 106448a..0be588a 100644
--- a/ql/src/test/results/clientpositive/part_inherit_tbl_props_with_star.q.out
+++ b/ql/src/test/results/clientpositive/part_inherit_tbl_props_with_star.q.out
@@ -38,9 +38,14 @@ Database: default
Table: mytbl
#### A masked pattern was here ####
Partition Parameters:
+ COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\"}
a myval
b yourval
c noval
+ numFiles 0
+ numRows 0
+ rawDataSize 0
+ totalSize 0
#### A masked pattern was here ####
# Storage Information
http://git-wip-us.apache.org/repos/asf/hive/blob/244ce09c/ql/src/test/results/clientpositive/partition_coltype_literals.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/partition_coltype_literals.q.out b/ql/src/test/results/clientpositive/partition_coltype_literals.q.out
index bc159eb..b7a09d2 100644
--- a/ql/src/test/results/clientpositive/partition_coltype_literals.q.out
+++ b/ql/src/test/results/clientpositive/partition_coltype_literals.q.out
@@ -45,6 +45,11 @@ Database: default
Table: partcoltypenum
#### A masked pattern was here ####
Partition Parameters:
+ COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\"}
+ numFiles 0
+ numRows 0
+ rawDataSize 0
+ totalSize 0
#### A masked pattern was here ####
# Storage Information
@@ -92,8 +97,11 @@ Database: default
Table: partcoltypenum
#### A masked pattern was here ####
Partition Parameters:
+ COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\"}
#### A masked pattern was here ####
numFiles 0
+ numRows 0
+ rawDataSize 0
totalSize 0
#### A masked pattern was here ####
@@ -142,8 +150,11 @@ Database: default
Table: partcoltypenum
#### A masked pattern was here ####
Partition Parameters:
+ COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\"}
#### A masked pattern was here ####
numFiles 0
+ numRows 0
+ rawDataSize 0
totalSize 0
#### A masked pattern was here ####
@@ -193,8 +204,11 @@ Database: default
Table: partcoltypenum
#### A masked pattern was here ####
Partition Parameters:
+ COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\"}
#### A masked pattern was here ####
numFiles 0
+ numRows 0
+ rawDataSize 0
totalSize 0
#### A masked pattern was here ####
http://git-wip-us.apache.org/repos/asf/hive/blob/244ce09c/ql/src/test/results/clientpositive/pcr.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/pcr.q.out b/ql/src/test/results/clientpositive/pcr.q.out
index b53226e..7222617 100644
--- a/ql/src/test/results/clientpositive/pcr.q.out
+++ b/ql/src/test/results/clientpositive/pcr.q.out
@@ -3419,15 +3419,20 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
+ COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
bucket_count -1
columns key,value
columns.comments
columns.types int:string
#### A masked pattern was here ####
name default.pcr_t2
+ numFiles 0
+ numRows 0
+ rawDataSize 0
serialization.ddl struct pcr_t2 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ totalSize 0
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.pcr_t2
@@ -3449,15 +3454,20 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
+ COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
bucket_count -1
columns key,value
columns.comments
columns.types int:string
#### A masked pattern was here ####
name default.pcr_t3
+ numFiles 0
+ numRows 0
+ rawDataSize 0
serialization.ddl struct pcr_t3 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ totalSize 0
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.pcr_t3
@@ -3533,15 +3543,20 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
+ COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
bucket_count -1
columns key,value
columns.comments
columns.types int:string
#### A masked pattern was here ####
name default.pcr_t2
+ numFiles 0
+ numRows 0
+ rawDataSize 0
serialization.ddl struct pcr_t2 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ totalSize 0
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.pcr_t2
@@ -3564,15 +3579,20 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
+ COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
bucket_count -1
columns key,value
columns.comments
columns.types int:string
#### A masked pattern was here ####
name default.pcr_t2
+ numFiles 0
+ numRows 0
+ rawDataSize 0
serialization.ddl struct pcr_t2 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ totalSize 0
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.pcr_t2
@@ -3588,30 +3608,40 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
+ COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
bucket_count -1
columns key,value
columns.comments
columns.types int:string
#### A masked pattern was here ####
name default.pcr_t2
+ numFiles 0
+ numRows 0
+ rawDataSize 0
serialization.ddl struct pcr_t2 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ totalSize 0
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
+ COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
bucket_count -1
columns key,value
columns.comments
columns.types int:string
#### A masked pattern was here ####
name default.pcr_t2
+ numFiles 0
+ numRows 0
+ rawDataSize 0
serialization.ddl struct pcr_t2 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ totalSize 0
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.pcr_t2
@@ -3633,15 +3663,20 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
+ COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
bucket_count -1
columns key,value
columns.comments
columns.types int:string
#### A masked pattern was here ####
name default.pcr_t2
+ numFiles 0
+ numRows 0
+ rawDataSize 0
serialization.ddl struct pcr_t2 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ totalSize 0
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.pcr_t2
@@ -3657,30 +3692,40 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
+ COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
bucket_count -1
columns key,value
columns.comments
columns.types int:string
#### A masked pattern was here ####
name default.pcr_t2
+ numFiles 0
+ numRows 0
+ rawDataSize 0
serialization.ddl struct pcr_t2 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ totalSize 0
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
+ COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
bucket_count -1
columns key,value
columns.comments
columns.types int:string
#### A masked pattern was here ####
name default.pcr_t2
+ numFiles 0
+ numRows 0
+ rawDataSize 0
serialization.ddl struct pcr_t2 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ totalSize 0
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.pcr_t2
@@ -3712,15 +3757,20 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
+ COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
bucket_count -1
columns key,value
columns.comments
columns.types int:string
#### A masked pattern was here ####
name default.pcr_t3
+ numFiles 0
+ numRows 0
+ rawDataSize 0
serialization.ddl struct pcr_t3 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ totalSize 0
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.pcr_t3
@@ -3743,15 +3793,20 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
+ COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
bucket_count -1
columns key,value
columns.comments
columns.types int:string
#### A masked pattern was here ####
name default.pcr_t3
+ numFiles 0
+ numRows 0
+ rawDataSize 0
serialization.ddl struct pcr_t3 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ totalSize 0
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.pcr_t3
@@ -3767,30 +3822,40 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
+ COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
bucket_count -1
columns key,value
columns.comments
columns.types int:string
#### A masked pattern was here ####
name default.pcr_t3
+ numFiles 0
+ numRows 0
+ rawDataSize 0
serialization.ddl struct pcr_t3 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ totalSize 0
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
+ COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
bucket_count -1
columns key,value
columns.comments
columns.types int:string
#### A masked pattern was here ####
name default.pcr_t3
+ numFiles 0
+ numRows 0
+ rawDataSize 0
serialization.ddl struct pcr_t3 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ totalSize 0
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.pcr_t3
@@ -3812,15 +3877,20 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
+ COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
bucket_count -1
columns key,value
columns.comments
columns.types int:string
#### A masked pattern was here ####
name default.pcr_t3
+ numFiles 0
+ numRows 0
+ rawDataSize 0
serialization.ddl struct pcr_t3 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ totalSize 0
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.pcr_t3
@@ -3836,30 +3906,40 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
+ COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
bucket_count -1
columns key,value
columns.comments
columns.types int:string
#### A masked pattern was here ####
name default.pcr_t3
+ numFiles 0
+ numRows 0
+ rawDataSize 0
serialization.ddl struct pcr_t3 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ totalSize 0
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
+ COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
bucket_count -1
columns key,value
columns.comments
columns.types int:string
#### A masked pattern was here ####
name default.pcr_t3
+ numFiles 0
+ numRows 0
+ rawDataSize 0
serialization.ddl struct pcr_t3 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ totalSize 0
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.pcr_t3