You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hive.apache.org by px...@apache.org on 2016/05/09 17:41:47 UTC
[05/21] hive git commit: HIVE-13341: Stats state is not captured
correctly: differentiate load table and create table (Pengcheng Xiong,
reviewed by Ashutosh Chauhan)
http://git-wip-us.apache.org/repos/asf/hive/blob/244ce09c/ql/src/test/results/clientpositive/spark/pcr.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/spark/pcr.q.out b/ql/src/test/results/clientpositive/spark/pcr.q.out
index 3dd2147..21b7519 100644
--- a/ql/src/test/results/clientpositive/spark/pcr.q.out
+++ b/ql/src/test/results/clientpositive/spark/pcr.q.out
@@ -3460,15 +3460,20 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
+ COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
bucket_count -1
columns key,value
columns.comments
columns.types int:string
#### A masked pattern was here ####
name default.pcr_t2
+ numFiles 0
+ numRows 0
+ rawDataSize 0
serialization.ddl struct pcr_t2 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ totalSize 0
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.pcr_t2
@@ -3490,15 +3495,20 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
+ COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
bucket_count -1
columns key,value
columns.comments
columns.types int:string
#### A masked pattern was here ####
name default.pcr_t3
+ numFiles 0
+ numRows 0
+ rawDataSize 0
serialization.ddl struct pcr_t3 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ totalSize 0
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.pcr_t3
@@ -3565,15 +3575,20 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
+ COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
bucket_count -1
columns key,value
columns.comments
columns.types int:string
#### A masked pattern was here ####
name default.pcr_t2
+ numFiles 0
+ numRows 0
+ rawDataSize 0
serialization.ddl struct pcr_t2 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ totalSize 0
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.pcr_t2
@@ -3591,15 +3606,20 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
+ COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
bucket_count -1
columns key,value
columns.comments
columns.types int:string
#### A masked pattern was here ####
name default.pcr_t3
+ numFiles 0
+ numRows 0
+ rawDataSize 0
serialization.ddl struct pcr_t3 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ totalSize 0
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.pcr_t3
http://git-wip-us.apache.org/repos/asf/hive/blob/244ce09c/ql/src/test/results/clientpositive/spark/reduce_deduplicate.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/spark/reduce_deduplicate.q.out b/ql/src/test/results/clientpositive/spark/reduce_deduplicate.q.out
index 3f5f649..b20e8fe 100644
--- a/ql/src/test/results/clientpositive/spark/reduce_deduplicate.q.out
+++ b/ql/src/test/results/clientpositive/spark/reduce_deduplicate.q.out
@@ -112,6 +112,7 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
+ COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
bucket_count 2
bucket_field_name key
columns key,value
@@ -119,9 +120,13 @@ STAGE PLANS:
columns.types string:string
#### A masked pattern was here ####
name default.bucket5_1
+ numFiles 0
+ numRows 0
+ rawDataSize 0
serialization.ddl struct bucket5_1 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ totalSize 0
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.bucket5_1
@@ -138,6 +143,7 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
+ COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
bucket_count 2
bucket_field_name key
columns key,value
@@ -145,9 +151,13 @@ STAGE PLANS:
columns.types string:string
#### A masked pattern was here ####
name default.bucket5_1
+ numFiles 0
+ numRows 0
+ rawDataSize 0
serialization.ddl struct bucket5_1 { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ totalSize 0
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.bucket5_1
http://git-wip-us.apache.org/repos/asf/hive/blob/244ce09c/ql/src/test/results/clientpositive/spark/sample1.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/spark/sample1.q.out b/ql/src/test/results/clientpositive/spark/sample1.q.out
index 15b1016..eb9d5f6 100644
--- a/ql/src/test/results/clientpositive/spark/sample1.q.out
+++ b/ql/src/test/results/clientpositive/spark/sample1.q.out
@@ -53,15 +53,20 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
+ COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
bucket_count -1
columns key,value,dt,hr
columns.comments
columns.types int:string:string:string
#### A masked pattern was here ####
name default.dest1
+ numFiles 0
+ numRows 0
+ rawDataSize 0
serialization.ddl struct dest1 { i32 key, string value, string dt, string hr}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ totalSize 0
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.dest1
@@ -129,15 +134,20 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
+ COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
bucket_count -1
columns key,value,dt,hr
columns.comments
columns.types int:string:string:string
#### A masked pattern was here ####
name default.dest1
+ numFiles 0
+ numRows 0
+ rawDataSize 0
serialization.ddl struct dest1 { i32 key, string value, string dt, string hr}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ totalSize 0
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.dest1
http://git-wip-us.apache.org/repos/asf/hive/blob/244ce09c/ql/src/test/results/clientpositive/spark/sample2.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/spark/sample2.q.out b/ql/src/test/results/clientpositive/spark/sample2.q.out
index 28108bb..b13f818 100644
--- a/ql/src/test/results/clientpositive/spark/sample2.q.out
+++ b/ql/src/test/results/clientpositive/spark/sample2.q.out
@@ -54,15 +54,20 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
+ COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
bucket_count -1
columns key,value
columns.comments
columns.types int:string
#### A masked pattern was here ####
name default.dest1
+ numFiles 0
+ numRows 0
+ rawDataSize 0
serialization.ddl struct dest1 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ totalSize 0
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.dest1
@@ -130,15 +135,20 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
+ COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
bucket_count -1
columns key,value
columns.comments
columns.types int:string
#### A masked pattern was here ####
name default.dest1
+ numFiles 0
+ numRows 0
+ rawDataSize 0
serialization.ddl struct dest1 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ totalSize 0
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.dest1
http://git-wip-us.apache.org/repos/asf/hive/blob/244ce09c/ql/src/test/results/clientpositive/spark/sample4.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/spark/sample4.q.out b/ql/src/test/results/clientpositive/spark/sample4.q.out
index 25f300d..69066c1 100644
--- a/ql/src/test/results/clientpositive/spark/sample4.q.out
+++ b/ql/src/test/results/clientpositive/spark/sample4.q.out
@@ -54,15 +54,20 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
+ COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
bucket_count -1
columns key,value
columns.comments
columns.types int:string
#### A masked pattern was here ####
name default.dest1
+ numFiles 0
+ numRows 0
+ rawDataSize 0
serialization.ddl struct dest1 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ totalSize 0
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.dest1
@@ -130,15 +135,20 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
+ COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
bucket_count -1
columns key,value
columns.comments
columns.types int:string
#### A masked pattern was here ####
name default.dest1
+ numFiles 0
+ numRows 0
+ rawDataSize 0
serialization.ddl struct dest1 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ totalSize 0
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.dest1
http://git-wip-us.apache.org/repos/asf/hive/blob/244ce09c/ql/src/test/results/clientpositive/spark/sample5.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/spark/sample5.q.out b/ql/src/test/results/clientpositive/spark/sample5.q.out
index 8aebfe5..819939c 100644
--- a/ql/src/test/results/clientpositive/spark/sample5.q.out
+++ b/ql/src/test/results/clientpositive/spark/sample5.q.out
@@ -55,15 +55,20 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
+ COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
bucket_count -1
columns key,value
columns.comments
columns.types int:string
#### A masked pattern was here ####
name default.dest1
+ numFiles 0
+ numRows 0
+ rawDataSize 0
serialization.ddl struct dest1 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ totalSize 0
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.dest1
@@ -131,15 +136,20 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
+ COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
bucket_count -1
columns key,value
columns.comments
columns.types int:string
#### A masked pattern was here ####
name default.dest1
+ numFiles 0
+ numRows 0
+ rawDataSize 0
serialization.ddl struct dest1 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ totalSize 0
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.dest1
http://git-wip-us.apache.org/repos/asf/hive/blob/244ce09c/ql/src/test/results/clientpositive/spark/sample6.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/spark/sample6.q.out b/ql/src/test/results/clientpositive/spark/sample6.q.out
index 2b49712..bf06004 100644
--- a/ql/src/test/results/clientpositive/spark/sample6.q.out
+++ b/ql/src/test/results/clientpositive/spark/sample6.q.out
@@ -52,15 +52,20 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
+ COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
bucket_count -1
columns key,value
columns.comments
columns.types int:string
#### A masked pattern was here ####
name default.dest1
+ numFiles 0
+ numRows 0
+ rawDataSize 0
serialization.ddl struct dest1 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ totalSize 0
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.dest1
@@ -128,15 +133,20 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
+ COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
bucket_count -1
columns key,value
columns.comments
columns.types int:string
#### A masked pattern was here ####
name default.dest1
+ numFiles 0
+ numRows 0
+ rawDataSize 0
serialization.ddl struct dest1 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ totalSize 0
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.dest1
http://git-wip-us.apache.org/repos/asf/hive/blob/244ce09c/ql/src/test/results/clientpositive/spark/sample7.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/spark/sample7.q.out b/ql/src/test/results/clientpositive/spark/sample7.q.out
index 6c17ab1..a821c76 100644
--- a/ql/src/test/results/clientpositive/spark/sample7.q.out
+++ b/ql/src/test/results/clientpositive/spark/sample7.q.out
@@ -53,15 +53,20 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
+ COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
bucket_count -1
columns key,value
columns.comments
columns.types int:string
#### A masked pattern was here ####
name default.dest1
+ numFiles 0
+ numRows 0
+ rawDataSize 0
serialization.ddl struct dest1 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ totalSize 0
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.dest1
@@ -129,15 +134,20 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
+ COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
bucket_count -1
columns key,value
columns.comments
columns.types int:string
#### A masked pattern was here ####
name default.dest1
+ numFiles 0
+ numRows 0
+ rawDataSize 0
serialization.ddl struct dest1 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ totalSize 0
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.dest1
http://git-wip-us.apache.org/repos/asf/hive/blob/244ce09c/ql/src/test/results/clientpositive/spark/stats0.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/spark/stats0.q.out b/ql/src/test/results/clientpositive/spark/stats0.q.out
index e741320..0b14e21 100644
--- a/ql/src/test/results/clientpositive/spark/stats0.q.out
+++ b/ql/src/test/results/clientpositive/spark/stats0.q.out
@@ -45,15 +45,20 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
+ COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
bucket_count -1
columns key,value
columns.comments
columns.types string:string
#### A masked pattern was here ####
name default.stats_non_partitioned
+ numFiles 0
+ numRows 0
+ rawDataSize 0
serialization.ddl struct stats_non_partitioned { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ totalSize 0
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.stats_non_partitioned
@@ -119,15 +124,20 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
+ COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
bucket_count -1
columns key,value
columns.comments
columns.types string:string
#### A masked pattern was here ####
name default.stats_non_partitioned
+ numFiles 0
+ numRows 0
+ rawDataSize 0
serialization.ddl struct stats_non_partitioned { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ totalSize 0
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.stats_non_partitioned
@@ -1349,15 +1359,20 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
+ COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
bucket_count -1
columns key,value
columns.comments
columns.types string:string
#### A masked pattern was here ####
name default.stats_non_partitioned
+ numFiles 0
+ numRows 0
+ rawDataSize 0
serialization.ddl struct stats_non_partitioned { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ totalSize 0
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.stats_non_partitioned
@@ -1423,15 +1438,20 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
+ COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
bucket_count -1
columns key,value
columns.comments
columns.types string:string
#### A masked pattern was here ####
name default.stats_non_partitioned
+ numFiles 0
+ numRows 0
+ rawDataSize 0
serialization.ddl struct stats_non_partitioned { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ totalSize 0
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.stats_non_partitioned
http://git-wip-us.apache.org/repos/asf/hive/blob/244ce09c/ql/src/test/results/clientpositive/spark/stats1.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/spark/stats1.q.out b/ql/src/test/results/clientpositive/spark/stats1.q.out
index 588a33d..2f29d27 100644
--- a/ql/src/test/results/clientpositive/spark/stats1.q.out
+++ b/ql/src/test/results/clientpositive/spark/stats1.q.out
@@ -224,6 +224,8 @@ Retention: 0
Table Type: MANAGED_TABLE
Table Parameters:
numFiles 3
+ numRows 26
+ rawDataSize 199
totalSize 1583
#### A masked pattern was here ####
http://git-wip-us.apache.org/repos/asf/hive/blob/244ce09c/ql/src/test/results/clientpositive/spark/stats18.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/spark/stats18.q.out b/ql/src/test/results/clientpositive/spark/stats18.q.out
index 6971e44..3ad9679 100644
--- a/ql/src/test/results/clientpositive/spark/stats18.q.out
+++ b/ql/src/test/results/clientpositive/spark/stats18.q.out
@@ -94,6 +94,8 @@ Table: stats_part
#### A masked pattern was here ####
Partition Parameters:
numFiles 2
+ numRows 500
+ rawDataSize 5312
totalSize 7170
#### A masked pattern was here ####
http://git-wip-us.apache.org/repos/asf/hive/blob/244ce09c/ql/src/test/results/clientpositive/spark/stats20.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/spark/stats20.q.out b/ql/src/test/results/clientpositive/spark/stats20.q.out
index d7e52b4..a824bc9 100644
--- a/ql/src/test/results/clientpositive/spark/stats20.q.out
+++ b/ql/src/test/results/clientpositive/spark/stats20.q.out
@@ -55,6 +55,45 @@ Bucket Columns: []
Sort Columns: []
Storage Desc Params:
serialization.format 1
+PREHOOK: query: describe formatted stats_partitioned partition (ds='1')
+PREHOOK: type: DESCTABLE
+PREHOOK: Input: default@stats_partitioned
+POSTHOOK: query: describe formatted stats_partitioned partition (ds='1')
+POSTHOOK: type: DESCTABLE
+POSTHOOK: Input: default@stats_partitioned
+# col_name data_type comment
+
+key string
+value string
+
+# Partition Information
+# col_name data_type comment
+
+ds string
+
+# Detailed Partition Information
+Partition Value: [1]
+Database: default
+Table: stats_partitioned
+#### A masked pattern was here ####
+Partition Parameters:
+ COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\"}
+ numFiles 1
+ numRows 500
+ rawDataSize 5312
+ totalSize 5812
+#### A masked pattern was here ####
+
+# Storage Information
+SerDe Library: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+InputFormat: org.apache.hadoop.mapred.TextInputFormat
+OutputFormat: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+Compressed: No
+Num Buckets: -1
+Bucket Columns: []
+Sort Columns: []
+Storage Desc Params:
+ serialization.format 1
PREHOOK: query: insert overwrite table stats_partitioned partition (ds='1')
select * from src
PREHOOK: type: QUERY
@@ -104,3 +143,42 @@ Bucket Columns: []
Sort Columns: []
Storage Desc Params:
serialization.format 1
+PREHOOK: query: describe formatted stats_partitioned partition (ds='1')
+PREHOOK: type: DESCTABLE
+PREHOOK: Input: default@stats_partitioned
+POSTHOOK: query: describe formatted stats_partitioned partition (ds='1')
+POSTHOOK: type: DESCTABLE
+POSTHOOK: Input: default@stats_partitioned
+# col_name data_type comment
+
+key string
+value string
+
+# Partition Information
+# col_name data_type comment
+
+ds string
+
+# Detailed Partition Information
+Partition Value: [1]
+Database: default
+Table: stats_partitioned
+#### A masked pattern was here ####
+Partition Parameters:
+ COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\"}
+ numFiles 1
+ numRows 500
+ rawDataSize 0
+ totalSize 5812
+#### A masked pattern was here ####
+
+# Storage Information
+SerDe Library: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+InputFormat: org.apache.hadoop.mapred.TextInputFormat
+OutputFormat: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+Compressed: No
+Num Buckets: -1
+Bucket Columns: []
+Sort Columns: []
+Storage Desc Params:
+ serialization.format 1
http://git-wip-us.apache.org/repos/asf/hive/blob/244ce09c/ql/src/test/results/clientpositive/spark/stats3.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/spark/stats3.q.out b/ql/src/test/results/clientpositive/spark/stats3.q.out
index 4dbda94..7db4fa0 100644
--- a/ql/src/test/results/clientpositive/spark/stats3.q.out
+++ b/ql/src/test/results/clientpositive/spark/stats3.q.out
@@ -34,15 +34,20 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
+ COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
bucket_count -1
columns col1
columns.comments
columns.types string
#### A masked pattern was here ####
name default.hive_test_src
+ numFiles 0
+ numRows 0
+ rawDataSize 0
serialization.ddl struct hive_test_src { string col1}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ totalSize 0
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.hive_test_src
@@ -76,6 +81,8 @@ Retention: 0
Table Type: MANAGED_TABLE
Table Parameters:
numFiles 1
+ numRows 0
+ rawDataSize 0
totalSize 11
#### A masked pattern was here ####
http://git-wip-us.apache.org/repos/asf/hive/blob/244ce09c/ql/src/test/results/clientpositive/spark/vectorized_ptf.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/spark/vectorized_ptf.q.out b/ql/src/test/results/clientpositive/spark/vectorized_ptf.q.out
index 7b56585..f4b9584 100644
--- a/ql/src/test/results/clientpositive/spark/vectorized_ptf.q.out
+++ b/ql/src/test/results/clientpositive/spark/vectorized_ptf.q.out
@@ -4592,15 +4592,20 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
+ COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
bucket_count -1
columns p_mfgr,p_name,p_size,r,dr,s
columns.comments
columns.types string:string:int:int:int:double
#### A masked pattern was here ####
name default.part_4
+ numFiles 0
+ numRows 0
+ rawDataSize 0
serialization.ddl struct part_4 { string p_mfgr, string p_name, i32 p_size, i32 r, i32 dr, double s}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ totalSize 0
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.part_4
@@ -4710,15 +4715,20 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
+ COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
bucket_count -1
columns p_mfgr,p_name,p_size,s2,r,dr,cud,fv1
columns.comments
columns.types string:string:int:int:int:int:double:int
#### A masked pattern was here ####
name default.part_5
+ numFiles 0
+ numRows 0
+ rawDataSize 0
serialization.ddl struct part_5 { string p_mfgr, string p_name, i32 p_size, i32 s2, i32 r, i32 dr, double cud, i32 fv1}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ totalSize 0
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.part_5
@@ -4791,15 +4801,20 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
+ COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
bucket_count -1
columns p_mfgr,p_name,p_size,r,dr,s
columns.comments
columns.types string:string:int:int:int:double
#### A masked pattern was here ####
name default.part_4
+ numFiles 0
+ numRows 0
+ rawDataSize 0
serialization.ddl struct part_4 { string p_mfgr, string p_name, i32 p_size, i32 r, i32 dr, double s}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ totalSize 0
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.part_4
@@ -4817,15 +4832,20 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
+ COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
bucket_count -1
columns p_mfgr,p_name,p_size,s2,r,dr,cud,fv1
columns.comments
columns.types string:string:int:int:int:int:double:int
#### A masked pattern was here ####
name default.part_5
+ numFiles 0
+ numRows 0
+ rawDataSize 0
serialization.ddl struct part_5 { string p_mfgr, string p_name, i32 p_size, i32 s2, i32 r, i32 dr, double cud, i32 fv1}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ totalSize 0
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.part_5
http://git-wip-us.apache.org/repos/asf/hive/blob/244ce09c/ql/src/test/results/clientpositive/stats0.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/stats0.q.out b/ql/src/test/results/clientpositive/stats0.q.out
index 90a9a70..bbe38c1 100644
--- a/ql/src/test/results/clientpositive/stats0.q.out
+++ b/ql/src/test/results/clientpositive/stats0.q.out
@@ -42,15 +42,20 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
+ COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
bucket_count -1
columns key,value
columns.comments
columns.types string:string
#### A masked pattern was here ####
name default.stats_non_partitioned
+ numFiles 0
+ numRows 0
+ rawDataSize 0
serialization.ddl struct stats_non_partitioned { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ totalSize 0
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.stats_non_partitioned
@@ -116,15 +121,20 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
+ COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
bucket_count -1
columns key,value
columns.comments
columns.types string:string
#### A masked pattern was here ####
name default.stats_non_partitioned
+ numFiles 0
+ numRows 0
+ rawDataSize 0
serialization.ddl struct stats_non_partitioned { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ totalSize 0
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.stats_non_partitioned
@@ -1345,15 +1355,20 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
+ COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
bucket_count -1
columns key,value
columns.comments
columns.types string:string
#### A masked pattern was here ####
name default.stats_non_partitioned
+ numFiles 0
+ numRows 0
+ rawDataSize 0
serialization.ddl struct stats_non_partitioned { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ totalSize 0
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.stats_non_partitioned
@@ -1428,15 +1443,20 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
+ COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
bucket_count -1
columns key,value
columns.comments
columns.types string:string
#### A masked pattern was here ####
name default.stats_non_partitioned
+ numFiles 0
+ numRows 0
+ rawDataSize 0
serialization.ddl struct stats_non_partitioned { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ totalSize 0
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.stats_non_partitioned
@@ -1459,15 +1479,20 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
+ COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
bucket_count -1
columns key,value
columns.comments
columns.types string:string
#### A masked pattern was here ####
name default.stats_non_partitioned
+ numFiles 0
+ numRows 0
+ rawDataSize 0
serialization.ddl struct stats_non_partitioned { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ totalSize 0
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.stats_non_partitioned
@@ -1483,30 +1508,40 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
+ COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
bucket_count -1
columns key,value
columns.comments
columns.types string:string
#### A masked pattern was here ####
name default.stats_non_partitioned
+ numFiles 0
+ numRows 0
+ rawDataSize 0
serialization.ddl struct stats_non_partitioned { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ totalSize 0
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
+ COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
bucket_count -1
columns key,value
columns.comments
columns.types string:string
#### A masked pattern was here ####
name default.stats_non_partitioned
+ numFiles 0
+ numRows 0
+ rawDataSize 0
serialization.ddl struct stats_non_partitioned { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ totalSize 0
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.stats_non_partitioned
@@ -1528,15 +1563,20 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
+ COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
bucket_count -1
columns key,value
columns.comments
columns.types string:string
#### A masked pattern was here ####
name default.stats_non_partitioned
+ numFiles 0
+ numRows 0
+ rawDataSize 0
serialization.ddl struct stats_non_partitioned { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ totalSize 0
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.stats_non_partitioned
@@ -1552,30 +1592,40 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
+ COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
bucket_count -1
columns key,value
columns.comments
columns.types string:string
#### A masked pattern was here ####
name default.stats_non_partitioned
+ numFiles 0
+ numRows 0
+ rawDataSize 0
serialization.ddl struct stats_non_partitioned { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ totalSize 0
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
+ COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
bucket_count -1
columns key,value
columns.comments
columns.types string:string
#### A masked pattern was here ####
name default.stats_non_partitioned
+ numFiles 0
+ numRows 0
+ rawDataSize 0
serialization.ddl struct stats_non_partitioned { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ totalSize 0
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.stats_non_partitioned
http://git-wip-us.apache.org/repos/asf/hive/blob/244ce09c/ql/src/test/results/clientpositive/stats1.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/stats1.q.out b/ql/src/test/results/clientpositive/stats1.q.out
index 72c53e3..ac076ec 100644
--- a/ql/src/test/results/clientpositive/stats1.q.out
+++ b/ql/src/test/results/clientpositive/stats1.q.out
@@ -232,6 +232,8 @@ Retention: 0
Table Type: MANAGED_TABLE
Table Parameters:
numFiles 3
+ numRows 26
+ rawDataSize 199
totalSize 1583
#### A masked pattern was here ####
http://git-wip-us.apache.org/repos/asf/hive/blob/244ce09c/ql/src/test/results/clientpositive/stats11.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/stats11.q.out b/ql/src/test/results/clientpositive/stats11.q.out
index 8ca5b27..9395773 100644
--- a/ql/src/test/results/clientpositive/stats11.q.out
+++ b/ql/src/test/results/clientpositive/stats11.q.out
@@ -88,6 +88,8 @@ Table: srcbucket_mapjoin_part
#### A masked pattern was here ####
Partition Parameters:
numFiles 1
+ numRows 0
+ rawDataSize 0
totalSize 1358
#### A masked pattern was here ####
@@ -132,6 +134,8 @@ Table: srcbucket_mapjoin_part
#### A masked pattern was here ####
Partition Parameters:
numFiles 2
+ numRows 0
+ rawDataSize 0
totalSize 2750
#### A masked pattern was here ####
@@ -176,6 +180,8 @@ Table: srcbucket_mapjoin_part
#### A masked pattern was here ####
Partition Parameters:
numFiles 3
+ numRows 0
+ rawDataSize 0
totalSize 4200
#### A masked pattern was here ####
@@ -220,6 +226,8 @@ Table: srcbucket_mapjoin_part
#### A masked pattern was here ####
Partition Parameters:
numFiles 4
+ numRows 0
+ rawDataSize 0
totalSize 5812
#### A masked pattern was here ####
@@ -327,8 +335,10 @@ STAGE PLANS:
#### A masked pattern was here ####
name default.srcbucket_mapjoin_part
numFiles 4
+ numRows 0
partition_columns ds
partition_columns.types string
+ rawDataSize 0
serialization.ddl struct srcbucket_mapjoin_part { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -414,15 +424,20 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
+ COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
bucket_count -1
columns key,value1,value2
columns.comments
columns.types string:string:string
#### A masked pattern was here ####
name default.bucketmapjoin_tmp_result
+ numFiles 0
+ numRows 0
+ rawDataSize 0
serialization.ddl struct bucketmapjoin_tmp_result { string key, string value1, string value2}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ totalSize 0
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.bucketmapjoin_tmp_result
@@ -448,6 +463,8 @@ STAGE PLANS:
#### A masked pattern was here ####
name default.srcbucket_mapjoin
numFiles 2
+ numRows 0
+ rawDataSize 0
serialization.ddl struct srcbucket_mapjoin { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -466,6 +483,8 @@ STAGE PLANS:
#### A masked pattern was here ####
name default.srcbucket_mapjoin
numFiles 2
+ numRows 0
+ rawDataSize 0
serialization.ddl struct srcbucket_mapjoin { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -495,15 +514,20 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
+ COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
bucket_count -1
columns key,value1,value2
columns.comments
columns.types string:string:string
#### A masked pattern was here ####
name default.bucketmapjoin_tmp_result
+ numFiles 0
+ numRows 0
+ rawDataSize 0
serialization.ddl struct bucketmapjoin_tmp_result { string key, string value1, string value2}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ totalSize 0
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.bucketmapjoin_tmp_result
@@ -526,15 +550,20 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
+ COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
bucket_count -1
columns key,value1,value2
columns.comments
columns.types string:string:string
#### A masked pattern was here ####
name default.bucketmapjoin_tmp_result
+ numFiles 0
+ numRows 0
+ rawDataSize 0
serialization.ddl struct bucketmapjoin_tmp_result { string key, string value1, string value2}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ totalSize 0
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.bucketmapjoin_tmp_result
@@ -550,30 +579,40 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
+ COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
bucket_count -1
columns key,value1,value2
columns.comments
columns.types string:string:string
#### A masked pattern was here ####
name default.bucketmapjoin_tmp_result
+ numFiles 0
+ numRows 0
+ rawDataSize 0
serialization.ddl struct bucketmapjoin_tmp_result { string key, string value1, string value2}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ totalSize 0
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
+ COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
bucket_count -1
columns key,value1,value2
columns.comments
columns.types string:string:string
#### A masked pattern was here ####
name default.bucketmapjoin_tmp_result
+ numFiles 0
+ numRows 0
+ rawDataSize 0
serialization.ddl struct bucketmapjoin_tmp_result { string key, string value1, string value2}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ totalSize 0
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.bucketmapjoin_tmp_result
@@ -595,15 +634,20 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
+ COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
bucket_count -1
columns key,value1,value2
columns.comments
columns.types string:string:string
#### A masked pattern was here ####
name default.bucketmapjoin_tmp_result
+ numFiles 0
+ numRows 0
+ rawDataSize 0
serialization.ddl struct bucketmapjoin_tmp_result { string key, string value1, string value2}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ totalSize 0
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.bucketmapjoin_tmp_result
@@ -619,30 +663,40 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
+ COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
bucket_count -1
columns key,value1,value2
columns.comments
columns.types string:string:string
#### A masked pattern was here ####
name default.bucketmapjoin_tmp_result
+ numFiles 0
+ numRows 0
+ rawDataSize 0
serialization.ddl struct bucketmapjoin_tmp_result { string key, string value1, string value2}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ totalSize 0
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
+ COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
bucket_count -1
columns key,value1,value2
columns.comments
columns.types string:string:string
#### A masked pattern was here ####
name default.bucketmapjoin_tmp_result
+ numFiles 0
+ numRows 0
+ rawDataSize 0
serialization.ddl struct bucketmapjoin_tmp_result { string key, string value1, string value2}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ totalSize 0
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.bucketmapjoin_tmp_result
@@ -887,8 +941,10 @@ STAGE PLANS:
#### A masked pattern was here ####
name default.srcbucket_mapjoin_part
numFiles 4
+ numRows 0
partition_columns ds
partition_columns.types string
+ rawDataSize 0
serialization.ddl struct srcbucket_mapjoin_part { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
http://git-wip-us.apache.org/repos/asf/hive/blob/244ce09c/ql/src/test/results/clientpositive/stats18.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/stats18.q.out b/ql/src/test/results/clientpositive/stats18.q.out
index 6971e44..3ad9679 100644
--- a/ql/src/test/results/clientpositive/stats18.q.out
+++ b/ql/src/test/results/clientpositive/stats18.q.out
@@ -94,6 +94,8 @@ Table: stats_part
#### A masked pattern was here ####
Partition Parameters:
numFiles 2
+ numRows 500
+ rawDataSize 5312
totalSize 7170
#### A masked pattern was here ####
http://git-wip-us.apache.org/repos/asf/hive/blob/244ce09c/ql/src/test/results/clientpositive/stats20.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/stats20.q.out b/ql/src/test/results/clientpositive/stats20.q.out
index d7e52b4..a824bc9 100644
--- a/ql/src/test/results/clientpositive/stats20.q.out
+++ b/ql/src/test/results/clientpositive/stats20.q.out
@@ -55,6 +55,45 @@ Bucket Columns: []
Sort Columns: []
Storage Desc Params:
serialization.format 1
+PREHOOK: query: describe formatted stats_partitioned partition (ds='1')
+PREHOOK: type: DESCTABLE
+PREHOOK: Input: default@stats_partitioned
+POSTHOOK: query: describe formatted stats_partitioned partition (ds='1')
+POSTHOOK: type: DESCTABLE
+POSTHOOK: Input: default@stats_partitioned
+# col_name data_type comment
+
+key string
+value string
+
+# Partition Information
+# col_name data_type comment
+
+ds string
+
+# Detailed Partition Information
+Partition Value: [1]
+Database: default
+Table: stats_partitioned
+#### A masked pattern was here ####
+Partition Parameters:
+ COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\"}
+ numFiles 1
+ numRows 500
+ rawDataSize 5312
+ totalSize 5812
+#### A masked pattern was here ####
+
+# Storage Information
+SerDe Library: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+InputFormat: org.apache.hadoop.mapred.TextInputFormat
+OutputFormat: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+Compressed: No
+Num Buckets: -1
+Bucket Columns: []
+Sort Columns: []
+Storage Desc Params:
+ serialization.format 1
PREHOOK: query: insert overwrite table stats_partitioned partition (ds='1')
select * from src
PREHOOK: type: QUERY
@@ -104,3 +143,42 @@ Bucket Columns: []
Sort Columns: []
Storage Desc Params:
serialization.format 1
+PREHOOK: query: describe formatted stats_partitioned partition (ds='1')
+PREHOOK: type: DESCTABLE
+PREHOOK: Input: default@stats_partitioned
+POSTHOOK: query: describe formatted stats_partitioned partition (ds='1')
+POSTHOOK: type: DESCTABLE
+POSTHOOK: Input: default@stats_partitioned
+# col_name data_type comment
+
+key string
+value string
+
+# Partition Information
+# col_name data_type comment
+
+ds string
+
+# Detailed Partition Information
+Partition Value: [1]
+Database: default
+Table: stats_partitioned
+#### A masked pattern was here ####
+Partition Parameters:
+ COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\"}
+ numFiles 1
+ numRows 500
+ rawDataSize 0
+ totalSize 5812
+#### A masked pattern was here ####
+
+# Storage Information
+SerDe Library: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+InputFormat: org.apache.hadoop.mapred.TextInputFormat
+OutputFormat: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+Compressed: No
+Num Buckets: -1
+Bucket Columns: []
+Sort Columns: []
+Storage Desc Params:
+ serialization.format 1
http://git-wip-us.apache.org/repos/asf/hive/blob/244ce09c/ql/src/test/results/clientpositive/stats3.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/stats3.q.out b/ql/src/test/results/clientpositive/stats3.q.out
index 4dbda94..7db4fa0 100644
--- a/ql/src/test/results/clientpositive/stats3.q.out
+++ b/ql/src/test/results/clientpositive/stats3.q.out
@@ -34,15 +34,20 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
+ COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
bucket_count -1
columns col1
columns.comments
columns.types string
#### A masked pattern was here ####
name default.hive_test_src
+ numFiles 0
+ numRows 0
+ rawDataSize 0
serialization.ddl struct hive_test_src { string col1}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ totalSize 0
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.hive_test_src
@@ -76,6 +81,8 @@ Retention: 0
Table Type: MANAGED_TABLE
Table Parameters:
numFiles 1
+ numRows 0
+ rawDataSize 0
totalSize 11
#### A masked pattern was here ####
http://git-wip-us.apache.org/repos/asf/hive/blob/244ce09c/ql/src/test/results/clientpositive/tez/auto_sortmerge_join_1.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/tez/auto_sortmerge_join_1.q.out b/ql/src/test/results/clientpositive/tez/auto_sortmerge_join_1.q.out
index 892539d..dffe096 100644
--- a/ql/src/test/results/clientpositive/tez/auto_sortmerge_join_1.q.out
+++ b/ql/src/test/results/clientpositive/tez/auto_sortmerge_join_1.q.out
@@ -163,8 +163,10 @@ STAGE PLANS:
#### A masked pattern was here ####
name default.bucket_small
numFiles 2
+ numRows 0
partition_columns ds
partition_columns.types string
+ rawDataSize 0
serialization.ddl struct bucket_small { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -251,8 +253,10 @@ STAGE PLANS:
#### A masked pattern was here ####
name default.bucket_big
numFiles 4
+ numRows 0
partition_columns ds
partition_columns.types string
+ rawDataSize 0
serialization.ddl struct bucket_big { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -296,8 +300,10 @@ STAGE PLANS:
#### A masked pattern was here ####
name default.bucket_big
numFiles 4
+ numRows 0
partition_columns ds
partition_columns.types string
+ rawDataSize 0
serialization.ddl struct bucket_big { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -456,8 +462,10 @@ STAGE PLANS:
#### A masked pattern was here ####
name default.bucket_big
numFiles 4
+ numRows 0
partition_columns ds
partition_columns.types string
+ rawDataSize 0
serialization.ddl struct bucket_big { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -501,8 +509,10 @@ STAGE PLANS:
#### A masked pattern was here ####
name default.bucket_big
numFiles 4
+ numRows 0
partition_columns ds
partition_columns.types string
+ rawDataSize 0
serialization.ddl struct bucket_big { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -574,8 +584,10 @@ STAGE PLANS:
#### A masked pattern was here ####
name default.bucket_small
numFiles 2
+ numRows 0
partition_columns ds
partition_columns.types string
+ rawDataSize 0
serialization.ddl struct bucket_small { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -733,8 +745,10 @@ STAGE PLANS:
#### A masked pattern was here ####
name default.bucket_big
numFiles 4
+ numRows 0
partition_columns ds
partition_columns.types string
+ rawDataSize 0
serialization.ddl struct bucket_big { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -778,8 +792,10 @@ STAGE PLANS:
#### A masked pattern was here ####
name default.bucket_big
numFiles 4
+ numRows 0
partition_columns ds
partition_columns.types string
+ rawDataSize 0
serialization.ddl struct bucket_big { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -851,8 +867,10 @@ STAGE PLANS:
#### A masked pattern was here ####
name default.bucket_small
numFiles 2
+ numRows 0
partition_columns ds
partition_columns.types string
+ rawDataSize 0
serialization.ddl struct bucket_small { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
http://git-wip-us.apache.org/repos/asf/hive/blob/244ce09c/ql/src/test/results/clientpositive/tez/auto_sortmerge_join_11.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/tez/auto_sortmerge_join_11.q.out b/ql/src/test/results/clientpositive/tez/auto_sortmerge_join_11.q.out
index 954fd96..8a553d7 100644
--- a/ql/src/test/results/clientpositive/tez/auto_sortmerge_join_11.q.out
+++ b/ql/src/test/results/clientpositive/tez/auto_sortmerge_join_11.q.out
@@ -159,8 +159,10 @@ STAGE PLANS:
#### A masked pattern was here ####
name default.bucket_small
numFiles 2
+ numRows 0
partition_columns ds
partition_columns.types string
+ rawDataSize 0
serialization.ddl struct bucket_small { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -246,8 +248,10 @@ STAGE PLANS:
#### A masked pattern was here ####
name default.bucket_big
numFiles 4
+ numRows 0
partition_columns ds
partition_columns.types string
+ rawDataSize 0
serialization.ddl struct bucket_big { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -290,8 +294,10 @@ STAGE PLANS:
#### A masked pattern was here ####
name default.bucket_big
numFiles 4
+ numRows 0
partition_columns ds
partition_columns.types string
+ rawDataSize 0
serialization.ddl struct bucket_big { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -441,8 +447,10 @@ STAGE PLANS:
#### A masked pattern was here ####
name default.bucket_small
numFiles 2
+ numRows 0
partition_columns ds
partition_columns.types string
+ rawDataSize 0
serialization.ddl struct bucket_small { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -528,8 +536,10 @@ STAGE PLANS:
#### A masked pattern was here ####
name default.bucket_big
numFiles 4
+ numRows 0
partition_columns ds
partition_columns.types string
+ rawDataSize 0
serialization.ddl struct bucket_big { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -572,8 +582,10 @@ STAGE PLANS:
#### A masked pattern was here ####
name default.bucket_big
numFiles 4
+ numRows 0
partition_columns ds
partition_columns.types string
+ rawDataSize 0
serialization.ddl struct bucket_big { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -713,8 +725,10 @@ STAGE PLANS:
#### A masked pattern was here ####
name default.bucket_small
numFiles 2
+ numRows 0
partition_columns ds
partition_columns.types string
+ rawDataSize 0
serialization.ddl struct bucket_small { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -796,8 +810,10 @@ STAGE PLANS:
#### A masked pattern was here ####
name default.bucket_big
numFiles 4
+ numRows 0
partition_columns ds
partition_columns.types string
+ rawDataSize 0
serialization.ddl struct bucket_big { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -840,8 +856,10 @@ STAGE PLANS:
#### A masked pattern was here ####
name default.bucket_big
numFiles 4
+ numRows 0
partition_columns ds
partition_columns.types string
+ rawDataSize 0
serialization.ddl struct bucket_big { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -981,8 +999,10 @@ STAGE PLANS:
#### A masked pattern was here ####
name default.bucket_small
numFiles 2
+ numRows 0
partition_columns ds
partition_columns.types string
+ rawDataSize 0
serialization.ddl struct bucket_small { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -1067,8 +1087,10 @@ STAGE PLANS:
#### A masked pattern was here ####
name default.bucket_big
numFiles 4
+ numRows 0
partition_columns ds
partition_columns.types string
+ rawDataSize 0
serialization.ddl struct bucket_big { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -1111,8 +1133,10 @@ STAGE PLANS:
#### A masked pattern was here ####
name default.bucket_big
numFiles 4
+ numRows 0
partition_columns ds
partition_columns.types string
+ rawDataSize 0
serialization.ddl struct bucket_big { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -1179,8 +1203,10 @@ STAGE PLANS:
#### A masked pattern was here ####
name default.bucket_big
numFiles 4
+ numRows 0
partition_columns ds
partition_columns.types string
+ rawDataSize 0
serialization.ddl struct bucket_big { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -1223,8 +1249,10 @@ STAGE PLANS:
#### A masked pattern was here ####
name default.bucket_big
numFiles 4
+ numRows 0
partition_columns ds
partition_columns.types string
+ rawDataSize 0
serialization.ddl struct bucket_big { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
http://git-wip-us.apache.org/repos/asf/hive/blob/244ce09c/ql/src/test/results/clientpositive/tez/auto_sortmerge_join_12.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/tez/auto_sortmerge_join_12.q.out b/ql/src/test/results/clientpositive/tez/auto_sortmerge_join_12.q.out
index 578880d..dfcf7ea 100644
--- a/ql/src/test/results/clientpositive/tez/auto_sortmerge_join_12.q.out
+++ b/ql/src/test/results/clientpositive/tez/auto_sortmerge_join_12.q.out
@@ -197,8 +197,10 @@ STAGE PLANS:
#### A masked pattern was here ####
name default.bucket_small
numFiles 2
+ numRows 0
partition_columns ds
partition_columns.types string
+ rawDataSize 0
serialization.ddl struct bucket_small { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -269,8 +271,10 @@ STAGE PLANS:
#### A masked pattern was here ####
name default.bucket_medium
numFiles 3
+ numRows 0
partition_columns ds
partition_columns.types string
+ rawDataSize 0
serialization.ddl struct bucket_medium { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -371,8 +375,10 @@ STAGE PLANS:
#### A masked pattern was here ####
name default.bucket_big
numFiles 4
+ numRows 0
partition_columns ds
partition_columns.types string
+ rawDataSize 0
serialization.ddl struct bucket_big { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -416,8 +422,10 @@ STAGE PLANS:
#### A masked pattern was here ####
name default.bucket_big
numFiles 4
+ numRows 0
partition_columns ds
partition_columns.types string
+ rawDataSize 0
serialization.ddl struct bucket_big { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -481,8 +489,10 @@ STAGE PLANS:
#### A masked pattern was here ####
name default.bucket_medium
numFiles 3
+ numRows 0
partition_columns ds
partition_columns.types string
+ rawDataSize 0
serialization.ddl struct bucket_medium { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe