You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hive.apache.org by px...@apache.org on 2016/05/09 17:41:46 UTC
[04/21] hive git commit: HIVE-13341: Stats state is not captured
correctly: differentiate load table and create table (Pengcheng Xiong,
reviewed by Ashutosh Chauhan)
http://git-wip-us.apache.org/repos/asf/hive/blob/244ce09c/ql/src/test/results/clientpositive/tez/auto_sortmerge_join_2.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/tez/auto_sortmerge_join_2.q.out b/ql/src/test/results/clientpositive/tez/auto_sortmerge_join_2.q.out
index c283738..5d93fdc 100644
--- a/ql/src/test/results/clientpositive/tez/auto_sortmerge_join_2.q.out
+++ b/ql/src/test/results/clientpositive/tez/auto_sortmerge_join_2.q.out
@@ -159,8 +159,10 @@ STAGE PLANS:
#### A masked pattern was here ####
name default.bucket_big
numFiles 2
+ numRows 0
partition_columns ds
partition_columns.types string
+ rawDataSize 0
serialization.ddl struct bucket_big { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -204,8 +206,10 @@ STAGE PLANS:
#### A masked pattern was here ####
name default.bucket_big
numFiles 2
+ numRows 0
partition_columns ds
partition_columns.types string
+ rawDataSize 0
serialization.ddl struct bucket_big { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -277,8 +281,10 @@ STAGE PLANS:
#### A masked pattern was here ####
name default.bucket_small
numFiles 4
+ numRows 0
partition_columns ds
partition_columns.types string
+ rawDataSize 0
serialization.ddl struct bucket_small { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -438,8 +444,10 @@ STAGE PLANS:
#### A masked pattern was here ####
name default.bucket_big
numFiles 2
+ numRows 0
partition_columns ds
partition_columns.types string
+ rawDataSize 0
serialization.ddl struct bucket_big { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -483,8 +491,10 @@ STAGE PLANS:
#### A masked pattern was here ####
name default.bucket_big
numFiles 2
+ numRows 0
partition_columns ds
partition_columns.types string
+ rawDataSize 0
serialization.ddl struct bucket_big { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -556,8 +566,10 @@ STAGE PLANS:
#### A masked pattern was here ####
name default.bucket_small
numFiles 4
+ numRows 0
partition_columns ds
partition_columns.types string
+ rawDataSize 0
serialization.ddl struct bucket_small { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
http://git-wip-us.apache.org/repos/asf/hive/blob/244ce09c/ql/src/test/results/clientpositive/tez/auto_sortmerge_join_3.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/tez/auto_sortmerge_join_3.q.out b/ql/src/test/results/clientpositive/tez/auto_sortmerge_join_3.q.out
index 351fd20..15e9ef8 100644
--- a/ql/src/test/results/clientpositive/tez/auto_sortmerge_join_3.q.out
+++ b/ql/src/test/results/clientpositive/tez/auto_sortmerge_join_3.q.out
@@ -143,8 +143,10 @@ STAGE PLANS:
#### A masked pattern was here ####
name default.bucket_small
numFiles 2
+ numRows 0
partition_columns ds
partition_columns.types string
+ rawDataSize 0
serialization.ddl struct bucket_small { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -188,8 +190,10 @@ STAGE PLANS:
#### A masked pattern was here ####
name default.bucket_small
numFiles 2
+ numRows 0
partition_columns ds
partition_columns.types string
+ rawDataSize 0
serialization.ddl struct bucket_small { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -277,8 +281,10 @@ STAGE PLANS:
#### A masked pattern was here ####
name default.bucket_big
numFiles 4
+ numRows 0
partition_columns ds
partition_columns.types string
+ rawDataSize 0
serialization.ddl struct bucket_big { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -436,8 +442,10 @@ STAGE PLANS:
#### A masked pattern was here ####
name default.bucket_big
numFiles 4
+ numRows 0
partition_columns ds
partition_columns.types string
+ rawDataSize 0
serialization.ddl struct bucket_big { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -508,8 +516,10 @@ STAGE PLANS:
#### A masked pattern was here ####
name default.bucket_small
numFiles 2
+ numRows 0
partition_columns ds
partition_columns.types string
+ rawDataSize 0
serialization.ddl struct bucket_small { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -553,8 +563,10 @@ STAGE PLANS:
#### A masked pattern was here ####
name default.bucket_small
numFiles 2
+ numRows 0
partition_columns ds
partition_columns.types string
+ rawDataSize 0
serialization.ddl struct bucket_small { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -713,8 +725,10 @@ STAGE PLANS:
#### A masked pattern was here ####
name default.bucket_big
numFiles 4
+ numRows 0
partition_columns ds
partition_columns.types string
+ rawDataSize 0
serialization.ddl struct bucket_big { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -785,8 +799,10 @@ STAGE PLANS:
#### A masked pattern was here ####
name default.bucket_small
numFiles 2
+ numRows 0
partition_columns ds
partition_columns.types string
+ rawDataSize 0
serialization.ddl struct bucket_small { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -830,8 +846,10 @@ STAGE PLANS:
#### A masked pattern was here ####
name default.bucket_small
numFiles 2
+ numRows 0
partition_columns ds
partition_columns.types string
+ rawDataSize 0
serialization.ddl struct bucket_small { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
http://git-wip-us.apache.org/repos/asf/hive/blob/244ce09c/ql/src/test/results/clientpositive/tez/auto_sortmerge_join_4.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/tez/auto_sortmerge_join_4.q.out b/ql/src/test/results/clientpositive/tez/auto_sortmerge_join_4.q.out
index 2d9cdf8..f43cad2 100644
--- a/ql/src/test/results/clientpositive/tez/auto_sortmerge_join_4.q.out
+++ b/ql/src/test/results/clientpositive/tez/auto_sortmerge_join_4.q.out
@@ -159,8 +159,10 @@ STAGE PLANS:
#### A masked pattern was here ####
name default.bucket_small
numFiles 4
+ numRows 0
partition_columns ds
partition_columns.types string
+ rawDataSize 0
serialization.ddl struct bucket_small { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -204,8 +206,10 @@ STAGE PLANS:
#### A masked pattern was here ####
name default.bucket_small
numFiles 4
+ numRows 0
partition_columns ds
partition_columns.types string
+ rawDataSize 0
serialization.ddl struct bucket_small { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -293,8 +297,10 @@ STAGE PLANS:
#### A masked pattern was here ####
name default.bucket_big
numFiles 2
+ numRows 0
partition_columns ds
partition_columns.types string
+ rawDataSize 0
serialization.ddl struct bucket_big { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -452,8 +458,10 @@ STAGE PLANS:
#### A masked pattern was here ####
name default.bucket_big
numFiles 2
+ numRows 0
partition_columns ds
partition_columns.types string
+ rawDataSize 0
serialization.ddl struct bucket_big { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -524,8 +532,10 @@ STAGE PLANS:
#### A masked pattern was here ####
name default.bucket_small
numFiles 4
+ numRows 0
partition_columns ds
partition_columns.types string
+ rawDataSize 0
serialization.ddl struct bucket_small { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -569,8 +579,10 @@ STAGE PLANS:
#### A masked pattern was here ####
name default.bucket_small
numFiles 4
+ numRows 0
partition_columns ds
partition_columns.types string
+ rawDataSize 0
serialization.ddl struct bucket_small { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -729,8 +741,10 @@ STAGE PLANS:
#### A masked pattern was here ####
name default.bucket_big
numFiles 2
+ numRows 0
partition_columns ds
partition_columns.types string
+ rawDataSize 0
serialization.ddl struct bucket_big { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -801,8 +815,10 @@ STAGE PLANS:
#### A masked pattern was here ####
name default.bucket_small
numFiles 4
+ numRows 0
partition_columns ds
partition_columns.types string
+ rawDataSize 0
serialization.ddl struct bucket_small { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -846,8 +862,10 @@ STAGE PLANS:
#### A masked pattern was here ####
name default.bucket_small
numFiles 4
+ numRows 0
partition_columns ds
partition_columns.types string
+ rawDataSize 0
serialization.ddl struct bucket_small { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
http://git-wip-us.apache.org/repos/asf/hive/blob/244ce09c/ql/src/test/results/clientpositive/tez/auto_sortmerge_join_5.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/tez/auto_sortmerge_join_5.q.out b/ql/src/test/results/clientpositive/tez/auto_sortmerge_join_5.q.out
index 8bc203a..3f019bc 100644
--- a/ql/src/test/results/clientpositive/tez/auto_sortmerge_join_5.q.out
+++ b/ql/src/test/results/clientpositive/tez/auto_sortmerge_join_5.q.out
@@ -119,6 +119,8 @@ STAGE PLANS:
#### A masked pattern was here ####
name default.bucket_small
numFiles 4
+ numRows 0
+ rawDataSize 0
serialization.ddl struct bucket_small { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -138,6 +140,8 @@ STAGE PLANS:
#### A masked pattern was here ####
name default.bucket_small
numFiles 4
+ numRows 0
+ rawDataSize 0
serialization.ddl struct bucket_small { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -198,6 +202,8 @@ STAGE PLANS:
#### A masked pattern was here ####
name default.bucket_big
numFiles 2
+ numRows 0
+ rawDataSize 0
serialization.ddl struct bucket_big { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -217,6 +223,8 @@ STAGE PLANS:
#### A masked pattern was here ####
name default.bucket_big
numFiles 2
+ numRows 0
+ rawDataSize 0
serialization.ddl struct bucket_big { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -322,6 +330,8 @@ STAGE PLANS:
#### A masked pattern was here ####
name default.bucket_small
numFiles 4
+ numRows 0
+ rawDataSize 0
serialization.ddl struct bucket_small { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -341,6 +351,8 @@ STAGE PLANS:
#### A masked pattern was here ####
name default.bucket_small
numFiles 4
+ numRows 0
+ rawDataSize 0
serialization.ddl struct bucket_small { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -401,6 +413,8 @@ STAGE PLANS:
#### A masked pattern was here ####
name default.bucket_big
numFiles 2
+ numRows 0
+ rawDataSize 0
serialization.ddl struct bucket_big { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -420,6 +434,8 @@ STAGE PLANS:
#### A masked pattern was here ####
name default.bucket_big
numFiles 2
+ numRows 0
+ rawDataSize 0
serialization.ddl struct bucket_big { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -551,6 +567,8 @@ STAGE PLANS:
#### A masked pattern was here ####
name default.bucket_big
numFiles 2
+ numRows 0
+ rawDataSize 0
serialization.ddl struct bucket_big { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -570,6 +588,8 @@ STAGE PLANS:
#### A masked pattern was here ####
name default.bucket_big
numFiles 2
+ numRows 0
+ rawDataSize 0
serialization.ddl struct bucket_big { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -620,6 +640,8 @@ STAGE PLANS:
#### A masked pattern was here ####
name default.bucket_small
numFiles 4
+ numRows 0
+ rawDataSize 0
serialization.ddl struct bucket_small { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -639,6 +661,8 @@ STAGE PLANS:
#### A masked pattern was here ####
name default.bucket_small
numFiles 4
+ numRows 0
+ rawDataSize 0
serialization.ddl struct bucket_small { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
http://git-wip-us.apache.org/repos/asf/hive/blob/244ce09c/ql/src/test/results/clientpositive/tez/auto_sortmerge_join_7.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/tez/auto_sortmerge_join_7.q.out b/ql/src/test/results/clientpositive/tez/auto_sortmerge_join_7.q.out
index 18fc95c..cb17d38 100644
--- a/ql/src/test/results/clientpositive/tez/auto_sortmerge_join_7.q.out
+++ b/ql/src/test/results/clientpositive/tez/auto_sortmerge_join_7.q.out
@@ -176,8 +176,10 @@ STAGE PLANS:
#### A masked pattern was here ####
name default.bucket_small
numFiles 4
+ numRows 0
partition_columns ds
partition_columns.types string
+ rawDataSize 0
serialization.ddl struct bucket_small { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -221,8 +223,10 @@ STAGE PLANS:
#### A masked pattern was here ####
name default.bucket_small
numFiles 4
+ numRows 0
partition_columns ds
partition_columns.types string
+ rawDataSize 0
serialization.ddl struct bucket_small { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -310,8 +314,10 @@ STAGE PLANS:
#### A masked pattern was here ####
name default.bucket_big
numFiles 2
+ numRows 0
partition_columns ds
partition_columns.types string
+ rawDataSize 0
serialization.ddl struct bucket_big { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -355,8 +361,10 @@ STAGE PLANS:
#### A masked pattern was here ####
name default.bucket_big
numFiles 2
+ numRows 0
partition_columns ds
partition_columns.types string
+ rawDataSize 0
serialization.ddl struct bucket_big { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -517,8 +525,10 @@ STAGE PLANS:
#### A masked pattern was here ####
name default.bucket_big
numFiles 2
+ numRows 0
partition_columns ds
partition_columns.types string
+ rawDataSize 0
serialization.ddl struct bucket_big { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -562,8 +572,10 @@ STAGE PLANS:
#### A masked pattern was here ####
name default.bucket_big
numFiles 2
+ numRows 0
partition_columns ds
partition_columns.types string
+ rawDataSize 0
serialization.ddl struct bucket_big { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -635,8 +647,10 @@ STAGE PLANS:
#### A masked pattern was here ####
name default.bucket_small
numFiles 4
+ numRows 0
partition_columns ds
partition_columns.types string
+ rawDataSize 0
serialization.ddl struct bucket_small { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -680,8 +694,10 @@ STAGE PLANS:
#### A masked pattern was here ####
name default.bucket_small
numFiles 4
+ numRows 0
partition_columns ds
partition_columns.types string
+ rawDataSize 0
serialization.ddl struct bucket_small { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -842,8 +858,10 @@ STAGE PLANS:
#### A masked pattern was here ####
name default.bucket_big
numFiles 2
+ numRows 0
partition_columns ds
partition_columns.types string
+ rawDataSize 0
serialization.ddl struct bucket_big { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -887,8 +905,10 @@ STAGE PLANS:
#### A masked pattern was here ####
name default.bucket_big
numFiles 2
+ numRows 0
partition_columns ds
partition_columns.types string
+ rawDataSize 0
serialization.ddl struct bucket_big { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -960,8 +980,10 @@ STAGE PLANS:
#### A masked pattern was here ####
name default.bucket_small
numFiles 4
+ numRows 0
partition_columns ds
partition_columns.types string
+ rawDataSize 0
serialization.ddl struct bucket_small { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -1005,8 +1027,10 @@ STAGE PLANS:
#### A masked pattern was here ####
name default.bucket_small
numFiles 4
+ numRows 0
partition_columns ds
partition_columns.types string
+ rawDataSize 0
serialization.ddl struct bucket_small { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
http://git-wip-us.apache.org/repos/asf/hive/blob/244ce09c/ql/src/test/results/clientpositive/tez/auto_sortmerge_join_8.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/tez/auto_sortmerge_join_8.q.out b/ql/src/test/results/clientpositive/tez/auto_sortmerge_join_8.q.out
index 9a1010c..d24b3b2 100644
--- a/ql/src/test/results/clientpositive/tez/auto_sortmerge_join_8.q.out
+++ b/ql/src/test/results/clientpositive/tez/auto_sortmerge_join_8.q.out
@@ -176,8 +176,10 @@ STAGE PLANS:
#### A masked pattern was here ####
name default.bucket_small
numFiles 2
+ numRows 0
partition_columns ds
partition_columns.types string
+ rawDataSize 0
serialization.ddl struct bucket_small { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -221,8 +223,10 @@ STAGE PLANS:
#### A masked pattern was here ####
name default.bucket_small
numFiles 2
+ numRows 0
partition_columns ds
partition_columns.types string
+ rawDataSize 0
serialization.ddl struct bucket_small { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -310,8 +314,10 @@ STAGE PLANS:
#### A masked pattern was here ####
name default.bucket_big
numFiles 4
+ numRows 0
partition_columns ds
partition_columns.types string
+ rawDataSize 0
serialization.ddl struct bucket_big { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -355,8 +361,10 @@ STAGE PLANS:
#### A masked pattern was here ####
name default.bucket_big
numFiles 4
+ numRows 0
partition_columns ds
partition_columns.types string
+ rawDataSize 0
serialization.ddl struct bucket_big { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -517,8 +525,10 @@ STAGE PLANS:
#### A masked pattern was here ####
name default.bucket_big
numFiles 4
+ numRows 0
partition_columns ds
partition_columns.types string
+ rawDataSize 0
serialization.ddl struct bucket_big { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -562,8 +572,10 @@ STAGE PLANS:
#### A masked pattern was here ####
name default.bucket_big
numFiles 4
+ numRows 0
partition_columns ds
partition_columns.types string
+ rawDataSize 0
serialization.ddl struct bucket_big { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -635,8 +647,10 @@ STAGE PLANS:
#### A masked pattern was here ####
name default.bucket_small
numFiles 2
+ numRows 0
partition_columns ds
partition_columns.types string
+ rawDataSize 0
serialization.ddl struct bucket_small { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -680,8 +694,10 @@ STAGE PLANS:
#### A masked pattern was here ####
name default.bucket_small
numFiles 2
+ numRows 0
partition_columns ds
partition_columns.types string
+ rawDataSize 0
serialization.ddl struct bucket_small { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -844,8 +860,10 @@ STAGE PLANS:
#### A masked pattern was here ####
name default.bucket_big
numFiles 4
+ numRows 0
partition_columns ds
partition_columns.types string
+ rawDataSize 0
serialization.ddl struct bucket_big { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -889,8 +907,10 @@ STAGE PLANS:
#### A masked pattern was here ####
name default.bucket_big
numFiles 4
+ numRows 0
partition_columns ds
partition_columns.types string
+ rawDataSize 0
serialization.ddl struct bucket_big { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -962,8 +982,10 @@ STAGE PLANS:
#### A masked pattern was here ####
name default.bucket_small
numFiles 2
+ numRows 0
partition_columns ds
partition_columns.types string
+ rawDataSize 0
serialization.ddl struct bucket_small { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -1007,8 +1029,10 @@ STAGE PLANS:
#### A masked pattern was here ####
name default.bucket_small
numFiles 2
+ numRows 0
partition_columns ds
partition_columns.types string
+ rawDataSize 0
serialization.ddl struct bucket_small { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
http://git-wip-us.apache.org/repos/asf/hive/blob/244ce09c/ql/src/test/results/clientpositive/tez/bucket2.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/tez/bucket2.q.out b/ql/src/test/results/clientpositive/tez/bucket2.q.out
index e9432a8..800edf3 100644
--- a/ql/src/test/results/clientpositive/tez/bucket2.q.out
+++ b/ql/src/test/results/clientpositive/tez/bucket2.q.out
@@ -117,6 +117,7 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
+ COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
bucket_count 2
bucket_field_name key
columns key,value
@@ -124,9 +125,13 @@ STAGE PLANS:
columns.types int:string
#### A masked pattern was here ####
name default.bucket2_1
+ numFiles 0
+ numRows 0
+ rawDataSize 0
serialization.ddl struct bucket2_1 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ totalSize 0
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.bucket2_1
@@ -146,6 +151,7 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
+ COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
bucket_count 2
bucket_field_name key
columns key,value
@@ -153,9 +159,13 @@ STAGE PLANS:
columns.types int:string
#### A masked pattern was here ####
name default.bucket2_1
+ numFiles 0
+ numRows 0
+ rawDataSize 0
serialization.ddl struct bucket2_1 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ totalSize 0
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.bucket2_1
http://git-wip-us.apache.org/repos/asf/hive/blob/244ce09c/ql/src/test/results/clientpositive/tez/bucket4.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/tez/bucket4.q.out b/ql/src/test/results/clientpositive/tez/bucket4.q.out
index a0f1177..4291e44 100644
--- a/ql/src/test/results/clientpositive/tez/bucket4.q.out
+++ b/ql/src/test/results/clientpositive/tez/bucket4.q.out
@@ -114,6 +114,7 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
+ COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
SORTBUCKETCOLSPREFIX TRUE
bucket_count 2
bucket_field_name key
@@ -122,9 +123,13 @@ STAGE PLANS:
columns.types int:string
#### A masked pattern was here ####
name default.bucket4_1
+ numFiles 0
+ numRows 0
+ rawDataSize 0
serialization.ddl struct bucket4_1 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ totalSize 0
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.bucket4_1
@@ -144,6 +149,7 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
+ COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
SORTBUCKETCOLSPREFIX TRUE
bucket_count 2
bucket_field_name key
@@ -152,9 +158,13 @@ STAGE PLANS:
columns.types int:string
#### A masked pattern was here ####
name default.bucket4_1
+ numFiles 0
+ numRows 0
+ rawDataSize 0
serialization.ddl struct bucket4_1 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ totalSize 0
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.bucket4_1
http://git-wip-us.apache.org/repos/asf/hive/blob/244ce09c/ql/src/test/results/clientpositive/tez/column_names_with_leading_and_trailing_spaces.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/tez/column_names_with_leading_and_trailing_spaces.q.out b/ql/src/test/results/clientpositive/tez/column_names_with_leading_and_trailing_spaces.q.out
index 46c285e..18314b5 100644
--- a/ql/src/test/results/clientpositive/tez/column_names_with_leading_and_trailing_spaces.q.out
+++ b/ql/src/test/results/clientpositive/tez/column_names_with_leading_and_trailing_spaces.q.out
@@ -25,6 +25,11 @@ Retention: 0
#### A masked pattern was here ####
Table Type: MANAGED_TABLE
Table Parameters:
+ COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\"}
+ numFiles 0
+ numRows 0
+ rawDataSize 0
+ totalSize 0
#### A masked pattern was here ####
# Storage Information
http://git-wip-us.apache.org/repos/asf/hive/blob/244ce09c/ql/src/test/results/clientpositive/tez/disable_merge_for_bucketing.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/tez/disable_merge_for_bucketing.q.out b/ql/src/test/results/clientpositive/tez/disable_merge_for_bucketing.q.out
index bb62e1f..fb71214 100644
--- a/ql/src/test/results/clientpositive/tez/disable_merge_for_bucketing.q.out
+++ b/ql/src/test/results/clientpositive/tez/disable_merge_for_bucketing.q.out
@@ -113,6 +113,7 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
+ COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
bucket_count 2
bucket_field_name key
columns key,value
@@ -120,9 +121,13 @@ STAGE PLANS:
columns.types int:string
#### A masked pattern was here ####
name default.bucket2_1
+ numFiles 0
+ numRows 0
+ rawDataSize 0
serialization.ddl struct bucket2_1 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ totalSize 0
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.bucket2_1
@@ -142,6 +147,7 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
+ COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
bucket_count 2
bucket_field_name key
columns key,value
@@ -149,9 +155,13 @@ STAGE PLANS:
columns.types int:string
#### A masked pattern was here ####
name default.bucket2_1
+ numFiles 0
+ numRows 0
+ rawDataSize 0
serialization.ddl struct bucket2_1 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ totalSize 0
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.bucket2_1
http://git-wip-us.apache.org/repos/asf/hive/blob/244ce09c/ql/src/test/results/clientpositive/tez/explainuser_1.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/tez/explainuser_1.q.out b/ql/src/test/results/clientpositive/tez/explainuser_1.q.out
index bdb8830..965577e 100644
--- a/ql/src/test/results/clientpositive/tez/explainuser_1.q.out
+++ b/ql/src/test/results/clientpositive/tez/explainuser_1.q.out
@@ -186,9 +186,9 @@ Stage-0
SHUFFLE [RS_5]
Group By Operator [GBY_4] (rows=1 width=8)
Output:["_col0"],aggregations:["count(1)"]
- Select Operator [SEL_2] (rows=1 width=2515)
- TableScan [TS_0] (rows=1 width=2515)
- default@src_orc_merge_test_part,src_orc_merge_test_part,Tbl:PARTIAL,Col:NONE
+ Select Operator [SEL_2] (rows=500 width=94)
+ TableScan [TS_0] (rows=500 width=94)
+ default@src_orc_merge_test_part,src_orc_merge_test_part,Tbl:COMPLETE,Col:NONE
PREHOOK: query: explain select sum(hash(key)), sum(hash(value)) from src_orc_merge_test_part where ds='2012-01-03' and ts='2012-01-03+14:46:31'
PREHOOK: type: QUERY
@@ -211,9 +211,9 @@ Stage-0
SHUFFLE [RS_5]
Group By Operator [GBY_4] (rows=1 width=16)
Output:["_col0","_col1"],aggregations:["sum(_col0)","sum(_col1)"]
- Select Operator [SEL_2] (rows=24 width=104)
+ Select Operator [SEL_2] (rows=500 width=94)
Output:["_col0","_col1"]
- TableScan [TS_0] (rows=24 width=104)
+ TableScan [TS_0] (rows=500 width=94)
default@src_orc_merge_test_part,src_orc_merge_test_part,Tbl:COMPLETE,Col:NONE,Output:["key","value"]
PREHOOK: query: drop table src_orc_merge_test_part
@@ -3238,8 +3238,8 @@ Stage-0
Output:["_col0"],aggregations:["count(1)"]
<-Map 1 [SIMPLE_EDGE]
SHUFFLE [RS_3]
- Select Operator [SEL_1] (rows=1 width=171)
- TableScan [TS_0] (rows=1 width=171)
+ Select Operator [SEL_1] (rows=5 width=6)
+ TableScan [TS_0] (rows=5 width=6)
default@tgt_rc_merge_test,tgt_rc_merge_test,Tbl:COMPLETE,Col:COMPLETE
PREHOOK: query: explain select sum(hash(key)), sum(hash(value)) from tgt_rc_merge_test
@@ -3261,9 +3261,9 @@ Stage-0
Output:["_col0","_col1"],aggregations:["sum(VALUE._col0)","sum(VALUE._col1)"]
<-Map 1 [SIMPLE_EDGE]
SHUFFLE [RS_3]
- Select Operator [SEL_1] (rows=1 width=171)
+ Select Operator [SEL_1] (rows=5 width=6)
Output:["_col0","_col1"]
- TableScan [TS_0] (rows=1 width=171)
+ TableScan [TS_0] (rows=5 width=6)
default@tgt_rc_merge_test,tgt_rc_merge_test,Tbl:COMPLETE,Col:NONE,Output:["key","value"]
PREHOOK: query: drop table src_rc_merge_test
http://git-wip-us.apache.org/repos/asf/hive/blob/244ce09c/ql/src/test/results/clientpositive/tez/metadataonly1.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/tez/metadataonly1.q.out b/ql/src/test/results/clientpositive/tez/metadataonly1.q.out
index e43a7e4..15f5ed5 100644
--- a/ql/src/test/results/clientpositive/tez/metadataonly1.q.out
+++ b/ql/src/test/results/clientpositive/tez/metadataonly1.q.out
@@ -145,17 +145,22 @@ STAGE PLANS:
partition values:
ds 1
properties:
+ COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
bucket_count -1
columns a,b
columns.comments
columns.types int:double
#### A masked pattern was here ####
name default.test1
+ numFiles 0
+ numRows 0
partition_columns ds
partition_columns.types string
+ rawDataSize 0
serialization.ddl struct test1 { i32 a, double b}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.NullStructSerDe
+ totalSize 0
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.NullStructSerDe
@@ -276,17 +281,22 @@ STAGE PLANS:
partition values:
ds 1
properties:
+ COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
bucket_count -1
columns a,b
columns.comments
columns.types int:double
#### A masked pattern was here ####
name default.test1
+ numFiles 0
+ numRows 0
partition_columns ds
partition_columns.types string
+ rawDataSize 0
serialization.ddl struct test1 { i32 a, double b}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.NullStructSerDe
+ totalSize 0
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.NullStructSerDe
@@ -407,17 +417,22 @@ STAGE PLANS:
partition values:
ds 1
properties:
+ COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
bucket_count -1
columns a,b
columns.comments
columns.types int:double
#### A masked pattern was here ####
name default.test1
+ numFiles 0
+ numRows 0
partition_columns ds
partition_columns.types string
+ rawDataSize 0
serialization.ddl struct test1 { i32 a, double b}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ totalSize 0
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -545,17 +560,22 @@ STAGE PLANS:
partition values:
ds 1
properties:
+ COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
bucket_count -1
columns a,b
columns.comments
columns.types int:double
#### A masked pattern was here ####
name default.test1
+ numFiles 0
+ numRows 0
partition_columns ds
partition_columns.types string
+ rawDataSize 0
serialization.ddl struct test1 { i32 a, double b}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ totalSize 0
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -585,17 +605,22 @@ STAGE PLANS:
partition values:
ds 2
properties:
+ COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
bucket_count -1
columns a,b
columns.comments
columns.types int:double
#### A masked pattern was here ####
name default.test1
+ numFiles 0
+ numRows 0
partition_columns ds
partition_columns.types string
+ rawDataSize 0
serialization.ddl struct test1 { i32 a, double b}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ totalSize 0
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -653,17 +678,22 @@ STAGE PLANS:
partition values:
ds 1
properties:
+ COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
bucket_count -1
columns a,b
columns.comments
columns.types int:double
#### A masked pattern was here ####
name default.test1
+ numFiles 0
+ numRows 0
partition_columns ds
partition_columns.types string
+ rawDataSize 0
serialization.ddl struct test1 { i32 a, double b}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ totalSize 0
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -693,17 +723,22 @@ STAGE PLANS:
partition values:
ds 2
properties:
+ COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
bucket_count -1
columns a,b
columns.comments
columns.types int:double
#### A masked pattern was here ####
name default.test1
+ numFiles 0
+ numRows 0
partition_columns ds
partition_columns.types string
+ rawDataSize 0
serialization.ddl struct test1 { i32 a, double b}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ totalSize 0
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -918,17 +953,22 @@ STAGE PLANS:
ds 1
hr 1
properties:
+ COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
bucket_count -1
columns a,b
columns.comments
columns.types int:double
#### A masked pattern was here ####
name default.test2
+ numFiles 0
+ numRows 0
partition_columns ds/hr
partition_columns.types string:string
+ rawDataSize 0
serialization.ddl struct test2 { i32 a, double b}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.NullStructSerDe
+ totalSize 0
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.NullStructSerDe
@@ -958,17 +998,22 @@ STAGE PLANS:
ds 1
hr 2
properties:
+ COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
bucket_count -1
columns a,b
columns.comments
columns.types int:double
#### A masked pattern was here ####
name default.test2
+ numFiles 0
+ numRows 0
partition_columns ds/hr
partition_columns.types string:string
+ rawDataSize 0
serialization.ddl struct test2 { i32 a, double b}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.NullStructSerDe
+ totalSize 0
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.NullStructSerDe
@@ -998,17 +1043,22 @@ STAGE PLANS:
ds 1
hr 3
properties:
+ COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
bucket_count -1
columns a,b
columns.comments
columns.types int:double
#### A masked pattern was here ####
name default.test2
+ numFiles 0
+ numRows 0
partition_columns ds/hr
partition_columns.types string:string
+ rawDataSize 0
serialization.ddl struct test2 { i32 a, double b}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.NullStructSerDe
+ totalSize 0
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.NullStructSerDe
@@ -1140,17 +1190,22 @@ STAGE PLANS:
ds 1
hr 1
properties:
+ COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
bucket_count -1
columns a,b
columns.comments
columns.types int:double
#### A masked pattern was here ####
name default.test2
+ numFiles 0
+ numRows 0
partition_columns ds/hr
partition_columns.types string:string
+ rawDataSize 0
serialization.ddl struct test2 { i32 a, double b}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ totalSize 0
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -1181,17 +1236,22 @@ STAGE PLANS:
ds 1
hr 2
properties:
+ COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
bucket_count -1
columns a,b
columns.comments
columns.types int:double
#### A masked pattern was here ####
name default.test2
+ numFiles 0
+ numRows 0
partition_columns ds/hr
partition_columns.types string:string
+ rawDataSize 0
serialization.ddl struct test2 { i32 a, double b}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ totalSize 0
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -1222,17 +1282,22 @@ STAGE PLANS:
ds 1
hr 3
properties:
+ COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
bucket_count -1
columns a,b
columns.comments
columns.types int:double
#### A masked pattern was here ####
name default.test2
+ numFiles 0
+ numRows 0
partition_columns ds/hr
partition_columns.types string:string
+ rawDataSize 0
serialization.ddl struct test2 { i32 a, double b}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ totalSize 0
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
@@ -1359,17 +1424,22 @@ STAGE PLANS:
partition values:
ds 1
properties:
+ COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
bucket_count -1
columns a,b
columns.comments
columns.types int:double
#### A masked pattern was here ####
name default.test1
+ numFiles 0
+ numRows 0
partition_columns ds
partition_columns.types string
+ rawDataSize 0
serialization.ddl struct test1 { i32 a, double b}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.NullStructSerDe
+ totalSize 0
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.NullStructSerDe
@@ -1398,17 +1468,22 @@ STAGE PLANS:
partition values:
ds 2
properties:
+ COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
bucket_count -1
columns a,b
columns.comments
columns.types int:double
#### A masked pattern was here ####
name default.test1
+ numFiles 0
+ numRows 0
partition_columns ds
partition_columns.types string
+ rawDataSize 0
serialization.ddl struct test1 { i32 a, double b}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.NullStructSerDe
+ totalSize 0
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.NullStructSerDe
@@ -1589,17 +1664,22 @@ STAGE PLANS:
ds 01:10:10
hr 01
properties:
+ COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
bucket_count -1
columns a,b
columns.comments
columns.types int:double
#### A masked pattern was here ####
name default.test2
+ numFiles 0
+ numRows 0
partition_columns ds/hr
partition_columns.types string:string
+ rawDataSize 0
serialization.ddl struct test2 { i32 a, double b}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.NullStructSerDe
+ totalSize 0
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.NullStructSerDe
@@ -1629,17 +1709,22 @@ STAGE PLANS:
ds 01:10:20
hr 02
properties:
+ COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
bucket_count -1
columns a,b
columns.comments
columns.types int:double
#### A masked pattern was here ####
name default.test2
+ numFiles 0
+ numRows 0
partition_columns ds/hr
partition_columns.types string:string
+ rawDataSize 0
serialization.ddl struct test2 { i32 a, double b}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.NullStructSerDe
+ totalSize 0
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.NullStructSerDe
@@ -1669,17 +1754,22 @@ STAGE PLANS:
ds 1
hr 1
properties:
+ COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
bucket_count -1
columns a,b
columns.comments
columns.types int:double
#### A masked pattern was here ####
name default.test2
+ numFiles 0
+ numRows 0
partition_columns ds/hr
partition_columns.types string:string
+ rawDataSize 0
serialization.ddl struct test2 { i32 a, double b}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.NullStructSerDe
+ totalSize 0
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.NullStructSerDe
@@ -1709,17 +1799,22 @@ STAGE PLANS:
ds 1
hr 2
properties:
+ COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
bucket_count -1
columns a,b
columns.comments
columns.types int:double
#### A masked pattern was here ####
name default.test2
+ numFiles 0
+ numRows 0
partition_columns ds/hr
partition_columns.types string:string
+ rawDataSize 0
serialization.ddl struct test2 { i32 a, double b}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.NullStructSerDe
+ totalSize 0
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.NullStructSerDe
@@ -1749,17 +1844,22 @@ STAGE PLANS:
ds 1
hr 3
properties:
+ COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
bucket_count -1
columns a,b
columns.comments
columns.types int:double
#### A masked pattern was here ####
name default.test2
+ numFiles 0
+ numRows 0
partition_columns ds/hr
partition_columns.types string:string
+ rawDataSize 0
serialization.ddl struct test2 { i32 a, double b}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.NullStructSerDe
+ totalSize 0
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.NullStructSerDe
http://git-wip-us.apache.org/repos/asf/hive/blob/244ce09c/ql/src/test/results/clientpositive/tez/sample1.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/tez/sample1.q.out b/ql/src/test/results/clientpositive/tez/sample1.q.out
index 009969e..2120a1ff 100644
--- a/ql/src/test/results/clientpositive/tez/sample1.q.out
+++ b/ql/src/test/results/clientpositive/tez/sample1.q.out
@@ -54,15 +54,20 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
+ COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
bucket_count -1
columns key,value,dt,hr
columns.comments
columns.types int:string:string:string
#### A masked pattern was here ####
name default.dest1
+ numFiles 0
+ numRows 0
+ rawDataSize 0
serialization.ddl struct dest1 { i32 key, string value, string dt, string hr}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ totalSize 0
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.dest1
@@ -133,15 +138,20 @@ STAGE PLANS:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
properties:
+ COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
bucket_count -1
columns key,value,dt,hr
columns.comments
columns.types int:string:string:string
#### A masked pattern was here ####
name default.dest1
+ numFiles 0
+ numRows 0
+ rawDataSize 0
serialization.ddl struct dest1 { i32 key, string value, string dt, string hr}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ totalSize 0
#### A masked pattern was here ####
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.dest1