You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hive.apache.org by ha...@apache.org on 2017/12/12 00:00:48 UTC
[50/51] [partial] hive git commit: HIVE-13567 : Enable auto-gather
column stats by default (Zoltan Haindrich,
Pengcheng Xiong via Ashutosh Chauhan)
http://git-wip-us.apache.org/repos/asf/hive/blob/3bbc24d2/itests/hive-blobstore/src/test/results/clientpositive/write_final_output_blobstore.q.out
----------------------------------------------------------------------
diff --git a/itests/hive-blobstore/src/test/results/clientpositive/write_final_output_blobstore.q.out b/itests/hive-blobstore/src/test/results/clientpositive/write_final_output_blobstore.q.out
index 3d53292..a1be085 100644
--- a/itests/hive-blobstore/src/test/results/clientpositive/write_final_output_blobstore.q.out
+++ b/itests/hive-blobstore/src/test/results/clientpositive/write_final_output_blobstore.q.out
@@ -214,6 +214,41 @@ STAGE PLANS:
TotalFiles: 1
GatherStats: true
MultiFileSpray: false
+ Select Operator
+ expressions: _col0 (type: int)
+ outputColumnNames: key
+ Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL Column stats: NONE
+ Group By Operator
+ aggregations: compute_stats(key, 'hll')
+ mode: complete
+ outputColumnNames: _col0
+ Statistics: Num rows: 1 Data size: 424 Basic stats: PARTIAL Column stats: NONE
+ Select Operator
+ expressions: _col0 (type: struct<columntype:string,min:bigint,max:bigint,countnulls:bigint,numdistinctvalues:bigint,ndvbitvector:binary>)
+ outputColumnNames: _col0
+ Statistics: Num rows: 1 Data size: 424 Basic stats: PARTIAL Column stats: NONE
+ File Output Operator
+ compressed: false
+ GlobalTableId: 0
+#### A masked pattern was here ####
+ NumFilesPerFileSink: 1
+ Statistics: Num rows: 1 Data size: 424 Basic stats: PARTIAL Column stats: NONE
+#### A masked pattern was here ####
+ table:
+ input format: org.apache.hadoop.mapred.SequenceFileInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
+ properties:
+ columns _col0
+ columns.types struct<columntype:string,min:bigint,max:bigint,countnulls:bigint,numdistinctvalues:bigint,ndvbitvector:binary>
+ escape.delim \
+ hive.serialization.extend.additional.nesting.levels true
+ serialization.escape.crlf true
+ serialization.format 1
+ serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ TotalFiles: 1
+ GatherStats: false
+ MultiFileSpray: false
Stage: Stage-0
Move Operator
@@ -248,6 +283,11 @@ STAGE PLANS:
Stats Work
Basic Stats Work:
Stats Aggregation Key Prefix: ### BLOBSTORE_STAGING_PATH ###
+ Column Stats Desc:
+ Columns: key
+ Column Types: int
+ Table: default.blobstore_table
+ Is Table Level Stats: true
PREHOOK: query: EXPLAIN EXTENDED FROM hdfs_table INSERT OVERWRITE TABLE blobstore_table SELECT hdfs_table.key GROUP BY hdfs_table.key ORDER BY hdfs_table.key
PREHOOK: type: QUERY
@@ -439,6 +479,41 @@ STAGE PLANS:
TotalFiles: 1
GatherStats: true
MultiFileSpray: false
+ Select Operator
+ expressions: _col0 (type: int)
+ outputColumnNames: key
+ Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL Column stats: NONE
+ Group By Operator
+ aggregations: compute_stats(key, 'hll')
+ mode: complete
+ outputColumnNames: _col0
+ Statistics: Num rows: 1 Data size: 424 Basic stats: PARTIAL Column stats: NONE
+ Select Operator
+ expressions: _col0 (type: struct<columntype:string,min:bigint,max:bigint,countnulls:bigint,numdistinctvalues:bigint,ndvbitvector:binary>)
+ outputColumnNames: _col0
+ Statistics: Num rows: 1 Data size: 424 Basic stats: PARTIAL Column stats: NONE
+ File Output Operator
+ compressed: false
+ GlobalTableId: 0
+#### A masked pattern was here ####
+ NumFilesPerFileSink: 1
+ Statistics: Num rows: 1 Data size: 424 Basic stats: PARTIAL Column stats: NONE
+#### A masked pattern was here ####
+ table:
+ input format: org.apache.hadoop.mapred.SequenceFileInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
+ properties:
+ columns _col0
+ columns.types struct<columntype:string,min:bigint,max:bigint,countnulls:bigint,numdistinctvalues:bigint,ndvbitvector:binary>
+ escape.delim \
+ hive.serialization.extend.additional.nesting.levels true
+ serialization.escape.crlf true
+ serialization.format 1
+ serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+ TotalFiles: 1
+ GatherStats: false
+ MultiFileSpray: false
Stage: Stage-0
Move Operator
@@ -473,6 +548,11 @@ STAGE PLANS:
Stats Work
Basic Stats Work:
Stats Aggregation Key Prefix: ### BLOBSTORE_STAGING_PATH ###
+ Column Stats Desc:
+ Columns: key
+ Column Types: int
+ Table: default.blobstore_table
+ Is Table Level Stats: true
PREHOOK: query: DROP TABLE hdfs_table
PREHOOK: type: DROPTABLE
http://git-wip-us.apache.org/repos/asf/hive/blob/3bbc24d2/itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/TestAcidOnTez.java
----------------------------------------------------------------------
diff --git a/itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/TestAcidOnTez.java b/itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/TestAcidOnTez.java
index 26377e0..1fd84ac 100644
--- a/itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/TestAcidOnTez.java
+++ b/itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/TestAcidOnTez.java
@@ -675,6 +675,7 @@ ekoifman:apache-hive-3.0.0-SNAPSHOT-bin ekoifman$ tree ~/dev/hiverwgit/itests/h
@Test
public void testBucketedAcidInsertWithRemoveUnion() throws Exception {
HiveConf confForTez = new HiveConf(hiveConf); // make a clone of existing hive conf
+ confForTez.setBoolean("hive.stats.column.autogather", false);
setupTez(confForTez);
int[][] values = {{1,2},{2,4},{5,6},{6,8},{9,10}};
runStatementOnDriver("delete from " + Table.ACIDTBL, confForTez);
http://git-wip-us.apache.org/repos/asf/hive/blob/3bbc24d2/ql/src/test/queries/clientpositive/alterColumnStatsPart.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/alterColumnStatsPart.q b/ql/src/test/queries/clientpositive/alterColumnStatsPart.q
index f46d9d5..9bfc01f 100644
--- a/ql/src/test/queries/clientpositive/alterColumnStatsPart.q
+++ b/ql/src/test/queries/clientpositive/alterColumnStatsPart.q
@@ -1,3 +1,4 @@
+set hive.stats.column.autogather=false;
set hive.mapred.mode=nonstrict;
set hive.exec.dynamic.partition.mode=nonstrict;
http://git-wip-us.apache.org/repos/asf/hive/blob/3bbc24d2/ql/src/test/queries/clientpositive/auto_sortmerge_join_13.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/auto_sortmerge_join_13.q b/ql/src/test/queries/clientpositive/auto_sortmerge_join_13.q
index 1c868dc..b5706be 100644
--- a/ql/src/test/queries/clientpositive/auto_sortmerge_join_13.q
+++ b/ql/src/test/queries/clientpositive/auto_sortmerge_join_13.q
@@ -1,5 +1,4 @@
set hive.explain.user=false;
-;
set hive.exec.reducers.max = 1;
http://git-wip-us.apache.org/repos/asf/hive/blob/3bbc24d2/ql/src/test/queries/clientpositive/bucket_map_join_tez2.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/bucket_map_join_tez2.q b/ql/src/test/queries/clientpositive/bucket_map_join_tez2.q
index b79d89d..2ce859f 100644
--- a/ql/src/test/queries/clientpositive/bucket_map_join_tez2.q
+++ b/ql/src/test/queries/clientpositive/bucket_map_join_tez2.q
@@ -1,3 +1,4 @@
+set hive.stats.column.autogather=false;
set hive.strict.checks.bucketing=false;
set hive.mapred.mode=nonstrict;
http://git-wip-us.apache.org/repos/asf/hive/blob/3bbc24d2/ql/src/test/queries/clientpositive/bucket_num_reducers.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/bucket_num_reducers.q b/ql/src/test/queries/clientpositive/bucket_num_reducers.q
index 06f334e..5c5008e 100644
--- a/ql/src/test/queries/clientpositive/bucket_num_reducers.q
+++ b/ql/src/test/queries/clientpositive/bucket_num_reducers.q
@@ -1,4 +1,4 @@
-;
+set hive.stats.column.autogather=false;
set hive.exec.mode.local.auto=false;
set mapred.reduce.tasks = 10;
http://git-wip-us.apache.org/repos/asf/hive/blob/3bbc24d2/ql/src/test/queries/clientpositive/bucket_num_reducers_acid2.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/bucket_num_reducers_acid2.q b/ql/src/test/queries/clientpositive/bucket_num_reducers_acid2.q
index 2e6aa61..9776785 100644
--- a/ql/src/test/queries/clientpositive/bucket_num_reducers_acid2.q
+++ b/ql/src/test/queries/clientpositive/bucket_num_reducers_acid2.q
@@ -1,3 +1,4 @@
+set hive.stats.column.autogather=false;
set hive.support.concurrency=true;
set hive.txn.manager=org.apache.hadoop.hive.ql.lockmgr.DbTxnManager;
set hive.exec.mode.local.auto=false;
http://git-wip-us.apache.org/repos/asf/hive/blob/3bbc24d2/ql/src/test/queries/clientpositive/columnStatsUpdateForStatsOptimizer_1.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/columnStatsUpdateForStatsOptimizer_1.q b/ql/src/test/queries/clientpositive/columnStatsUpdateForStatsOptimizer_1.q
index 9cd9150..a11e7f0 100644
--- a/ql/src/test/queries/clientpositive/columnStatsUpdateForStatsOptimizer_1.q
+++ b/ql/src/test/queries/clientpositive/columnStatsUpdateForStatsOptimizer_1.q
@@ -1,3 +1,4 @@
+set hive.stats.column.autogather=false;
set hive.stats.fetch.column.stats=true;
set hive.compute.query.using.stats=true;
set hive.mapred.mode=nonstrict;
http://git-wip-us.apache.org/repos/asf/hive/blob/3bbc24d2/ql/src/test/queries/clientpositive/columnStatsUpdateForStatsOptimizer_2.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/columnStatsUpdateForStatsOptimizer_2.q b/ql/src/test/queries/clientpositive/columnStatsUpdateForStatsOptimizer_2.q
index 5fa7aec..2ae85e2 100644
--- a/ql/src/test/queries/clientpositive/columnStatsUpdateForStatsOptimizer_2.q
+++ b/ql/src/test/queries/clientpositive/columnStatsUpdateForStatsOptimizer_2.q
@@ -1,3 +1,4 @@
+set hive.stats.column.autogather=false;
set hive.stats.fetch.column.stats=true;
set hive.compute.query.using.stats=true;
http://git-wip-us.apache.org/repos/asf/hive/blob/3bbc24d2/ql/src/test/queries/clientpositive/combine1.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/combine1.q b/ql/src/test/queries/clientpositive/combine1.q
index 3bcb8b1..b300830 100644
--- a/ql/src/test/queries/clientpositive/combine1.q
+++ b/ql/src/test/queries/clientpositive/combine1.q
@@ -7,6 +7,8 @@ set mapred.max.split.size=256;
set mapred.output.compression.codec=org.apache.hadoop.io.compress.GzipCodec;
+set hive.stats.column.autogather=false;
+
-- SORT_QUERY_RESULTS
create table combine1_1(key string, value string) stored as textfile;
http://git-wip-us.apache.org/repos/asf/hive/blob/3bbc24d2/ql/src/test/queries/clientpositive/correlationoptimizer5.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/correlationoptimizer5.q b/ql/src/test/queries/clientpositive/correlationoptimizer5.q
index 45b8cb9..002fb12 100644
--- a/ql/src/test/queries/clientpositive/correlationoptimizer5.q
+++ b/ql/src/test/queries/clientpositive/correlationoptimizer5.q
@@ -1,3 +1,5 @@
+set hive.stats.column.autogather=false;
+-- Currently, a query with multiple FileSinkOperators are not supported.
set hive.mapred.mode=nonstrict;
CREATE TABLE T1(key INT, val STRING);
LOAD DATA LOCAL INPATH '../../data/files/kv1.txt' INTO TABLE T1;
http://git-wip-us.apache.org/repos/asf/hive/blob/3bbc24d2/ql/src/test/queries/clientpositive/dp_counter_mm.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/dp_counter_mm.q b/ql/src/test/queries/clientpositive/dp_counter_mm.q
index 4f3b100..5a7b859 100644
--- a/ql/src/test/queries/clientpositive/dp_counter_mm.q
+++ b/ql/src/test/queries/clientpositive/dp_counter_mm.q
@@ -1,3 +1,5 @@
+-- remove disable after HIVE-18237
+set hive.stats.column.autogather=false;
set hive.exec.dynamic.partition.mode=nonstrict;
set hive.exec.max.dynamic.partitions.pernode=200;
set hive.exec.max.dynamic.partitions=200;
http://git-wip-us.apache.org/repos/asf/hive/blob/3bbc24d2/ql/src/test/queries/clientpositive/encryption_insert_values.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/encryption_insert_values.q b/ql/src/test/queries/clientpositive/encryption_insert_values.q
index 2dd3e9a..c8d1d51 100644
--- a/ql/src/test/queries/clientpositive/encryption_insert_values.q
+++ b/ql/src/test/queries/clientpositive/encryption_insert_values.q
@@ -1,4 +1,5 @@
-- SORT_QUERY_RESULTS;
+set hive.stats.column.autogather=false;
DROP TABLE IF EXISTS encrypted_table PURGE;
CREATE TABLE encrypted_table (key INT, value STRING) LOCATION '${hiveconf:hive.metastore.warehouse.dir}/default/encrypted_table';
@@ -12,4 +13,4 @@ select * from encrypted_table;
-- this checks that we've actually created temp table data under encrypted_table folder
describe formatted values__tmp__table__1;
-CRYPTO DELETE_KEY --keyName key_128;
\ No newline at end of file
+CRYPTO DELETE_KEY --keyName key_128;
http://git-wip-us.apache.org/repos/asf/hive/blob/3bbc24d2/ql/src/test/queries/clientpositive/encryption_join_with_different_encryption_keys.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/encryption_join_with_different_encryption_keys.q b/ql/src/test/queries/clientpositive/encryption_join_with_different_encryption_keys.q
index 4dcea1f..7159ad5 100644
--- a/ql/src/test/queries/clientpositive/encryption_join_with_different_encryption_keys.q
+++ b/ql/src/test/queries/clientpositive/encryption_join_with_different_encryption_keys.q
@@ -1,3 +1,4 @@
+set hive.stats.column.autogather=false;
--SORT_QUERY_RESULTS
-- Java JCE must be installed in order to hava a key length of 256 bits
http://git-wip-us.apache.org/repos/asf/hive/blob/3bbc24d2/ql/src/test/queries/clientpositive/encryption_move_tbl.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/encryption_move_tbl.q b/ql/src/test/queries/clientpositive/encryption_move_tbl.q
index 0b7771c..8d865aa 100644
--- a/ql/src/test/queries/clientpositive/encryption_move_tbl.q
+++ b/ql/src/test/queries/clientpositive/encryption_move_tbl.q
@@ -1,4 +1,5 @@
-- SORT_QUERY_RESULTS;
+set hive.stats.column.autogather=false;
-- we're setting this so that TestNegaiveCliDriver.vm doesn't stop processing after ALTER TABLE fails;
http://git-wip-us.apache.org/repos/asf/hive/blob/3bbc24d2/ql/src/test/queries/clientpositive/groupby1.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/groupby1.q b/ql/src/test/queries/clientpositive/groupby1.q
index a8c9a8d..cd3a12b 100755
--- a/ql/src/test/queries/clientpositive/groupby1.q
+++ b/ql/src/test/queries/clientpositive/groupby1.q
@@ -1,3 +1,5 @@
+-- due to testMTQueries1
+set hive.stats.column.autogather=false;
set hive.mapred.mode=nonstrict;
set hive.explain.user=false;
set hive.map.aggr=false;
http://git-wip-us.apache.org/repos/asf/hive/blob/3bbc24d2/ql/src/test/queries/clientpositive/groupby1_limit.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/groupby1_limit.q b/ql/src/test/queries/clientpositive/groupby1_limit.q
index b8e389e..6c40e19 100644
--- a/ql/src/test/queries/clientpositive/groupby1_limit.q
+++ b/ql/src/test/queries/clientpositive/groupby1_limit.q
@@ -1,3 +1,4 @@
+set hive.stats.column.autogather=false;
set hive.mapred.mode=nonstrict;
set mapred.reduce.tasks=31;
http://git-wip-us.apache.org/repos/asf/hive/blob/3bbc24d2/ql/src/test/queries/clientpositive/groupby_multi_single_reducer.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/groupby_multi_single_reducer.q b/ql/src/test/queries/clientpositive/groupby_multi_single_reducer.q
index 2b799f8..40976ee 100644
--- a/ql/src/test/queries/clientpositive/groupby_multi_single_reducer.q
+++ b/ql/src/test/queries/clientpositive/groupby_multi_single_reducer.q
@@ -1,3 +1,6 @@
+set hive.stats.column.autogather=false;
+-- due to L137 in LimitPushDownOptimization Not safe to continue for RS-GBY-GBY-LIM kind of pipelines. See HIVE-10607 for more.
+
set hive.multigroupby.singlereducer=true;
-- SORT_QUERY_RESULTS
http://git-wip-us.apache.org/repos/asf/hive/blob/3bbc24d2/ql/src/test/queries/clientpositive/hybridgrace_hashjoin_1.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/hybridgrace_hashjoin_1.q b/ql/src/test/queries/clientpositive/hybridgrace_hashjoin_1.q
index e404dd0..8955b6a 100644
--- a/ql/src/test/queries/clientpositive/hybridgrace_hashjoin_1.q
+++ b/ql/src/test/queries/clientpositive/hybridgrace_hashjoin_1.q
@@ -1,3 +1,4 @@
+set hive.stats.column.autogather=false;
set hive.mapred.mode=nonstrict;
set hive.explain.user=false;
set tez.cartesian-product.max-parallelism=1;
http://git-wip-us.apache.org/repos/asf/hive/blob/3bbc24d2/ql/src/test/queries/clientpositive/infer_bucket_sort_convert_join.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/infer_bucket_sort_convert_join.q b/ql/src/test/queries/clientpositive/infer_bucket_sort_convert_join.q
index 6809b72..e417028 100644
--- a/ql/src/test/queries/clientpositive/infer_bucket_sort_convert_join.q
+++ b/ql/src/test/queries/clientpositive/infer_bucket_sort_convert_join.q
@@ -1,3 +1,12 @@
+set hive.stats.column.autogather=false;
+-- sounds weird:
+-- on master, when auto=true, hive.mapjoin.localtask.max.memory.usage will be 0.55 as there is a gby
+-- L132 of LocalMapJoinProcFactory
+-- when execute in CLI, hive.exec.submit.local.task.via.child is true and we can see the error
+-- if set hive.exec.submit.local.task.via.child=false, we can see it.
+-- with patch, we just merge the tasks. hive.exec.submit.local.task.via.child=false due to pom.xml setting
+-- however, even after change it to true, it still fails.
+
set hive.mapred.mode=nonstrict;
set hive.exec.infer.bucket.sort=true;
set hive.exec.infer.bucket.sort.num.buckets.power.two=true;
http://git-wip-us.apache.org/repos/asf/hive/blob/3bbc24d2/ql/src/test/queries/clientpositive/infer_bucket_sort_reducers_power_two.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/infer_bucket_sort_reducers_power_two.q b/ql/src/test/queries/clientpositive/infer_bucket_sort_reducers_power_two.q
index 6824c1c..c0ddb8b 100644
--- a/ql/src/test/queries/clientpositive/infer_bucket_sort_reducers_power_two.q
+++ b/ql/src/test/queries/clientpositive/infer_bucket_sort_reducers_power_two.q
@@ -1,3 +1,4 @@
+set hive.stats.column.autogather=false;
set hive.mapred.mode=nonstrict;
set hive.exec.infer.bucket.sort=true;
set hive.exec.infer.bucket.sort.num.buckets.power.two=true;
http://git-wip-us.apache.org/repos/asf/hive/blob/3bbc24d2/ql/src/test/queries/clientpositive/input11_limit.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/input11_limit.q b/ql/src/test/queries/clientpositive/input11_limit.q
index 052a72e..211c37a 100644
--- a/ql/src/test/queries/clientpositive/input11_limit.q
+++ b/ql/src/test/queries/clientpositive/input11_limit.q
@@ -1,3 +1,4 @@
+set hive.stats.column.autogather=false;
set hive.mapred.mode=nonstrict;
-- SORT_QUERY_RESULTS
http://git-wip-us.apache.org/repos/asf/hive/blob/3bbc24d2/ql/src/test/queries/clientpositive/input14_limit.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/input14_limit.q b/ql/src/test/queries/clientpositive/input14_limit.q
index 7316752..2f6e4e4 100644
--- a/ql/src/test/queries/clientpositive/input14_limit.q
+++ b/ql/src/test/queries/clientpositive/input14_limit.q
@@ -1,3 +1,5 @@
+set hive.stats.column.autogather=false;
+
CREATE TABLE dest1(key INT, value STRING) STORED AS TEXTFILE;
EXPLAIN
http://git-wip-us.apache.org/repos/asf/hive/blob/3bbc24d2/ql/src/test/queries/clientpositive/join2.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/join2.q b/ql/src/test/queries/clientpositive/join2.q
index 8aedd56..c3c7c24 100644
--- a/ql/src/test/queries/clientpositive/join2.q
+++ b/ql/src/test/queries/clientpositive/join2.q
@@ -1,3 +1,5 @@
+-- due to testMTQueries1
+set hive.stats.column.autogather=false;
set hive.mapred.mode=nonstrict;
-- SORT_QUERY_RESULTS
http://git-wip-us.apache.org/repos/asf/hive/blob/3bbc24d2/ql/src/test/queries/clientpositive/metadata_only_queries.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/metadata_only_queries.q b/ql/src/test/queries/clientpositive/metadata_only_queries.q
index 8581a46..bcf320b 100644
--- a/ql/src/test/queries/clientpositive/metadata_only_queries.q
+++ b/ql/src/test/queries/clientpositive/metadata_only_queries.q
@@ -1,3 +1,4 @@
+set hive.stats.column.autogather=false;
set hive.mapred.mode=nonstrict;
set hive.explain.user=false;
set hive.compute.query.using.stats=true;
http://git-wip-us.apache.org/repos/asf/hive/blob/3bbc24d2/ql/src/test/queries/clientpositive/metadata_only_queries_with_filters.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/metadata_only_queries_with_filters.q b/ql/src/test/queries/clientpositive/metadata_only_queries_with_filters.q
index 1af813e..692c414 100644
--- a/ql/src/test/queries/clientpositive/metadata_only_queries_with_filters.q
+++ b/ql/src/test/queries/clientpositive/metadata_only_queries_with_filters.q
@@ -1,3 +1,4 @@
+set hive.stats.column.autogather=false;
set hive.stats.dbclass=fs;
set hive.compute.query.using.stats=true;
set hive.explain.user=false;
http://git-wip-us.apache.org/repos/asf/hive/blob/3bbc24d2/ql/src/test/queries/clientpositive/multiMapJoin1.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/multiMapJoin1.q b/ql/src/test/queries/clientpositive/multiMapJoin1.q
index 5c49b4c..6e16af4 100644
--- a/ql/src/test/queries/clientpositive/multiMapJoin1.q
+++ b/ql/src/test/queries/clientpositive/multiMapJoin1.q
@@ -1,3 +1,4 @@
+set hive.stats.column.autogather=false;
set hive.mapred.mode=nonstrict;
set hive.exec.post.hooks=org.apache.hadoop.hive.ql.hooks.PostExecutePrinter,org.apache.hadoop.hive.ql.hooks.PrintCompletedTasksHook;
http://git-wip-us.apache.org/repos/asf/hive/blob/3bbc24d2/ql/src/test/queries/clientpositive/orc_wide_table.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/orc_wide_table.q b/ql/src/test/queries/clientpositive/orc_wide_table.q
index 422a3c2..d2ec385 100644
--- a/ql/src/test/queries/clientpositive/orc_wide_table.q
+++ b/ql/src/test/queries/clientpositive/orc_wide_table.q
@@ -1,4 +1,5 @@
set hive.mapred.mode=nonstrict;
+set hive.stats.column.autogather=false;
drop table if exists test_txt;
drop table if exists test_orc;
create table test_txt(
http://git-wip-us.apache.org/repos/asf/hive/blob/3bbc24d2/ql/src/test/queries/clientpositive/partition_coltype_literals.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/partition_coltype_literals.q b/ql/src/test/queries/clientpositive/partition_coltype_literals.q
index eb56b1a..8da4876 100644
--- a/ql/src/test/queries/clientpositive/partition_coltype_literals.q
+++ b/ql/src/test/queries/clientpositive/partition_coltype_literals.q
@@ -1,3 +1,4 @@
+set hive.stats.column.autogather=false;
set hive.compute.query.using.stats=false;
drop table if exists partcoltypenum;
create table partcoltypenum (key int, value string) partitioned by (tint tinyint, sint smallint, bint bigint);
http://git-wip-us.apache.org/repos/asf/hive/blob/3bbc24d2/ql/src/test/queries/clientpositive/row__id.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/row__id.q b/ql/src/test/queries/clientpositive/row__id.q
index d9cb7b0..6aaa40f 100644
--- a/ql/src/test/queries/clientpositive/row__id.q
+++ b/ql/src/test/queries/clientpositive/row__id.q
@@ -1,3 +1,5 @@
+-- tid is flaky when compute column stats
+set hive.stats.column.autogather=false;
set hive.txn.manager=org.apache.hadoop.hive.ql.lockmgr.DbTxnManager;
drop table if exists hello_acid;
http://git-wip-us.apache.org/repos/asf/hive/blob/3bbc24d2/ql/src/test/queries/clientpositive/smb_join_partition_key.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/smb_join_partition_key.q b/ql/src/test/queries/clientpositive/smb_join_partition_key.q
index 160bf5e..23027f8 100644
--- a/ql/src/test/queries/clientpositive/smb_join_partition_key.q
+++ b/ql/src/test/queries/clientpositive/smb_join_partition_key.q
@@ -1,3 +1,5 @@
+--because p1 is decimal, in derby, when it retrieves partition with decimal, it will use partval = 100.0, rather than 100. As a result, the partition will not be found and it throws exception.
+set hive.stats.column.autogather=false;
set hive.mapred.mode=nonstrict;
SET hive.enforce.sortmergebucketmapjoin=false;
SET hive.auto.convert.sortmerge.join=true;
http://git-wip-us.apache.org/repos/asf/hive/blob/3bbc24d2/ql/src/test/queries/clientpositive/stats_only_null.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/stats_only_null.q b/ql/src/test/queries/clientpositive/stats_only_null.q
index 7fd22c5..fa83c6f 100644
--- a/ql/src/test/queries/clientpositive/stats_only_null.q
+++ b/ql/src/test/queries/clientpositive/stats_only_null.q
@@ -1,3 +1,4 @@
+set hive.stats.column.autogather=false;
set hive.mapred.mode=nonstrict;
set hive.explain.user=false;
set hive.stats.dbclass=fs;
http://git-wip-us.apache.org/repos/asf/hive/blob/3bbc24d2/ql/src/test/queries/clientpositive/tez_smb_1.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/tez_smb_1.q b/ql/src/test/queries/clientpositive/tez_smb_1.q
index 4e8d1c0..1ae35ef 100644
--- a/ql/src/test/queries/clientpositive/tez_smb_1.q
+++ b/ql/src/test/queries/clientpositive/tez_smb_1.q
@@ -1,3 +1,4 @@
+set hive.stats.column.autogather=false;
set hive.strict.checks.bucketing=false;
set hive.mapred.mode=nonstrict;
http://git-wip-us.apache.org/repos/asf/hive/blob/3bbc24d2/ql/src/test/queries/clientpositive/tez_smb_main.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/tez_smb_main.q b/ql/src/test/queries/clientpositive/tez_smb_main.q
index e4ab75a..6bace67 100644
--- a/ql/src/test/queries/clientpositive/tez_smb_main.q
+++ b/ql/src/test/queries/clientpositive/tez_smb_main.q
@@ -1,3 +1,4 @@
+set hive.stats.column.autogather=false;
set hive.strict.checks.bucketing=false;
set hive.mapred.mode=nonstrict;
http://git-wip-us.apache.org/repos/asf/hive/blob/3bbc24d2/ql/src/test/queries/clientpositive/udf_round_2.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/udf_round_2.q b/ql/src/test/queries/clientpositive/udf_round_2.q
index 43988c1..38885a9 100644
--- a/ql/src/test/queries/clientpositive/udf_round_2.q
+++ b/ql/src/test/queries/clientpositive/udf_round_2.q
@@ -1,4 +1,5 @@
set hive.fetch.task.conversion=more;
+set hive.stats.column.autogather=false;
-- test for NaN (not-a-number)
create table tstTbl1(n double);
http://git-wip-us.apache.org/repos/asf/hive/blob/3bbc24d2/ql/src/test/queries/clientpositive/union_remove_26.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/union_remove_26.q b/ql/src/test/queries/clientpositive/union_remove_26.q
index 4e77d73..92ccbc3 100644
--- a/ql/src/test/queries/clientpositive/union_remove_26.q
+++ b/ql/src/test/queries/clientpositive/union_remove_26.q
@@ -1,3 +1,4 @@
+set hive.stats.column.autogather=false;
set hive.stats.autogather=true;
-- This is to test the union remove optimization with stats optimization
http://git-wip-us.apache.org/repos/asf/hive/blob/3bbc24d2/ql/src/test/queries/clientpositive/vector_auto_smb_mapjoin_14.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/vector_auto_smb_mapjoin_14.q b/ql/src/test/queries/clientpositive/vector_auto_smb_mapjoin_14.q
index 3c53853..81d6541 100644
--- a/ql/src/test/queries/clientpositive/vector_auto_smb_mapjoin_14.q
+++ b/ql/src/test/queries/clientpositive/vector_auto_smb_mapjoin_14.q
@@ -1,3 +1,4 @@
+set hive.stats.column.autogather=false;
set hive.mapred.mode=nonstrict;
set hive.explain.user=false;
SET hive.vectorized.execution.enabled=true;
http://git-wip-us.apache.org/repos/asf/hive/blob/3bbc24d2/ql/src/test/queries/clientpositive/vector_bucket.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/vector_bucket.q b/ql/src/test/queries/clientpositive/vector_bucket.q
index b67592e..25d3b56 100644
--- a/ql/src/test/queries/clientpositive/vector_bucket.q
+++ b/ql/src/test/queries/clientpositive/vector_bucket.q
@@ -1,3 +1,4 @@
+set hive.stats.column.autogather=false;
set hive.mapred.mode=nonstrict;
set hive.explain.user=false;
SET hive.vectorized.execution.enabled=true;
http://git-wip-us.apache.org/repos/asf/hive/blob/3bbc24d2/ql/src/test/queries/clientpositive/vector_char_4.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/vector_char_4.q b/ql/src/test/queries/clientpositive/vector_char_4.q
index edc59cd..df72ac1 100644
--- a/ql/src/test/queries/clientpositive/vector_char_4.q
+++ b/ql/src/test/queries/clientpositive/vector_char_4.q
@@ -1,3 +1,4 @@
+set hive.stats.column.autogather=false;
set hive.explain.user=false;
SET hive.vectorized.execution.enabled=true;
set hive.fetch.task.conversion=none;
@@ -50,4 +51,4 @@ insert overwrite table char_lazy_binary_columnar select t, si, i, b, f, d, s fro
-- insert overwrite table char_lazy_binary_columnar select t, si, i, b, f, d, s from vectortab2korc;
--- select count(*) as cnt from char_lazy_binary_columnar group by cs order by cnt asc;
\ No newline at end of file
+-- select count(*) as cnt from char_lazy_binary_columnar group by cs order by cnt asc;
http://git-wip-us.apache.org/repos/asf/hive/blob/3bbc24d2/ql/src/test/queries/clientpositive/vector_char_simple.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/vector_char_simple.q b/ql/src/test/queries/clientpositive/vector_char_simple.q
index c315241..70b0900 100644
--- a/ql/src/test/queries/clientpositive/vector_char_simple.q
+++ b/ql/src/test/queries/clientpositive/vector_char_simple.q
@@ -1,3 +1,4 @@
+set hive.stats.column.autogather=false;
set hive.explain.user=false;
SET hive.vectorized.execution.enabled=true;
set hive.fetch.task.conversion=none;
http://git-wip-us.apache.org/repos/asf/hive/blob/3bbc24d2/ql/src/test/queries/clientpositive/vector_decimal_mapjoin.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/vector_decimal_mapjoin.q b/ql/src/test/queries/clientpositive/vector_decimal_mapjoin.q
index 940cc12..ed27ce9 100644
--- a/ql/src/test/queries/clientpositive/vector_decimal_mapjoin.q
+++ b/ql/src/test/queries/clientpositive/vector_decimal_mapjoin.q
@@ -1,3 +1,5 @@
+-- until HIVE-18225 is fixed
+set hive.stats.column.autogather=false;
set hive.mapred.mode=nonstrict;
set hive.explain.user=false;
SET hive.auto.convert.join=true;
http://git-wip-us.apache.org/repos/asf/hive/blob/3bbc24d2/ql/src/test/queries/clientpositive/vector_groupby_rollup1.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/vector_groupby_rollup1.q b/ql/src/test/queries/clientpositive/vector_groupby_rollup1.q
index 17858ff..39bc2c1 100644
--- a/ql/src/test/queries/clientpositive/vector_groupby_rollup1.q
+++ b/ql/src/test/queries/clientpositive/vector_groupby_rollup1.q
@@ -1,3 +1,4 @@
+set hive.stats.column.autogather=false;
set hive.explain.user=false;
SET hive.vectorized.execution.enabled=true;
SET hive.vectorized.execution.reduce.enabled=true;
http://git-wip-us.apache.org/repos/asf/hive/blob/3bbc24d2/ql/src/test/queries/clientpositive/vector_multi_insert.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/vector_multi_insert.q b/ql/src/test/queries/clientpositive/vector_multi_insert.q
index c56ee1c..e6bfb96 100644
--- a/ql/src/test/queries/clientpositive/vector_multi_insert.q
+++ b/ql/src/test/queries/clientpositive/vector_multi_insert.q
@@ -1,3 +1,4 @@
+set hive.stats.column.autogather=false;
set hive.explain.user=false;
SET hive.vectorized.execution.enabled=true;
set hive.fetch.task.conversion=none;
http://git-wip-us.apache.org/repos/asf/hive/blob/3bbc24d2/ql/src/test/queries/clientpositive/vector_tablesample_rows.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/vector_tablesample_rows.q b/ql/src/test/queries/clientpositive/vector_tablesample_rows.q
index bb3c5a4..e748e90 100644
--- a/ql/src/test/queries/clientpositive/vector_tablesample_rows.q
+++ b/ql/src/test/queries/clientpositive/vector_tablesample_rows.q
@@ -1,3 +1,4 @@
+set hive.stats.column.autogather=false;
set hive.cli.print.header=true;
set hive.explain.user=false;
SET hive.vectorized.execution.enabled=true;
@@ -35,4 +36,4 @@ create temporary table dual as select 1;
create temporary table dual as select 1;
-select * from dual;
\ No newline at end of file
+select * from dual;
http://git-wip-us.apache.org/repos/asf/hive/blob/3bbc24d2/ql/src/test/queries/clientpositive/vector_udf_character_length.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/vector_udf_character_length.q b/ql/src/test/queries/clientpositive/vector_udf_character_length.q
index 19a5260..e49a091 100644
--- a/ql/src/test/queries/clientpositive/vector_udf_character_length.q
+++ b/ql/src/test/queries/clientpositive/vector_udf_character_length.q
@@ -1,3 +1,4 @@
+set hive.stats.column.autogather=false;
SET hive.vectorized.execution.enabled=true;
set hive.fetch.task.conversion=none;
http://git-wip-us.apache.org/repos/asf/hive/blob/3bbc24d2/ql/src/test/queries/clientpositive/vector_udf_octet_length.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/vector_udf_octet_length.q b/ql/src/test/queries/clientpositive/vector_udf_octet_length.q
index 06a4985..af4c7c4 100644
--- a/ql/src/test/queries/clientpositive/vector_udf_octet_length.q
+++ b/ql/src/test/queries/clientpositive/vector_udf_octet_length.q
@@ -1,3 +1,4 @@
+set hive.stats.column.autogather=false;
SET hive.vectorized.execution.enabled=true;
set hive.fetch.task.conversion=none;
http://git-wip-us.apache.org/repos/asf/hive/blob/3bbc24d2/ql/src/test/queries/clientpositive/vector_varchar_4.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/vector_varchar_4.q b/ql/src/test/queries/clientpositive/vector_varchar_4.q
index 80f84d8..b3402d0 100644
--- a/ql/src/test/queries/clientpositive/vector_varchar_4.q
+++ b/ql/src/test/queries/clientpositive/vector_varchar_4.q
@@ -1,3 +1,4 @@
+set hive.stats.column.autogather=false;
set hive.explain.user=false;
SET hive.vectorized.execution.enabled=true;
set hive.fetch.task.conversion=none;
@@ -50,4 +51,4 @@ insert overwrite table varchar_lazy_binary_columnar select t, si, i, b, f, d, s
-- insert overwrite table varchar_lazy_binary_columnar select t, si, i, b, f, d, s from vectortab2korc;
--- select count(*) as cnt from varchar_lazy_binary_columnar group by vs order by cnt asc;
\ No newline at end of file
+-- select count(*) as cnt from varchar_lazy_binary_columnar group by vs order by cnt asc;
http://git-wip-us.apache.org/repos/asf/hive/blob/3bbc24d2/ql/src/test/queries/clientpositive/vector_varchar_simple.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/vector_varchar_simple.q b/ql/src/test/queries/clientpositive/vector_varchar_simple.q
index 6f753a7..352ec3a 100644
--- a/ql/src/test/queries/clientpositive/vector_varchar_simple.q
+++ b/ql/src/test/queries/clientpositive/vector_varchar_simple.q
@@ -1,3 +1,4 @@
+set hive.stats.column.autogather=false;
set hive.explain.user=false;
SET hive.vectorized.execution.enabled=true;
set hive.fetch.task.conversion=none;
http://git-wip-us.apache.org/repos/asf/hive/blob/3bbc24d2/ql/src/test/queries/clientpositive/vector_windowing_expressions.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/vector_windowing_expressions.q b/ql/src/test/queries/clientpositive/vector_windowing_expressions.q
index 6a37c4e..3a72900 100644
--- a/ql/src/test/queries/clientpositive/vector_windowing_expressions.q
+++ b/ql/src/test/queries/clientpositive/vector_windowing_expressions.q
@@ -1,3 +1,4 @@
+set hive.stats.column.autogather=false;
set hive.cli.print.header=true;
SET hive.vectorized.execution.enabled=true;
SET hive.vectorized.execution.reduce.enabled=true;
http://git-wip-us.apache.org/repos/asf/hive/blob/3bbc24d2/ql/src/test/queries/clientpositive/vectorized_insert_into_bucketed_table.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/vectorized_insert_into_bucketed_table.q b/ql/src/test/queries/clientpositive/vectorized_insert_into_bucketed_table.q
index 8a71a70..ad87d0a 100644
--- a/ql/src/test/queries/clientpositive/vectorized_insert_into_bucketed_table.q
+++ b/ql/src/test/queries/clientpositive/vectorized_insert_into_bucketed_table.q
@@ -1,6 +1,7 @@
+set hive.stats.column.autogather=false;
set hive.vectorized.execution.enabled=true;
create temporary table foo (x int) clustered by (x) into 4 buckets stored as orc;
explain vectorization detail insert overwrite table foo values(1),(2),(3),(4),(9);
insert overwrite table foo values(1),(2),(3),(4),(9);
select *, regexp_extract(INPUT__FILE__NAME, '.*/(.*)', 1) from foo;
-drop table foo;
\ No newline at end of file
+drop table foo;
http://git-wip-us.apache.org/repos/asf/hive/blob/3bbc24d2/ql/src/test/results/clientnegative/fileformat_void_input.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientnegative/fileformat_void_input.q.out b/ql/src/test/results/clientnegative/fileformat_void_input.q.out
index 6043258..59242ea 100644
--- a/ql/src/test/results/clientnegative/fileformat_void_input.q.out
+++ b/ql/src/test/results/clientnegative/fileformat_void_input.q.out
@@ -10,16 +10,4 @@ POSTHOOK: query: CREATE TABLE dest1(key INT, value STRING) STORED AS
POSTHOOK: type: CREATETABLE
POSTHOOK: Output: database:default
POSTHOOK: Output: default@dest1
-PREHOOK: query: FROM src
-INSERT OVERWRITE TABLE dest1 SELECT src.key, src.value WHERE src.key < 10
-PREHOOK: type: QUERY
-PREHOOK: Input: default@src
-PREHOOK: Output: default@dest1
-POSTHOOK: query: FROM src
-INSERT OVERWRITE TABLE dest1 SELECT src.key, src.value WHERE src.key < 10
-POSTHOOK: type: QUERY
-POSTHOOK: Input: default@src
-POSTHOOK: Output: default@dest1
-POSTHOOK: Lineage: dest1.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
-POSTHOOK: Lineage: dest1.value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ]
-FAILED: SemanticException 3:20 Input format must implement InputFormat. Error encountered near token 'dest1'
+FAILED: SemanticException 1:72 Input format must implement InputFormat. Error encountered near token 'dest1'
http://git-wip-us.apache.org/repos/asf/hive/blob/3bbc24d2/ql/src/test/results/clientpositive/acid_table_stats.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/acid_table_stats.q.out b/ql/src/test/results/clientpositive/acid_table_stats.q.out
index 4c8297e..8a25e5a 100644
--- a/ql/src/test/results/clientpositive/acid_table_stats.q.out
+++ b/ql/src/test/results/clientpositive/acid_table_stats.q.out
@@ -92,7 +92,6 @@ Database: default
Table: acid
#### A masked pattern was here ####
Partition Parameters:
- COLUMN_STATS_ACCURATE {}
numFiles 2
numRows 0
rawDataSize 0
@@ -384,7 +383,6 @@ Database: default
Table: acid
#### A masked pattern was here ####
Partition Parameters:
- COLUMN_STATS_ACCURATE {}
numFiles 4
numRows 1000
rawDataSize 208000
@@ -663,7 +661,7 @@ Database: default
Table: acid
#### A masked pattern was here ####
Partition Parameters:
- COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\"}
+ COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\",\"COLUMN_STATS\":{\"key\":\"true\",\"value\":\"true\"}}
numFiles 2
numRows 1000
rawDataSize 176000
http://git-wip-us.apache.org/repos/asf/hive/blob/3bbc24d2/ql/src/test/results/clientpositive/alterColumnStats.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/alterColumnStats.q.out b/ql/src/test/results/clientpositive/alterColumnStats.q.out
index 09f2fe1..6136392 100644
--- a/ql/src/test/results/clientpositive/alterColumnStats.q.out
+++ b/ql/src/test/results/clientpositive/alterColumnStats.q.out
@@ -72,7 +72,7 @@ Retention: 0
#### A masked pattern was here ####
Table Type: MANAGED_TABLE
Table Parameters:
- COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\"}
+ COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\",\"COLUMN_STATS\":{\"c1\":\"true\",\"c2\":\"true\",\"insert_num\":\"true\"}}
numFiles 1
numRows 1
rawDataSize 8
@@ -115,7 +115,7 @@ Retention: 0
#### A masked pattern was here ####
Table Type: MANAGED_TABLE
Table Parameters:
- COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\"}
+ COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\",\"COLUMN_STATS\":{\"insert_num\":\"true\"}}
#### A masked pattern was here ####
numFiles 1
numRows 1
@@ -151,7 +151,7 @@ num_trues
num_falses
bitVector
comment from deserializer
-COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\"}
+COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\",\"COLUMN_STATS\":{\"insert_num\":\"true\"}}
PREHOOK: query: desc formatted p c2
PREHOOK: type: DESCTABLE
PREHOOK: Input: default@p
@@ -170,4 +170,4 @@ num_trues
num_falses
bitVector
comment from deserializer
-COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\"}
+COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\",\"COLUMN_STATS\":{\"insert_num\":\"true\"}}
http://git-wip-us.apache.org/repos/asf/hive/blob/3bbc24d2/ql/src/test/results/clientpositive/alter_numbuckets_partitioned_table2_h23.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/alter_numbuckets_partitioned_table2_h23.q.out b/ql/src/test/results/clientpositive/alter_numbuckets_partitioned_table2_h23.q.out
index 3479ac3..2d6aca5 100644
--- a/ql/src/test/results/clientpositive/alter_numbuckets_partitioned_table2_h23.q.out
+++ b/ql/src/test/results/clientpositive/alter_numbuckets_partitioned_table2_h23.q.out
@@ -75,7 +75,7 @@ Database: default
Table: tst1
#### A masked pattern was here ####
Partition Parameters:
- COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\"}
+ COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\",\"COLUMN_STATS\":{\"key\":\"true\",\"value\":\"true\"}}
numFiles 1
numRows 500
rawDataSize 5312
@@ -170,7 +170,7 @@ Database: default
Table: tst1
#### A masked pattern was here ####
Partition Parameters:
- COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\"}
+ COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\",\"COLUMN_STATS\":{\"key\":\"true\",\"value\":\"true\"}}
numFiles 8
numRows 500
rawDataSize 5312
@@ -265,7 +265,7 @@ Database: default
Table: tst1
#### A masked pattern was here ####
Partition Parameters:
- COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\"}
+ COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\",\"COLUMN_STATS\":{\"key\":\"true\",\"value\":\"true\"}}
numFiles 8
numRows 500
rawDataSize 5312
@@ -360,7 +360,7 @@ Database: default
Table: tst1
#### A masked pattern was here ####
Partition Parameters:
- COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\"}
+ COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\",\"COLUMN_STATS\":{\"key\":\"true\",\"value\":\"true\"}}
numFiles 8
numRows 500
rawDataSize 5312
@@ -455,7 +455,7 @@ Database: default
Table: tst1
#### A masked pattern was here ####
Partition Parameters:
- COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\"}
+ COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\",\"COLUMN_STATS\":{\"key\":\"true\",\"value\":\"true\"}}
numFiles 4
numRows 500
rawDataSize 5312
@@ -550,7 +550,7 @@ Database: default
Table: tst1
#### A masked pattern was here ####
Partition Parameters:
- COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\"}
+ COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\",\"COLUMN_STATS\":{\"key\":\"true\",\"value\":\"true\"}}
numFiles 4
numRows 500
rawDataSize 5312
@@ -645,7 +645,7 @@ Database: default
Table: tst1
#### A masked pattern was here ####
Partition Parameters:
- COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\"}
+ COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\",\"COLUMN_STATS\":{\"key\":\"true\",\"value\":\"true\"}}
numFiles 4
numRows 500
rawDataSize 5312
@@ -740,7 +740,7 @@ Database: default
Table: tst1
#### A masked pattern was here ####
Partition Parameters:
- COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\"}
+ COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\",\"COLUMN_STATS\":{\"key\":\"true\",\"value\":\"true\"}}
numFiles 4
numRows 500
rawDataSize 5312
@@ -835,7 +835,7 @@ Database: default
Table: tst1
#### A masked pattern was here ####
Partition Parameters:
- COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\"}
+ COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\",\"COLUMN_STATS\":{\"key\":\"true\",\"value\":\"true\"}}
numFiles 1
numRows 500
rawDataSize 5312
http://git-wip-us.apache.org/repos/asf/hive/blob/3bbc24d2/ql/src/test/results/clientpositive/alter_numbuckets_partitioned_table_h23.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/alter_numbuckets_partitioned_table_h23.q.out b/ql/src/test/results/clientpositive/alter_numbuckets_partitioned_table_h23.q.out
index 537e776..60d3eba 100644
--- a/ql/src/test/results/clientpositive/alter_numbuckets_partitioned_table_h23.q.out
+++ b/ql/src/test/results/clientpositive/alter_numbuckets_partitioned_table_h23.q.out
@@ -84,7 +84,7 @@ Database: default
Table: tst1
#### A masked pattern was here ####
Partition Parameters:
- COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\"}
+ COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\",\"COLUMN_STATS\":{\"key\":\"true\",\"value\":\"true\"}}
numFiles 8
numRows 500
rawDataSize 5312
@@ -139,7 +139,7 @@ Database: default
Table: tst1
#### A masked pattern was here ####
Partition Parameters:
- COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\"}
+ COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\",\"COLUMN_STATS\":{\"key\":\"true\",\"value\":\"true\"}}
numFiles 12
numRows 500
rawDataSize 5312
@@ -263,7 +263,7 @@ Database: default
Table: tst1
#### A masked pattern was here ####
Partition Parameters:
- COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\"}
+ COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\",\"COLUMN_STATS\":{\"key\":\"true\",\"value\":\"true\"}}
numFiles 12
numRows 500
rawDataSize 5312
@@ -348,7 +348,7 @@ Database: default
Table: tst1
#### A masked pattern was here ####
Partition Parameters:
- COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\"}
+ COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\",\"COLUMN_STATS\":{\"key\":\"true\",\"value\":\"true\"}}
#### A masked pattern was here ####
numFiles 12
numRows 500
http://git-wip-us.apache.org/repos/asf/hive/blob/3bbc24d2/ql/src/test/results/clientpositive/alter_partition_coltype.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/alter_partition_coltype.q.out b/ql/src/test/results/clientpositive/alter_partition_coltype.q.out
index 72322d9..361b196 100644
--- a/ql/src/test/results/clientpositive/alter_partition_coltype.q.out
+++ b/ql/src/test/results/clientpositive/alter_partition_coltype.q.out
@@ -231,7 +231,7 @@ STAGE PLANS:
dt 100
ts 3.0
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
+ COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}}
bucket_count -1
column.name.delimiter ,
columns key,value
@@ -277,7 +277,7 @@ STAGE PLANS:
dt 100
ts 6.30
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
+ COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}}
bucket_count -1
column.name.delimiter ,
columns key,value
@@ -465,7 +465,7 @@ STAGE PLANS:
partcol1 1
partcol2 1
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
+ COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"intcol":"true"}}
bucket_count -1
column.name.delimiter ,
columns intcol
@@ -534,7 +534,7 @@ STAGE PLANS:
partcol1 2
partcol2 1
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
+ COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"intcol":"true"}}
bucket_count -1
column.name.delimiter ,
columns intcol
http://git-wip-us.apache.org/repos/asf/hive/blob/3bbc24d2/ql/src/test/results/clientpositive/alter_table_add_partition.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/alter_table_add_partition.q.out b/ql/src/test/results/clientpositive/alter_table_add_partition.q.out
index ca941dd..a8b4d5b 100644
--- a/ql/src/test/results/clientpositive/alter_table_add_partition.q.out
+++ b/ql/src/test/results/clientpositive/alter_table_add_partition.q.out
@@ -189,7 +189,7 @@ Database: default
Table: mp
#### A masked pattern was here ####
Partition Parameters:
- COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\"}
+ COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\",\"COLUMN_STATS\":{\"a\":\"true\"}}
numFiles 1
numRows 1
rawDataSize 1
http://git-wip-us.apache.org/repos/asf/hive/blob/3bbc24d2/ql/src/test/results/clientpositive/alter_table_serde2.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/alter_table_serde2.q.out b/ql/src/test/results/clientpositive/alter_table_serde2.q.out
index 1c798b0..81b1f5b 100644
--- a/ql/src/test/results/clientpositive/alter_table_serde2.q.out
+++ b/ql/src/test/results/clientpositive/alter_table_serde2.q.out
@@ -75,7 +75,7 @@ Database: default
Table: tst1
#### A masked pattern was here ####
Partition Parameters:
- COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\"}
+ COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\",\"COLUMN_STATS\":{\"key\":\"true\",\"value\":\"true\"}}
numFiles 1
numRows 500
rawDataSize 5312
@@ -171,7 +171,7 @@ Database: default
Table: tst1
#### A masked pattern was here ####
Partition Parameters:
- COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\"}
+ COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\",\"COLUMN_STATS\":{\"key\":\"true\",\"value\":\"true\"}}
numFiles 1
numRows 500
rawDataSize 5312
http://git-wip-us.apache.org/repos/asf/hive/blob/3bbc24d2/ql/src/test/results/clientpositive/analyze_table_null_partition.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/analyze_table_null_partition.q.out b/ql/src/test/results/clientpositive/analyze_table_null_partition.q.out
index 324b700..d48df75 100644
--- a/ql/src/test/results/clientpositive/analyze_table_null_partition.q.out
+++ b/ql/src/test/results/clientpositive/analyze_table_null_partition.q.out
@@ -101,7 +101,7 @@ STAGE PLANS:
partition values:
age 15
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
+ COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"name":"true"}}
bucket_count -1
column.name.delimiter ,
columns name
@@ -146,7 +146,7 @@ STAGE PLANS:
partition values:
age 30
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
+ COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"name":"true"}}
bucket_count -1
column.name.delimiter ,
columns name
@@ -191,7 +191,7 @@ STAGE PLANS:
partition values:
age 40
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
+ COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"name":"true"}}
bucket_count -1
column.name.delimiter ,
columns name
@@ -236,7 +236,7 @@ STAGE PLANS:
partition values:
age __HIVE_DEFAULT_PARTITION__
properties:
- COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
+ COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"name":"true"}}
bucket_count -1
column.name.delimiter ,
columns name
http://git-wip-us.apache.org/repos/asf/hive/blob/3bbc24d2/ql/src/test/results/clientpositive/annotate_stats_filter.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/annotate_stats_filter.q.out b/ql/src/test/results/clientpositive/annotate_stats_filter.q.out
index 20f870e..ae300aa 100644
--- a/ql/src/test/results/clientpositive/annotate_stats_filter.q.out
+++ b/ql/src/test/results/clientpositive/annotate_stats_filter.q.out
@@ -66,11 +66,11 @@ STAGE PLANS:
Processor Tree:
TableScan
alias: loc_orc
- Statistics: Num rows: 8 Data size: 1600 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 8 Data size: 816 Basic stats: COMPLETE Column stats: COMPLETE
Select Operator
expressions: state (type: string), locid (type: int), zip (type: bigint), year (type: int)
outputColumnNames: _col0, _col1, _col2, _col3
- Statistics: Num rows: 8 Data size: 1600 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 8 Data size: 816 Basic stats: COMPLETE Column stats: COMPLETE
ListSink
PREHOOK: query: explain select * from loc_orc where state='OH'
@@ -87,17 +87,17 @@ STAGE PLANS:
Map Operator Tree:
TableScan
alias: loc_orc
- Statistics: Num rows: 8 Data size: 1600 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 8 Data size: 816 Basic stats: COMPLETE Column stats: COMPLETE
Filter Operator
predicate: (state = 'OH') (type: boolean)
- Statistics: Num rows: 8 Data size: 1600 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 102 Basic stats: COMPLETE Column stats: COMPLETE
Select Operator
expressions: 'OH' (type: string), locid (type: int), zip (type: bigint), year (type: int)
outputColumnNames: _col0, _col1, _col2, _col3
- Statistics: Num rows: 8 Data size: 1600 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 102 Basic stats: COMPLETE Column stats: COMPLETE
File Output Operator
compressed: false
- Statistics: Num rows: 8 Data size: 1600 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 102 Basic stats: COMPLETE Column stats: COMPLETE
table:
input format: org.apache.hadoop.mapred.SequenceFileInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
http://git-wip-us.apache.org/repos/asf/hive/blob/3bbc24d2/ql/src/test/results/clientpositive/annotate_stats_groupby.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/annotate_stats_groupby.q.out b/ql/src/test/results/clientpositive/annotate_stats_groupby.q.out
index 4e67841..ed3d594 100644
--- a/ql/src/test/results/clientpositive/annotate_stats_groupby.q.out
+++ b/ql/src/test/results/clientpositive/annotate_stats_groupby.q.out
@@ -66,11 +66,11 @@ STAGE PLANS:
Processor Tree:
TableScan
alias: loc_orc
- Statistics: Num rows: 8 Data size: 1600 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 8 Data size: 816 Basic stats: COMPLETE Column stats: COMPLETE
Select Operator
expressions: state (type: string), locid (type: int), zip (type: bigint), year (type: int)
outputColumnNames: _col0, _col1, _col2, _col3
- Statistics: Num rows: 8 Data size: 1600 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 8 Data size: 816 Basic stats: COMPLETE Column stats: COMPLETE
ListSink
PREHOOK: query: analyze table loc_orc compute statistics for columns state
@@ -108,22 +108,22 @@ STAGE PLANS:
Map Operator Tree:
TableScan
alias: loc_orc
- Statistics: Num rows: 8 Data size: 720 Basic stats: COMPLETE Column stats: PARTIAL
+ Statistics: Num rows: 8 Data size: 720 Basic stats: COMPLETE Column stats: COMPLETE
Select Operator
expressions: state (type: string), locid (type: int)
outputColumnNames: state, locid
- Statistics: Num rows: 8 Data size: 720 Basic stats: COMPLETE Column stats: PARTIAL
+ Statistics: Num rows: 8 Data size: 720 Basic stats: COMPLETE Column stats: COMPLETE
Group By Operator
aggregations: count()
keys: state (type: string), locid (type: int)
mode: hash
outputColumnNames: _col0, _col1, _col2
- Statistics: Num rows: 8 Data size: 784 Basic stats: COMPLETE Column stats: PARTIAL
+ Statistics: Num rows: 8 Data size: 784 Basic stats: COMPLETE Column stats: COMPLETE
Reduce Output Operator
key expressions: _col0 (type: string), _col1 (type: int)
sort order: ++
Map-reduce partition columns: _col0 (type: string), _col1 (type: int)
- Statistics: Num rows: 8 Data size: 784 Basic stats: COMPLETE Column stats: PARTIAL
+ Statistics: Num rows: 8 Data size: 784 Basic stats: COMPLETE Column stats: COMPLETE
value expressions: _col2 (type: bigint)
Reduce Operator Tree:
Group By Operator
@@ -131,13 +131,13 @@ STAGE PLANS:
keys: KEY._col0 (type: string), KEY._col1 (type: int)
mode: mergepartial
outputColumnNames: _col0, _col1, _col2
- Statistics: Num rows: 6 Data size: 588 Basic stats: COMPLETE Column stats: PARTIAL
+ Statistics: Num rows: 8 Data size: 784 Basic stats: COMPLETE Column stats: COMPLETE
Group By Operator
aggregations: min(_col1)
keys: _col0 (type: string), _col2 (type: bigint)
mode: hash
outputColumnNames: _col0, _col1, _col2
- Statistics: Num rows: 6 Data size: 588 Basic stats: COMPLETE Column stats: PARTIAL
+ Statistics: Num rows: 8 Data size: 784 Basic stats: COMPLETE Column stats: COMPLETE
File Output Operator
compressed: false
table:
@@ -153,7 +153,7 @@ STAGE PLANS:
key expressions: _col0 (type: string), _col1 (type: bigint)
sort order: ++
Map-reduce partition columns: _col0 (type: string), _col1 (type: bigint)
- Statistics: Num rows: 6 Data size: 588 Basic stats: COMPLETE Column stats: PARTIAL
+ Statistics: Num rows: 8 Data size: 784 Basic stats: COMPLETE Column stats: COMPLETE
value expressions: _col2 (type: int)
Reduce Operator Tree:
Group By Operator
@@ -161,10 +161,10 @@ STAGE PLANS:
keys: KEY._col0 (type: string), KEY._col1 (type: bigint)
mode: mergepartial
outputColumnNames: _col0, _col1, _col2
- Statistics: Num rows: 6 Data size: 588 Basic stats: COMPLETE Column stats: PARTIAL
+ Statistics: Num rows: 8 Data size: 784 Basic stats: COMPLETE Column stats: COMPLETE
File Output Operator
compressed: false
- Statistics: Num rows: 6 Data size: 588 Basic stats: COMPLETE Column stats: PARTIAL
+ Statistics: Num rows: 8 Data size: 784 Basic stats: COMPLETE Column stats: COMPLETE
table:
input format: org.apache.hadoop.mapred.SequenceFileInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
@@ -747,30 +747,30 @@ STAGE PLANS:
Map Operator Tree:
TableScan
alias: loc_orc
- Statistics: Num rows: 8 Data size: 752 Basic stats: COMPLETE Column stats: PARTIAL
+ Statistics: Num rows: 8 Data size: 752 Basic stats: COMPLETE Column stats: COMPLETE
Select Operator
expressions: state (type: string), zip (type: bigint)
outputColumnNames: state, zip
- Statistics: Num rows: 8 Data size: 752 Basic stats: COMPLETE Column stats: PARTIAL
+ Statistics: Num rows: 8 Data size: 752 Basic stats: COMPLETE Column stats: COMPLETE
Group By Operator
keys: state (type: string), zip (type: bigint)
mode: hash
outputColumnNames: _col0, _col1
- Statistics: Num rows: 4 Data size: 376 Basic stats: COMPLETE Column stats: PARTIAL
+ Statistics: Num rows: 4 Data size: 376 Basic stats: COMPLETE Column stats: COMPLETE
Reduce Output Operator
key expressions: _col0 (type: string), _col1 (type: bigint)
sort order: ++
Map-reduce partition columns: _col0 (type: string), _col1 (type: bigint)
- Statistics: Num rows: 4 Data size: 376 Basic stats: COMPLETE Column stats: PARTIAL
+ Statistics: Num rows: 4 Data size: 376 Basic stats: COMPLETE Column stats: COMPLETE
Reduce Operator Tree:
Group By Operator
keys: KEY._col0 (type: string), KEY._col1 (type: bigint)
mode: mergepartial
outputColumnNames: _col0, _col1
- Statistics: Num rows: 4 Data size: 376 Basic stats: COMPLETE Column stats: PARTIAL
+ Statistics: Num rows: 4 Data size: 376 Basic stats: COMPLETE Column stats: COMPLETE
File Output Operator
compressed: false
- Statistics: Num rows: 4 Data size: 376 Basic stats: COMPLETE Column stats: PARTIAL
+ Statistics: Num rows: 4 Data size: 376 Basic stats: COMPLETE Column stats: COMPLETE
table:
input format: org.apache.hadoop.mapred.SequenceFileInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
http://git-wip-us.apache.org/repos/asf/hive/blob/3bbc24d2/ql/src/test/results/clientpositive/array_size_estimation.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/array_size_estimation.q.out b/ql/src/test/results/clientpositive/array_size_estimation.q.out
index 3cd205f..3a0a9df 100644
--- a/ql/src/test/results/clientpositive/array_size_estimation.q.out
+++ b/ql/src/test/results/clientpositive/array_size_estimation.q.out
@@ -64,14 +64,14 @@ STAGE PLANS:
Map Operator Tree:
TableScan
alias: t
- Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 85 Basic stats: COMPLETE Column stats: COMPLETE
Select Operator
expressions: array('b','d','c',col) (type: array<string>)
outputColumnNames: _col0
- Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 1920 Basic stats: COMPLETE Column stats: COMPLETE
File Output Operator
compressed: false
- Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 1920 Basic stats: COMPLETE Column stats: COMPLETE
table:
input format: org.apache.hadoop.mapred.SequenceFileInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
@@ -134,14 +134,14 @@ STAGE PLANS:
Map Operator Tree:
TableScan
alias: t
- Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 85 Basic stats: COMPLETE Column stats: COMPLETE
Select Operator
expressions: sort_array(array('b','d','c',col)) (type: array<string>), array('1','2') (type: array<string>)
outputColumnNames: _col0, _col1
- Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 2328 Basic stats: COMPLETE Column stats: COMPLETE
File Output Operator
compressed: false
- Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: NONE
+ Statistics: Num rows: 1 Data size: 2328 Basic stats: COMPLETE Column stats: COMPLETE
table:
input format: org.apache.hadoop.mapred.SequenceFileInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
http://git-wip-us.apache.org/repos/asf/hive/blob/3bbc24d2/ql/src/test/results/clientpositive/auto_join1.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/auto_join1.q.out b/ql/src/test/results/clientpositive/auto_join1.q.out
index c2d2473..2edfb86 100644
--- a/ql/src/test/results/clientpositive/auto_join1.q.out
+++ b/ql/src/test/results/clientpositive/auto_join1.q.out
@@ -15,13 +15,14 @@ FROM src src1 JOIN src src2 ON (src1.key = src2.key)
INSERT OVERWRITE TABLE dest_j1 SELECT src1.key, src2.value
POSTHOOK: type: QUERY
STAGE DEPENDENCIES:
- Stage-5 is a root stage
- Stage-4 depends on stages: Stage-5
- Stage-0 depends on stages: Stage-4
- Stage-2 depends on stages: Stage-0
+ Stage-6 is a root stage
+ Stage-5 depends on stages: Stage-6
+ Stage-0 depends on stages: Stage-5
+ Stage-2 depends on stages: Stage-0, Stage-3
+ Stage-3 depends on stages: Stage-5
STAGE PLANS:
- Stage: Stage-5
+ Stage: Stage-6
Map Reduce Local Work
Alias -> Map Local Tables:
$hdt$_0:src1
@@ -44,7 +45,7 @@ STAGE PLANS:
0 _col0 (type: string)
1 _col0 (type: string)
- Stage: Stage-4
+ Stage: Stage-5
Map Reduce
Map Operator Tree:
TableScan
@@ -77,6 +78,21 @@ STAGE PLANS:
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.dest_j1
+ Select Operator
+ expressions: _col0 (type: int), _col1 (type: string)
+ outputColumnNames: key, value
+ Statistics: Num rows: 550 Data size: 5843 Basic stats: COMPLETE Column stats: NONE
+ Group By Operator
+ aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll')
+ mode: hash
+ outputColumnNames: _col0, _col1
+ Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: NONE
+ File Output Operator
+ compressed: false
+ table:
+ input format: org.apache.hadoop.mapred.SequenceFileInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe
Local Work:
Map Reduce Local Work
@@ -93,6 +109,32 @@ STAGE PLANS:
Stage: Stage-2
Stats Work
Basic Stats Work:
+ Column Stats Desc:
+ Columns: key, value
+ Column Types: int, string
+ Table: default.dest_j1
+
+ Stage: Stage-3
+ Map Reduce
+ Map Operator Tree:
+ TableScan
+ Reduce Output Operator
+ sort order:
+ Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: NONE
+ value expressions: _col0 (type: struct<columntype:string,min:bigint,max:bigint,countnulls:bigint,bitvector:binary>), _col1 (type: struct<columntype:string,maxlength:bigint,sumlength:bigint,count:bigint,countnulls:bigint,bitvector:binary>)
+ Reduce Operator Tree:
+ Group By Operator
+ aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1)
+ mode: mergepartial
+ outputColumnNames: _col0, _col1
+ Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: NONE
+ File Output Operator
+ compressed: false
+ Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: NONE
+ table:
+ input format: org.apache.hadoop.mapred.SequenceFileInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
PREHOOK: query: FROM src src1 JOIN src src2 ON (src1.key = src2.key)
INSERT OVERWRITE TABLE dest_j1 SELECT src1.key, src2.value
http://git-wip-us.apache.org/repos/asf/hive/blob/3bbc24d2/ql/src/test/results/clientpositive/auto_join14.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/auto_join14.q.out b/ql/src/test/results/clientpositive/auto_join14.q.out
index e67506b..52aaa02 100644
--- a/ql/src/test/results/clientpositive/auto_join14.q.out
+++ b/ql/src/test/results/clientpositive/auto_join14.q.out
@@ -15,13 +15,14 @@ FROM src JOIN srcpart ON src.key = srcpart.key AND srcpart.ds = '2008-04-08' and
INSERT OVERWRITE TABLE dest1 SELECT src.key, srcpart.value
POSTHOOK: type: QUERY
STAGE DEPENDENCIES:
- Stage-5 is a root stage
- Stage-4 depends on stages: Stage-5
- Stage-0 depends on stages: Stage-4
- Stage-2 depends on stages: Stage-0
+ Stage-6 is a root stage
+ Stage-5 depends on stages: Stage-6
+ Stage-0 depends on stages: Stage-5
+ Stage-2 depends on stages: Stage-0, Stage-3
+ Stage-3 depends on stages: Stage-5
STAGE PLANS:
- Stage: Stage-5
+ Stage: Stage-6
Map Reduce Local Work
Alias -> Map Local Tables:
$hdt$_0:src
@@ -44,7 +45,7 @@ STAGE PLANS:
0 _col0 (type: string)
1 _col0 (type: string)
- Stage: Stage-4
+ Stage: Stage-5
Map Reduce
Map Operator Tree:
TableScan
@@ -77,6 +78,21 @@ STAGE PLANS:
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.dest1
+ Select Operator
+ expressions: _col0 (type: int), _col1 (type: string)
+ outputColumnNames: c1, c2
+ Statistics: Num rows: 366 Data size: 3890 Basic stats: COMPLETE Column stats: NONE
+ Group By Operator
+ aggregations: compute_stats(c1, 'hll'), compute_stats(c2, 'hll')
+ mode: hash
+ outputColumnNames: _col0, _col1
+ Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: NONE
+ File Output Operator
+ compressed: false
+ table:
+ input format: org.apache.hadoop.mapred.SequenceFileInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe
Local Work:
Map Reduce Local Work
@@ -93,6 +109,32 @@ STAGE PLANS:
Stage: Stage-2
Stats Work
Basic Stats Work:
+ Column Stats Desc:
+ Columns: c1, c2
+ Column Types: int, string
+ Table: default.dest1
+
+ Stage: Stage-3
+ Map Reduce
+ Map Operator Tree:
+ TableScan
+ Reduce Output Operator
+ sort order:
+ Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: NONE
+ value expressions: _col0 (type: struct<columntype:string,min:bigint,max:bigint,countnulls:bigint,bitvector:binary>), _col1 (type: struct<columntype:string,maxlength:bigint,sumlength:bigint,count:bigint,countnulls:bigint,bitvector:binary>)
+ Reduce Operator Tree:
+ Group By Operator
+ aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1)
+ mode: mergepartial
+ outputColumnNames: _col0, _col1
+ Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: NONE
+ File Output Operator
+ compressed: false
+ Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: NONE
+ table:
+ input format: org.apache.hadoop.mapred.SequenceFileInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
PREHOOK: query: FROM src JOIN srcpart ON src.key = srcpart.key AND srcpart.ds = '2008-04-08' and src.key > 100
INSERT OVERWRITE TABLE dest1 SELECT src.key, srcpart.value
http://git-wip-us.apache.org/repos/asf/hive/blob/3bbc24d2/ql/src/test/results/clientpositive/auto_join17.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/auto_join17.q.out b/ql/src/test/results/clientpositive/auto_join17.q.out
index 0239cf8..20adc30 100644
--- a/ql/src/test/results/clientpositive/auto_join17.q.out
+++ b/ql/src/test/results/clientpositive/auto_join17.q.out
@@ -15,13 +15,14 @@ FROM src src1 JOIN src src2 ON (src1.key = src2.key)
INSERT OVERWRITE TABLE dest1 SELECT src1.*, src2.*
POSTHOOK: type: QUERY
STAGE DEPENDENCIES:
- Stage-5 is a root stage
- Stage-4 depends on stages: Stage-5
- Stage-0 depends on stages: Stage-4
- Stage-2 depends on stages: Stage-0
+ Stage-6 is a root stage
+ Stage-5 depends on stages: Stage-6
+ Stage-0 depends on stages: Stage-5
+ Stage-2 depends on stages: Stage-0, Stage-3
+ Stage-3 depends on stages: Stage-5
STAGE PLANS:
- Stage: Stage-5
+ Stage: Stage-6
Map Reduce Local Work
Alias -> Map Local Tables:
$hdt$_0:src1
@@ -44,7 +45,7 @@ STAGE PLANS:
0 _col0 (type: string)
1 _col0 (type: string)
- Stage: Stage-4
+ Stage: Stage-5
Map Reduce
Map Operator Tree:
TableScan
@@ -77,6 +78,21 @@ STAGE PLANS:
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.dest1
+ Select Operator
+ expressions: _col0 (type: int), _col1 (type: string), _col2 (type: int), _col3 (type: string)
+ outputColumnNames: key1, value1, key2, value2
+ Statistics: Num rows: 550 Data size: 5843 Basic stats: COMPLETE Column stats: NONE
+ Group By Operator
+ aggregations: compute_stats(key1, 'hll'), compute_stats(value1, 'hll'), compute_stats(key2, 'hll'), compute_stats(value2, 'hll')
+ mode: hash
+ outputColumnNames: _col0, _col1, _col2, _col3
+ Statistics: Num rows: 1 Data size: 1728 Basic stats: COMPLETE Column stats: NONE
+ File Output Operator
+ compressed: false
+ table:
+ input format: org.apache.hadoop.mapred.SequenceFileInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe
Local Work:
Map Reduce Local Work
@@ -93,6 +109,32 @@ STAGE PLANS:
Stage: Stage-2
Stats Work
Basic Stats Work:
+ Column Stats Desc:
+ Columns: key1, value1, key2, value2
+ Column Types: int, string, int, string
+ Table: default.dest1
+
+ Stage: Stage-3
+ Map Reduce
+ Map Operator Tree:
+ TableScan
+ Reduce Output Operator
+ sort order:
+ Statistics: Num rows: 1 Data size: 1728 Basic stats: COMPLETE Column stats: NONE
+ value expressions: _col0 (type: struct<columntype:string,min:bigint,max:bigint,countnulls:bigint,bitvector:binary>), _col1 (type: struct<columntype:string,maxlength:bigint,sumlength:bigint,count:bigint,countnulls:bigint,bitvector:binary>), _col2 (type: struct<columntype:string,min:bigint,max:bigint,countnulls:bigint,bitvector:binary>), _col3 (type: struct<columntype:string,maxlength:bigint,sumlength:bigint,count:bigint,countnulls:bigint,bitvector:binary>)
+ Reduce Operator Tree:
+ Group By Operator
+ aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2), compute_stats(VALUE._col3)
+ mode: mergepartial
+ outputColumnNames: _col0, _col1, _col2, _col3
+ Statistics: Num rows: 1 Data size: 1760 Basic stats: COMPLETE Column stats: NONE
+ File Output Operator
+ compressed: false
+ Statistics: Num rows: 1 Data size: 1760 Basic stats: COMPLETE Column stats: NONE
+ table:
+ input format: org.apache.hadoop.mapred.SequenceFileInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
PREHOOK: query: FROM src src1 JOIN src src2 ON (src1.key = src2.key)
INSERT OVERWRITE TABLE dest1 SELECT src1.*, src2.*
http://git-wip-us.apache.org/repos/asf/hive/blob/3bbc24d2/ql/src/test/results/clientpositive/auto_join19.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/auto_join19.q.out b/ql/src/test/results/clientpositive/auto_join19.q.out
index e70c514..80615de 100644
--- a/ql/src/test/results/clientpositive/auto_join19.q.out
+++ b/ql/src/test/results/clientpositive/auto_join19.q.out
@@ -17,13 +17,14 @@ INSERT OVERWRITE TABLE dest1 SELECT src1.key, src2.value
where (src1.ds = '2008-04-08' or src1.ds = '2008-04-09' )and (src1.hr = '12' or src1.hr = '11')
POSTHOOK: type: QUERY
STAGE DEPENDENCIES:
- Stage-5 is a root stage
- Stage-4 depends on stages: Stage-5
- Stage-0 depends on stages: Stage-4
- Stage-2 depends on stages: Stage-0
+ Stage-6 is a root stage
+ Stage-5 depends on stages: Stage-6
+ Stage-0 depends on stages: Stage-5
+ Stage-2 depends on stages: Stage-0, Stage-3
+ Stage-3 depends on stages: Stage-5
STAGE PLANS:
- Stage: Stage-5
+ Stage: Stage-6
Map Reduce Local Work
Alias -> Map Local Tables:
$hdt$_1:src2
@@ -46,7 +47,7 @@ STAGE PLANS:
0 _col0 (type: string)
1 _col0 (type: string)
- Stage: Stage-4
+ Stage: Stage-5
Map Reduce
Map Operator Tree:
TableScan
@@ -79,6 +80,21 @@ STAGE PLANS:
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: default.dest1
+ Select Operator
+ expressions: _col0 (type: int), _col1 (type: string)
+ outputColumnNames: key, value
+ Statistics: Num rows: 2200 Data size: 23372 Basic stats: COMPLETE Column stats: NONE
+ Group By Operator
+ aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll')
+ mode: hash
+ outputColumnNames: _col0, _col1
+ Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: NONE
+ File Output Operator
+ compressed: false
+ table:
+ input format: org.apache.hadoop.mapred.SequenceFileInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe
Local Work:
Map Reduce Local Work
@@ -95,6 +111,32 @@ STAGE PLANS:
Stage: Stage-2
Stats Work
Basic Stats Work:
+ Column Stats Desc:
+ Columns: key, value
+ Column Types: int, string
+ Table: default.dest1
+
+ Stage: Stage-3
+ Map Reduce
+ Map Operator Tree:
+ TableScan
+ Reduce Output Operator
+ sort order:
+ Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: NONE
+ value expressions: _col0 (type: struct<columntype:string,min:bigint,max:bigint,countnulls:bigint,bitvector:binary>), _col1 (type: struct<columntype:string,maxlength:bigint,sumlength:bigint,count:bigint,countnulls:bigint,bitvector:binary>)
+ Reduce Operator Tree:
+ Group By Operator
+ aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1)
+ mode: mergepartial
+ outputColumnNames: _col0, _col1
+ Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: NONE
+ File Output Operator
+ compressed: false
+ Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: NONE
+ table:
+ input format: org.apache.hadoop.mapred.SequenceFileInputFormat
+ output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
+ serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
PREHOOK: query: FROM srcpart src1 JOIN src src2 ON (src1.key = src2.key)
INSERT OVERWRITE TABLE dest1 SELECT src1.key, src2.value