You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hive.apache.org by jc...@apache.org on 2018/05/21 20:25:15 UTC
[38/51] [abbrv] [partial] hive git commit: HIVE-19617: Rename test
tables to avoid collisions during execution in batches (Jesus Camacho
Rodriguez, reviewed by Gunther Hagleitner)
http://git-wip-us.apache.org/repos/asf/hive/blob/38d3b8e1/ql/src/test/queries/clientpositive/dynamic_semijoin_reduction.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/dynamic_semijoin_reduction.q b/ql/src/test/queries/clientpositive/dynamic_semijoin_reduction.q
index 705dfdc..32c2854 100644
--- a/ql/src/test/queries/clientpositive/dynamic_semijoin_reduction.q
+++ b/ql/src/test/queries/clientpositive/dynamic_semijoin_reduction.q
@@ -16,67 +16,67 @@ set hive.stats.fetch.column.stats=true;
set hive.tez.bloom.filter.factor=1.0f;
-- Create Tables
-create table alltypesorc_int ( cint int, cstring string ) stored as ORC;
-create table srcpart_date (key string, value string) partitioned by (ds string ) stored as ORC;
-CREATE TABLE srcpart_small(key1 STRING, value1 STRING) partitioned by (ds string) STORED as ORC;
+create table alltypesorc_int_n1 ( cint int, cstring string ) stored as ORC;
+create table srcpart_date_n7 (key string, value string) partitioned by (ds string ) stored as ORC;
+CREATE TABLE srcpart_small_n3(key1 STRING, value1 STRING) partitioned by (ds string) STORED as ORC;
-- Add Partitions
-alter table srcpart_date add partition (ds = "2008-04-08");
-alter table srcpart_date add partition (ds = "2008-04-09");
+alter table srcpart_date_n7 add partition (ds = "2008-04-08");
+alter table srcpart_date_n7 add partition (ds = "2008-04-09");
-alter table srcpart_small add partition (ds = "2008-04-08");
-alter table srcpart_small add partition (ds = "2008-04-09");
+alter table srcpart_small_n3 add partition (ds = "2008-04-08");
+alter table srcpart_small_n3 add partition (ds = "2008-04-09");
-- Load
-insert overwrite table alltypesorc_int select cint, cstring1 from alltypesorc;
-insert overwrite table srcpart_date partition (ds = "2008-04-08" ) select key, value from srcpart where ds = "2008-04-08";
-insert overwrite table srcpart_date partition (ds = "2008-04-09") select key, value from srcpart where ds = "2008-04-09";
-insert overwrite table srcpart_small partition (ds = "2008-04-09") select key, value from srcpart where ds = "2008-04-09" limit 20;
+insert overwrite table alltypesorc_int_n1 select cint, cstring1 from alltypesorc;
+insert overwrite table srcpart_date_n7 partition (ds = "2008-04-08" ) select key, value from srcpart where ds = "2008-04-08";
+insert overwrite table srcpart_date_n7 partition (ds = "2008-04-09") select key, value from srcpart where ds = "2008-04-09";
+insert overwrite table srcpart_small_n3 partition (ds = "2008-04-09") select key, value from srcpart where ds = "2008-04-09" limit 20;
set hive.tez.dynamic.semijoin.reduction=false;
-analyze table alltypesorc_int compute statistics for columns;
-analyze table srcpart_date compute statistics for columns;
-analyze table srcpart_small compute statistics for columns;
+analyze table alltypesorc_int_n1 compute statistics for columns;
+analyze table srcpart_date_n7 compute statistics for columns;
+analyze table srcpart_small_n3 compute statistics for columns;
-- single column, single key
-EXPLAIN select count(*) from srcpart_date join srcpart_small on (srcpart_date.key = srcpart_small.key1);
-select count(*) from srcpart_date join srcpart_small on (srcpart_date.key = srcpart_small.key1);
+EXPLAIN select count(*) from srcpart_date_n7 join srcpart_small_n3 on (srcpart_date_n7.key = srcpart_small_n3.key1);
+select count(*) from srcpart_date_n7 join srcpart_small_n3 on (srcpart_date_n7.key = srcpart_small_n3.key1);
set hive.tez.dynamic.semijoin.reduction=true;
-EXPLAIN select count(*) from srcpart_date join srcpart_small on (srcpart_date.key = srcpart_small.key1);
-select count(*) from srcpart_date join srcpart_small on (srcpart_date.key = srcpart_small.key1);
+EXPLAIN select count(*) from srcpart_date_n7 join srcpart_small_n3 on (srcpart_date_n7.key = srcpart_small_n3.key1);
+select count(*) from srcpart_date_n7 join srcpart_small_n3 on (srcpart_date_n7.key = srcpart_small_n3.key1);
set hive.tez.dynamic.semijoin.reduction=true;
-- Mix dynamic partition pruning(DPP) and min/max bloom filter optimizations. Should pick the DPP.
-EXPLAIN select count(*) from srcpart_date join srcpart_small on (srcpart_date.key = srcpart_small.ds);
-select count(*) from srcpart_date join srcpart_small on (srcpart_date.key = srcpart_small.ds);
+EXPLAIN select count(*) from srcpart_date_n7 join srcpart_small_n3 on (srcpart_date_n7.key = srcpart_small_n3.ds);
+select count(*) from srcpart_date_n7 join srcpart_small_n3 on (srcpart_date_n7.key = srcpart_small_n3.ds);
set hive.tez.dynamic.semijoin.reduction=false;
--multiple sources, single key
-EXPLAIN select count(*) from srcpart_date join srcpart_small on (srcpart_date.key = srcpart_small.key1) join alltypesorc_int on (srcpart_small.key1 = alltypesorc_int.cstring);
-select count(*) from srcpart_date join srcpart_small on (srcpart_date.key = srcpart_small.key1) join alltypesorc_int on (srcpart_small.key1 = alltypesorc_int.cstring);
+EXPLAIN select count(*) from srcpart_date_n7 join srcpart_small_n3 on (srcpart_date_n7.key = srcpart_small_n3.key1) join alltypesorc_int_n1 on (srcpart_small_n3.key1 = alltypesorc_int_n1.cstring);
+select count(*) from srcpart_date_n7 join srcpart_small_n3 on (srcpart_date_n7.key = srcpart_small_n3.key1) join alltypesorc_int_n1 on (srcpart_small_n3.key1 = alltypesorc_int_n1.cstring);
set hive.tez.dynamic.semijoin.reduction=true;
-EXPLAIN select count(*) from srcpart_date join srcpart_small on (srcpart_date.key = srcpart_small.key1) join alltypesorc_int on (srcpart_small.key1 = alltypesorc_int.cstring);
-select count(*) from srcpart_date join srcpart_small on (srcpart_date.key = srcpart_small.key1) join alltypesorc_int on (srcpart_small.key1 = alltypesorc_int.cstring);
+EXPLAIN select count(*) from srcpart_date_n7 join srcpart_small_n3 on (srcpart_date_n7.key = srcpart_small_n3.key1) join alltypesorc_int_n1 on (srcpart_small_n3.key1 = alltypesorc_int_n1.cstring);
+select count(*) from srcpart_date_n7 join srcpart_small_n3 on (srcpart_date_n7.key = srcpart_small_n3.key1) join alltypesorc_int_n1 on (srcpart_small_n3.key1 = alltypesorc_int_n1.cstring);
set hive.tez.dynamic.semijoin.reduction=false;
-- single source, multiple keys
-EXPLAIN select count(*) from srcpart_date join srcpart_small on (srcpart_date.key = srcpart_small.key1 and srcpart_date.value = srcpart_small.value1);
-select count(*) from srcpart_date join srcpart_small on (srcpart_date.key = srcpart_small.key1 and srcpart_date.value = srcpart_small.value1);
+EXPLAIN select count(*) from srcpart_date_n7 join srcpart_small_n3 on (srcpart_date_n7.key = srcpart_small_n3.key1 and srcpart_date_n7.value = srcpart_small_n3.value1);
+select count(*) from srcpart_date_n7 join srcpart_small_n3 on (srcpart_date_n7.key = srcpart_small_n3.key1 and srcpart_date_n7.value = srcpart_small_n3.value1);
set hive.tez.dynamic.semijoin.reduction=true;
-EXPLAIN select count(*) from srcpart_date join srcpart_small on (srcpart_date.key = srcpart_small.key1 and srcpart_date.value = srcpart_small.value1);
-select count(*) from srcpart_date join srcpart_small on (srcpart_date.key = srcpart_small.key1 and srcpart_date.value = srcpart_small.value1);
+EXPLAIN select count(*) from srcpart_date_n7 join srcpart_small_n3 on (srcpart_date_n7.key = srcpart_small_n3.key1 and srcpart_date_n7.value = srcpart_small_n3.value1);
+select count(*) from srcpart_date_n7 join srcpart_small_n3 on (srcpart_date_n7.key = srcpart_small_n3.key1 and srcpart_date_n7.value = srcpart_small_n3.value1);
set hive.tez.dynamic.semijoin.reduction=false;
-- multiple sources, different keys
-EXPLAIN select count(*) from srcpart_date join srcpart_small on (srcpart_date.key = srcpart_small.key1) join alltypesorc_int on (srcpart_date.value = alltypesorc_int.cstring);
-select count(*) from srcpart_date join srcpart_small on (srcpart_date.key = srcpart_small.key1) join alltypesorc_int on (srcpart_date.value = alltypesorc_int.cstring);
+EXPLAIN select count(*) from srcpart_date_n7 join srcpart_small_n3 on (srcpart_date_n7.key = srcpart_small_n3.key1) join alltypesorc_int_n1 on (srcpart_date_n7.value = alltypesorc_int_n1.cstring);
+select count(*) from srcpart_date_n7 join srcpart_small_n3 on (srcpart_date_n7.key = srcpart_small_n3.key1) join alltypesorc_int_n1 on (srcpart_date_n7.value = alltypesorc_int_n1.cstring);
set hive.tez.dynamic.semijoin.reduction=true;
-EXPLAIN select count(*) from srcpart_date join srcpart_small on (srcpart_date.key = srcpart_small.key1) join alltypesorc_int on (srcpart_date.value = alltypesorc_int.cstring);
-select count(*) from srcpart_date join srcpart_small on (srcpart_date.key = srcpart_small.key1) join alltypesorc_int on (srcpart_date.value = alltypesorc_int.cstring);
+EXPLAIN select count(*) from srcpart_date_n7 join srcpart_small_n3 on (srcpart_date_n7.key = srcpart_small_n3.key1) join alltypesorc_int_n1 on (srcpart_date_n7.value = alltypesorc_int_n1.cstring);
+select count(*) from srcpart_date_n7 join srcpart_small_n3 on (srcpart_date_n7.key = srcpart_small_n3.key1) join alltypesorc_int_n1 on (srcpart_date_n7.value = alltypesorc_int_n1.cstring);
-- Explain extended to verify fast start for Reducer in semijoin branch
-EXPLAIN extended select count(*) from srcpart_date join srcpart_small on (srcpart_date.key = srcpart_small.key1);
+EXPLAIN extended select count(*) from srcpart_date_n7 join srcpart_small_n3 on (srcpart_date_n7.key = srcpart_small_n3.key1);
set hive.tez.dynamic.semijoin.reduction=false;
-- With Mapjoins, there shouldn't be any semijoin parallel to mapjoin.
@@ -84,64 +84,64 @@ set hive.auto.convert.join=true;
set hive.auto.convert.join.noconditionaltask=true;
set hive.auto.convert.join.noconditionaltask.size=100000000000;
-EXPLAIN select count(*) from srcpart_date join srcpart_small on (srcpart_date.key = srcpart_small.key1);
-select count(*) from srcpart_date join srcpart_small on (srcpart_date.key = srcpart_small.key1);
+EXPLAIN select count(*) from srcpart_date_n7 join srcpart_small_n3 on (srcpart_date_n7.key = srcpart_small_n3.key1);
+select count(*) from srcpart_date_n7 join srcpart_small_n3 on (srcpart_date_n7.key = srcpart_small_n3.key1);
set hive.tez.dynamic.semijoin.reduction=true;
-EXPLAIN select count(*) from srcpart_date join srcpart_small on (srcpart_date.key = srcpart_small.key1);
-select count(*) from srcpart_date join srcpart_small on (srcpart_date.key = srcpart_small.key1);
+EXPLAIN select count(*) from srcpart_date_n7 join srcpart_small_n3 on (srcpart_date_n7.key = srcpart_small_n3.key1);
+select count(*) from srcpart_date_n7 join srcpart_small_n3 on (srcpart_date_n7.key = srcpart_small_n3.key1);
set hive.tez.dynamic.semijoin.reduction.for.mapjoin=true;
-- Enable semijoin parallel to mapjoins.
-EXPLAIN select count(*) from srcpart_date join srcpart_small on (srcpart_date.key = srcpart_small.key1);
-select count(*) from srcpart_date join srcpart_small on (srcpart_date.key = srcpart_small.key1);
+EXPLAIN select count(*) from srcpart_date_n7 join srcpart_small_n3 on (srcpart_date_n7.key = srcpart_small_n3.key1);
+select count(*) from srcpart_date_n7 join srcpart_small_n3 on (srcpart_date_n7.key = srcpart_small_n3.key1);
set hive.tez.dynamic.semijoin.reduction.for.mapjoin=false;
set hive.tez.dynamic.semijoin.reduction=false;
-- multiple sources, different keys
-EXPLAIN select count(*) from srcpart_date join srcpart_small on (srcpart_date.key = srcpart_small.key1) join alltypesorc_int on (srcpart_date.value = alltypesorc_int.cstring);
-select count(*) from srcpart_date join srcpart_small on (srcpart_date.key = srcpart_small.key1) join alltypesorc_int on (srcpart_date.value = alltypesorc_int.cstring);
+EXPLAIN select count(*) from srcpart_date_n7 join srcpart_small_n3 on (srcpart_date_n7.key = srcpart_small_n3.key1) join alltypesorc_int_n1 on (srcpart_date_n7.value = alltypesorc_int_n1.cstring);
+select count(*) from srcpart_date_n7 join srcpart_small_n3 on (srcpart_date_n7.key = srcpart_small_n3.key1) join alltypesorc_int_n1 on (srcpart_date_n7.value = alltypesorc_int_n1.cstring);
set hive.tez.dynamic.semijoin.reduction=true;
-EXPLAIN select count(*) from srcpart_date join srcpart_small on (srcpart_date.key = srcpart_small.key1) join alltypesorc_int on (srcpart_date.value = alltypesorc_int.cstring);
-select count(*) from srcpart_date join srcpart_small on (srcpart_date.key = srcpart_small.key1) join alltypesorc_int on (srcpart_date.value = alltypesorc_int.cstring);
+EXPLAIN select count(*) from srcpart_date_n7 join srcpart_small_n3 on (srcpart_date_n7.key = srcpart_small_n3.key1) join alltypesorc_int_n1 on (srcpart_date_n7.value = alltypesorc_int_n1.cstring);
+select count(*) from srcpart_date_n7 join srcpart_small_n3 on (srcpart_date_n7.key = srcpart_small_n3.key1) join alltypesorc_int_n1 on (srcpart_date_n7.value = alltypesorc_int_n1.cstring);
set hive.tez.dynamic.semijoin.reduction.for.mapjoin=true;
-- Enable semijoin parallel to mapjoins.
-EXPLAIN select count(*) from srcpart_date join srcpart_small on (srcpart_date.key = srcpart_small.key1) join alltypesorc_int on (srcpart_date.value = alltypesorc_int.cstring);
-select count(*) from srcpart_date join srcpart_small on (srcpart_date.key = srcpart_small.key1) join alltypesorc_int on (srcpart_date.value = alltypesorc_int.cstring);
+EXPLAIN select count(*) from srcpart_date_n7 join srcpart_small_n3 on (srcpart_date_n7.key = srcpart_small_n3.key1) join alltypesorc_int_n1 on (srcpart_date_n7.value = alltypesorc_int_n1.cstring);
+select count(*) from srcpart_date_n7 join srcpart_small_n3 on (srcpart_date_n7.key = srcpart_small_n3.key1) join alltypesorc_int_n1 on (srcpart_date_n7.value = alltypesorc_int_n1.cstring);
set hive.tez.dynamic.semijoin.reduction.for.mapjoin=false;
-- HIVE-17323 - with DPP, the 1st mapjoin is on a map with DPP and 2nd mapjoin is on a map which had semijoin but still removed.
-create table alltypesorc_int40 as select * from alltypesorc_int limit 40;
+create table alltypesorc_int40 as select * from alltypesorc_int_n1 limit 40;
set hive.tez.dynamic.semijoin.reduction=false;
-EXPLAIN select count(*) from srcpart_date join srcpart_small on (srcpart_date.ds = srcpart_small.ds) join alltypesorc_int40 on (srcpart_date.value = alltypesorc_int40.cstring);
-select count(*) from srcpart_date join srcpart_small on (srcpart_date.ds = srcpart_small.ds) join alltypesorc_int40 on (srcpart_date.value = alltypesorc_int40.cstring);
+EXPLAIN select count(*) from srcpart_date_n7 join srcpart_small_n3 on (srcpart_date_n7.ds = srcpart_small_n3.ds) join alltypesorc_int40 on (srcpart_date_n7.value = alltypesorc_int40.cstring);
+select count(*) from srcpart_date_n7 join srcpart_small_n3 on (srcpart_date_n7.ds = srcpart_small_n3.ds) join alltypesorc_int40 on (srcpart_date_n7.value = alltypesorc_int40.cstring);
set hive.tez.dynamic.semijoin.reduction=true;
-EXPLAIN select count(*) from srcpart_date join srcpart_small on (srcpart_date.ds = srcpart_small.ds) join alltypesorc_int40 on (srcpart_date.value = alltypesorc_int40.cstring);
-select count(*) from srcpart_date join srcpart_small on (srcpart_date.ds = srcpart_small.ds) join alltypesorc_int40 on (srcpart_date.value = alltypesorc_int40.cstring);
+EXPLAIN select count(*) from srcpart_date_n7 join srcpart_small_n3 on (srcpart_date_n7.ds = srcpart_small_n3.ds) join alltypesorc_int40 on (srcpart_date_n7.value = alltypesorc_int40.cstring);
+select count(*) from srcpart_date_n7 join srcpart_small_n3 on (srcpart_date_n7.ds = srcpart_small_n3.ds) join alltypesorc_int40 on (srcpart_date_n7.value = alltypesorc_int40.cstring);
-- HIVE-17399
-create table srcpart_small10 as select * from srcpart_small limit 10;
+create table srcpart_small10 as select * from srcpart_small_n3 limit 10;
analyze table srcpart_small10 compute statistics for columns;
set hive.tez.dynamic.semijoin.reduction=false;
-EXPLAIN select count(*) from srcpart_small10, srcpart_small, srcpart_date where srcpart_small.key1 = srcpart_small10.key1 and srcpart_date.ds = srcpart_small.ds;
-select count(*) from srcpart_small10, srcpart_small, srcpart_date where srcpart_small.key1 = srcpart_small10.key1 and srcpart_date.ds = srcpart_small.ds;
+EXPLAIN select count(*) from srcpart_small10, srcpart_small_n3, srcpart_date_n7 where srcpart_small_n3.key1 = srcpart_small10.key1 and srcpart_date_n7.ds = srcpart_small_n3.ds;
+select count(*) from srcpart_small10, srcpart_small_n3, srcpart_date_n7 where srcpart_small_n3.key1 = srcpart_small10.key1 and srcpart_date_n7.ds = srcpart_small_n3.ds;
set hive.tez.dynamic.semijoin.reduction=true;
set hive.llap.object.cache.enabled=false;
-EXPLAIN select count(*) from srcpart_small10, srcpart_small, srcpart_date where srcpart_small.key1 = srcpart_small10.key1 and srcpart_date.ds = srcpart_small.ds;
-select count(*) from srcpart_small10, srcpart_small, srcpart_date where srcpart_small.key1 = srcpart_small10.key1 and srcpart_date.ds = srcpart_small.ds;
+EXPLAIN select count(*) from srcpart_small10, srcpart_small_n3, srcpart_date_n7 where srcpart_small_n3.key1 = srcpart_small10.key1 and srcpart_date_n7.ds = srcpart_small_n3.ds;
+select count(*) from srcpart_small10, srcpart_small_n3, srcpart_date_n7 where srcpart_small_n3.key1 = srcpart_small10.key1 and srcpart_date_n7.ds = srcpart_small_n3.ds;
-- HIVE-17936
set hive.tez.dynamic.semijoin.reduction.for.dpp.factor = 0.75;
-EXPLAIN select count(*) from srcpart_small10, srcpart_small, srcpart_date where srcpart_small.key1 = srcpart_small10.key1 and srcpart_date.ds = srcpart_small.ds;
+EXPLAIN select count(*) from srcpart_small10, srcpart_small_n3, srcpart_date_n7 where srcpart_small_n3.key1 = srcpart_small10.key1 and srcpart_date_n7.ds = srcpart_small_n3.ds;
-- semijoin branch should be removed.
set hive.tez.dynamic.semijoin.reduction.for.dpp.factor = 0.4;
-EXPLAIN select count(*) from srcpart_small10, srcpart_small, srcpart_date where srcpart_small.key1 = srcpart_small10.key1 and srcpart_date.ds = srcpart_small.ds;
+EXPLAIN select count(*) from srcpart_small10, srcpart_small_n3, srcpart_date_n7 where srcpart_small_n3.key1 = srcpart_small10.key1 and srcpart_date_n7.ds = srcpart_small_n3.ds;
-- With unions
-explain select * from alltypesorc_int join
- (select srcpart_date.key as key from srcpart_date
+explain select * from alltypesorc_int_n1 join
+ (select srcpart_date_n7.key as key from srcpart_date_n7
union all
- select srcpart_small.key1 as key from srcpart_small) unionsrc on (alltypesorc_int.cstring = unionsrc.key);
+ select srcpart_small_n3.key1 as key from srcpart_small_n3) unionsrc on (alltypesorc_int_n1.cstring = unionsrc.key);
-drop table srcpart_date;
-drop table srcpart_small;
-drop table alltypesorc_int;
+drop table srcpart_date_n7;
+drop table srcpart_small_n3;
+drop table alltypesorc_int_n1;
http://git-wip-us.apache.org/repos/asf/hive/blob/38d3b8e1/ql/src/test/queries/clientpositive/dynamic_semijoin_reduction_2.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/dynamic_semijoin_reduction_2.q b/ql/src/test/queries/clientpositive/dynamic_semijoin_reduction_2.q
index 044ce5a..20b657e 100644
--- a/ql/src/test/queries/clientpositive/dynamic_semijoin_reduction_2.q
+++ b/ql/src/test/queries/clientpositive/dynamic_semijoin_reduction_2.q
@@ -12,14 +12,14 @@ set hive.tez.bigtable.minsize.semijoin.reduction=1;
set hive.tez.min.bloom.filter.entries=1;
set hive.tez.dynamic.semijoin.reduction.threshold=-999999999999;
-CREATE TABLE `table_1`(
+CREATE TABLE `table_1_n2`(
`bigint_col_7` bigint,
`decimal2016_col_26` decimal(20,16),
`tinyint_col_3` tinyint,
`decimal2612_col_77` decimal(26,12),
`timestamp_col_9` timestamp);
-CREATE TABLE `table_18`(
+CREATE TABLE `table_18_n2`(
`tinyint_col_15` tinyint,
`decimal2709_col_9` decimal(27,9),
`tinyint_col_20` tinyint,
@@ -31,29 +31,29 @@ CREATE TABLE `table_18`(
EXPLAIN
SELECT
COUNT(*)
-FROM table_1 t1
+FROM table_1_n2 t1
-INNER JOIN table_18 t2 ON (((t2.tinyint_col_15) = (t1.bigint_col_7)) AND
+INNER JOIN table_18_n2 t2 ON (((t2.tinyint_col_15) = (t1.bigint_col_7)) AND
((t2.decimal2709_col_9) = (t1.decimal2016_col_26))) AND
((t2.tinyint_col_20) = (t1.tinyint_col_3))
WHERE (t2.smallint_col_19) IN (SELECT
COALESCE(-92, -994) AS int_col
-FROM table_1 tt1
-INNER JOIN table_18 tt2 ON (tt2.decimal1911_col_16) = (tt1.decimal2612_col_77)
+FROM table_1_n2 tt1
+INNER JOIN table_18_n2 tt2 ON (tt2.decimal1911_col_16) = (tt1.decimal2612_col_77)
WHERE (t1.timestamp_col_9) = (tt2.timestamp_col_18));
-drop table table_1;
-drop table table_18;
+drop table table_1_n2;
+drop table table_18_n2;
-- Hive 15699
-CREATE TABLE srcbucket_mapjoin(key int, value string) partitioned by (ds string) CLUSTERED BY (key) INTO 2 BUCKETS STORED AS TEXTFILE;
+CREATE TABLE srcbucket_mapjoin_n20(key int, value string) partitioned by (ds string) CLUSTERED BY (key) INTO 2 BUCKETS STORED AS TEXTFILE;
-CREATE TABLE src2 as select * from src1;
-insert into src2 select * from src2;
-insert into src2 select * from src2;
+CREATE TABLE src2_n7 as select * from src1;
+insert into src2_n7 select * from src2_n7;
+insert into src2_n7 select * from src2_n7;
-load data local inpath '../../data/files/bmj/000000_0' INTO TABLE srcbucket_mapjoin partition(ds='2008-04-08');
-load data local inpath '../../data/files/bmj1/000001_0' INTO TABLE srcbucket_mapjoin partition(ds='2008-04-08');
+load data local inpath '../../data/files/bmj/000000_0' INTO TABLE srcbucket_mapjoin_n20 partition(ds='2008-04-08');
+load data local inpath '../../data/files/bmj1/000001_0' INTO TABLE srcbucket_mapjoin_n20 partition(ds='2008-04-08');
set hive.strict.checks.bucketing=false;
set hive.join.emit.interval=2;
@@ -61,9 +61,9 @@ set hive.stats.fetch.column.stats=true;
set hive.optimize.bucketingsorting=false;
set hive.stats.autogather=true;
-CREATE TABLE tab(key int, value string) PARTITIONED BY(ds STRING) CLUSTERED BY (key) SORTED BY (key) INTO 2 BUCKETS STORED AS TEXTFILE;
-insert overwrite table tab partition (ds='2008-04-08')
-select key,value from srcbucket_mapjoin;
+CREATE TABLE tab_n12(key int, value string) PARTITIONED BY(ds STRING) CLUSTERED BY (key) SORTED BY (key) INTO 2 BUCKETS STORED AS TEXTFILE;
+insert overwrite table tab_n12 partition (ds='2008-04-08')
+select key,value from srcbucket_mapjoin_n20;
set hive.convert.join.bucket.mapjoin.tez = true;
set hive.auto.convert.sortmerge.join = true;
@@ -74,20 +74,20 @@ set hive.mapjoin.hybridgrace.minnumpartitions=4;
set hive.llap.memory.oversubscription.max.executors.per.query=3;
-CREATE TABLE tab2 (key int, value string, ds string);
+CREATE TABLE tab2_n6 (key int, value string, ds string);
set hive.exec.dynamic.partition.mode=nonstrict
-insert into tab2select key, value, ds from tab;
-analyze table tab2 compute statistics;
-analyze table tab2 compute statistics for columns;
+insert into tab2select key, value, ds from tab_n12;
+analyze table tab2_n6 compute statistics;
+analyze table tab2_n6 compute statistics for columns;
explain
select
count(*)
from
- (select x.key as key, min(x.value) as value from tab2 x group by x.key) a
+ (select x.key as key, min(x.value) as value from tab2_n6 x group by x.key) a
join
- (select x.key as key, min(x.value) as value from tab2 x group by x.key) b
+ (select x.key as key, min(x.value) as value from tab2_n6 x group by x.key) b
on
a.key = b.key join src1 c on a.value = c.value where c.key < 0;
http://git-wip-us.apache.org/repos/asf/hive/blob/38d3b8e1/ql/src/test/queries/clientpositive/dynamic_semijoin_reduction_4.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/dynamic_semijoin_reduction_4.q b/ql/src/test/queries/clientpositive/dynamic_semijoin_reduction_4.q
index 67bf7c8..a04ab66 100644
--- a/ql/src/test/queries/clientpositive/dynamic_semijoin_reduction_4.q
+++ b/ql/src/test/queries/clientpositive/dynamic_semijoin_reduction_4.q
@@ -16,17 +16,17 @@ set hive.tez.bloom.filter.factor=1.0f;
set hive.disable.unsafe.external.table.operations=true;
-- Create Tables
-create table srcpart_date (key string, value string) partitioned by (ds string ) stored as ORC;
-CREATE TABLE srcpart_small(key1 STRING, value1 STRING) partitioned by (ds string) STORED as ORC;
+create table srcpart_date_n1 (key string, value string) partitioned by (ds string ) stored as ORC;
+CREATE TABLE srcpart_small_n0(key1 STRING, value1 STRING) partitioned by (ds string) STORED as ORC;
create external table srcpart_date_ext (key string, value string) partitioned by (ds string ) stored as ORC;
CREATE external TABLE srcpart_small_ext(key1 STRING, value1 STRING) partitioned by (ds string) STORED as ORC;
-- Add Partitions
-alter table srcpart_date add partition (ds = "2008-04-08");
-alter table srcpart_date add partition (ds = "2008-04-09");
+alter table srcpart_date_n1 add partition (ds = "2008-04-08");
+alter table srcpart_date_n1 add partition (ds = "2008-04-09");
-alter table srcpart_small add partition (ds = "2008-04-08");
-alter table srcpart_small add partition (ds = "2008-04-09");
+alter table srcpart_small_n0 add partition (ds = "2008-04-08");
+alter table srcpart_small_n0 add partition (ds = "2008-04-09");
alter table srcpart_date_ext add partition (ds = "2008-04-08");
alter table srcpart_date_ext add partition (ds = "2008-04-09");
@@ -35,16 +35,16 @@ alter table srcpart_small_ext add partition (ds = "2008-04-08");
alter table srcpart_small_ext add partition (ds = "2008-04-09");
-- Load
-insert overwrite table srcpart_date partition (ds = "2008-04-08" ) select key, value from srcpart where ds = "2008-04-08";
-insert overwrite table srcpart_date partition (ds = "2008-04-09") select key, value from srcpart where ds = "2008-04-09";
-insert overwrite table srcpart_small partition (ds = "2008-04-09") select key, value from srcpart where ds = "2008-04-09" limit 20;
+insert overwrite table srcpart_date_n1 partition (ds = "2008-04-08" ) select key, value from srcpart where ds = "2008-04-08";
+insert overwrite table srcpart_date_n1 partition (ds = "2008-04-09") select key, value from srcpart where ds = "2008-04-09";
+insert overwrite table srcpart_small_n0 partition (ds = "2008-04-09") select key, value from srcpart where ds = "2008-04-09" limit 20;
insert overwrite table srcpart_date_ext partition (ds = "2008-04-08" ) select key, value from srcpart where ds = "2008-04-08";
insert overwrite table srcpart_date_ext partition (ds = "2008-04-09") select key, value from srcpart where ds = "2008-04-09";
insert overwrite table srcpart_small_ext partition (ds = "2008-04-09") select key, value from srcpart where ds = "2008-04-09" limit 20;
-analyze table srcpart_date compute statistics for columns;
-analyze table srcpart_small compute statistics for columns;
+analyze table srcpart_date_n1 compute statistics for columns;
+analyze table srcpart_small_n0 compute statistics for columns;
analyze table srcpart_date_ext compute statistics for columns;
analyze table srcpart_small_ext compute statistics for columns;
@@ -53,13 +53,13 @@ analyze table srcpart_small_ext compute statistics for columns;
-- single column, single key
set test.comment=This query should use semijoin reduction optimization;
set test.comment;
-EXPLAIN select count(*) from srcpart_date join srcpart_small on (srcpart_date.key = srcpart_small.key1);
+EXPLAIN select count(*) from srcpart_date_n1 join srcpart_small_n0 on (srcpart_date_n1.key = srcpart_small_n0.key1);
set test.comment=Big table is external table - no semijoin reduction opt;
set test.comment;
-EXPLAIN select count(*) from srcpart_date_ext join srcpart_small on (srcpart_date_ext.key = srcpart_small.key1);
+EXPLAIN select count(*) from srcpart_date_ext join srcpart_small_n0 on (srcpart_date_ext.key = srcpart_small_n0.key1);
set test.comment=Small table is external table - no semijoin reduction opt;
set test.comment;
-EXPLAIN select count(*) from srcpart_date join srcpart_small_ext on (srcpart_date.key = srcpart_small_ext.key1);
+EXPLAIN select count(*) from srcpart_date_n1 join srcpart_small_ext on (srcpart_date_n1.key = srcpart_small_ext.key1);
http://git-wip-us.apache.org/repos/asf/hive/blob/38d3b8e1/ql/src/test/queries/clientpositive/dynamic_semijoin_reduction_sw.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/dynamic_semijoin_reduction_sw.q b/ql/src/test/queries/clientpositive/dynamic_semijoin_reduction_sw.q
index 3372cee..e4ca184 100644
--- a/ql/src/test/queries/clientpositive/dynamic_semijoin_reduction_sw.q
+++ b/ql/src/test/queries/clientpositive/dynamic_semijoin_reduction_sw.q
@@ -15,48 +15,48 @@ set hive.tez.min.bloom.filter.entries=1;
set hive.stats.fetch.column.stats=true;
-- Create Tables
-create table alltypesorc_int ( cint int, cstring string ) stored as ORC;
-create table srcpart_date (key string, value string) partitioned by (ds string ) stored as ORC;
-CREATE TABLE srcpart_small(key1 STRING, value1 STRING) partitioned by (ds1 string) STORED as ORC;
+create table alltypesorc_int_n0 ( cint int, cstring string ) stored as ORC;
+create table srcpart_date_n6 (key string, value string) partitioned by (ds string ) stored as ORC;
+CREATE TABLE srcpart_small_n2(key1 STRING, value1 STRING) partitioned by (ds1 string) STORED as ORC;
-- Add Partitions
-alter table srcpart_date add partition (ds = "2008-04-08");
-alter table srcpart_date add partition (ds = "2008-04-09");
+alter table srcpart_date_n6 add partition (ds = "2008-04-08");
+alter table srcpart_date_n6 add partition (ds = "2008-04-09");
-alter table srcpart_small add partition (ds1 = "2008-04-08");
-alter table srcpart_small add partition (ds1 = "2008-04-09");
+alter table srcpart_small_n2 add partition (ds1 = "2008-04-08");
+alter table srcpart_small_n2 add partition (ds1 = "2008-04-09");
-- Load
-insert overwrite table alltypesorc_int select cint, cstring1 from alltypesorc;
-insert overwrite table srcpart_date partition (ds = "2008-04-08" ) select key, value from srcpart where ds = "2008-04-08";
-insert overwrite table srcpart_date partition (ds = "2008-04-09") select key, value from srcpart where ds = "2008-04-09";
-insert overwrite table srcpart_small partition (ds1 = "2008-04-09") select key, value from srcpart where ds = "2008-04-09" limit 20;
+insert overwrite table alltypesorc_int_n0 select cint, cstring1 from alltypesorc;
+insert overwrite table srcpart_date_n6 partition (ds = "2008-04-08" ) select key, value from srcpart where ds = "2008-04-08";
+insert overwrite table srcpart_date_n6 partition (ds = "2008-04-09") select key, value from srcpart where ds = "2008-04-09";
+insert overwrite table srcpart_small_n2 partition (ds1 = "2008-04-09") select key, value from srcpart where ds = "2008-04-09" limit 20;
set hive.tez.dynamic.semijoin.reduction=false;
-analyze table alltypesorc_int compute statistics for columns;
-analyze table srcpart_date compute statistics for columns;
-analyze table srcpart_small compute statistics for columns;
+analyze table alltypesorc_int_n0 compute statistics for columns;
+analyze table srcpart_date_n6 compute statistics for columns;
+analyze table srcpart_small_n2 compute statistics for columns;
set hive.tez.dynamic.semijoin.reduction=true;
EXPLAIN
SELECT count(*)
FROM (
SELECT *
- FROM (SELECT * FROM srcpart_date WHERE ds = "2008-04-09") `srcpart_date`
- JOIN (SELECT * FROM srcpart_small WHERE ds1 = "2008-04-08") `srcpart_small`
- ON (srcpart_date.key = srcpart_small.key1)
- JOIN alltypesorc_int
- ON (srcpart_small.key1 = alltypesorc_int.cstring)) a
+ FROM (SELECT * FROM srcpart_date_n6 WHERE ds = "2008-04-09") `srcpart_date_n6`
+ JOIN (SELECT * FROM srcpart_small_n2 WHERE ds1 = "2008-04-08") `srcpart_small_n2`
+ ON (srcpart_date_n6.key = srcpart_small_n2.key1)
+ JOIN alltypesorc_int_n0
+ ON (srcpart_small_n2.key1 = alltypesorc_int_n0.cstring)) a
JOIN (
SELECT *
- FROM (SELECT * FROM srcpart_date WHERE ds = "2008-04-08") `srcpart_date`
- JOIN (SELECT * FROM srcpart_small WHERE ds1 = "2008-04-08") `srcpart_small`
- ON (srcpart_date.key = srcpart_small.key1)
- JOIN alltypesorc_int
- ON (srcpart_small.key1 = alltypesorc_int.cstring)) b
+ FROM (SELECT * FROM srcpart_date_n6 WHERE ds = "2008-04-08") `srcpart_date_n6`
+ JOIN (SELECT * FROM srcpart_small_n2 WHERE ds1 = "2008-04-08") `srcpart_small_n2`
+ ON (srcpart_date_n6.key = srcpart_small_n2.key1)
+ JOIN alltypesorc_int_n0
+ ON (srcpart_small_n2.key1 = alltypesorc_int_n0.cstring)) b
ON ('1' = '1');
-drop table srcpart_date;
-drop table srcpart_small;
-drop table alltypesorc_int;
+drop table srcpart_date_n6;
+drop table srcpart_small_n2;
+drop table alltypesorc_int_n0;
http://git-wip-us.apache.org/repos/asf/hive/blob/38d3b8e1/ql/src/test/queries/clientpositive/dynamic_semijoin_user_level.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/dynamic_semijoin_user_level.q b/ql/src/test/queries/clientpositive/dynamic_semijoin_user_level.q
index e9d0109..5e3a1b8 100644
--- a/ql/src/test/queries/clientpositive/dynamic_semijoin_user_level.q
+++ b/ql/src/test/queries/clientpositive/dynamic_semijoin_user_level.q
@@ -17,65 +17,65 @@ set hive.stats.fetch.column.stats=true;
set hive.llap.memory.oversubscription.max.executors.per.query=0;
-- Create Tables
-create table alltypesorc_int ( cint int, cstring string ) stored as ORC;
-create table srcpart_date (key string, value string) partitioned by (ds string ) stored as ORC;
-CREATE TABLE srcpart_small(key1 STRING, value1 STRING) partitioned by (ds string) STORED as ORC;
+create table alltypesorc_int_n2 ( cint int, cstring string ) stored as ORC;
+create table srcpart_date_n9 (key string, value string) partitioned by (ds string ) stored as ORC;
+CREATE TABLE srcpart_small_n4(key1 STRING, value1 STRING) partitioned by (ds string) STORED as ORC;
-- Add Partitions
-alter table srcpart_date add partition (ds = "2008-04-08");
-alter table srcpart_date add partition (ds = "2008-04-09");
+alter table srcpart_date_n9 add partition (ds = "2008-04-08");
+alter table srcpart_date_n9 add partition (ds = "2008-04-09");
-alter table srcpart_small add partition (ds = "2008-04-08");
-alter table srcpart_small add partition (ds = "2008-04-09");
+alter table srcpart_small_n4 add partition (ds = "2008-04-08");
+alter table srcpart_small_n4 add partition (ds = "2008-04-09");
-- Load
-insert overwrite table alltypesorc_int select cint, cstring1 from alltypesorc;
-insert overwrite table srcpart_date partition (ds = "2008-04-08" ) select key, value from srcpart where ds = "2008-04-08";
-insert overwrite table srcpart_date partition (ds = "2008-04-09") select key, value from srcpart where ds = "2008-04-09";
-insert overwrite table srcpart_small partition (ds = "2008-04-09") select key, value from srcpart where ds = "2008-04-09" limit 20;
+insert overwrite table alltypesorc_int_n2 select cint, cstring1 from alltypesorc;
+insert overwrite table srcpart_date_n9 partition (ds = "2008-04-08" ) select key, value from srcpart where ds = "2008-04-08";
+insert overwrite table srcpart_date_n9 partition (ds = "2008-04-09") select key, value from srcpart where ds = "2008-04-09";
+insert overwrite table srcpart_small_n4 partition (ds = "2008-04-09") select key, value from srcpart where ds = "2008-04-09" limit 20;
set hive.tez.dynamic.semijoin.reduction=false;
-analyze table alltypesorc_int compute statistics for columns;
-analyze table srcpart_date compute statistics for columns;
-analyze table srcpart_small compute statistics for columns;
+analyze table alltypesorc_int_n2 compute statistics for columns;
+analyze table srcpart_date_n9 compute statistics for columns;
+analyze table srcpart_small_n4 compute statistics for columns;
-- single column, single key
set hive.tez.dynamic.semijoin.reduction=true;
-EXPLAIN select count(*) from srcpart_date join srcpart_small on (srcpart_date.key = srcpart_small.key1);
-select count(*) from srcpart_date join srcpart_small on (srcpart_date.key = srcpart_small.key1);
+EXPLAIN select count(*) from srcpart_date_n9 join srcpart_small_n4 on (srcpart_date_n9.key = srcpart_small_n4.key1);
+select count(*) from srcpart_date_n9 join srcpart_small_n4 on (srcpart_date_n9.key = srcpart_small_n4.key1);
set hive.tez.dynamic.semijoin.reduction=true;
-- Mix dynamic partition pruning(DPP) and min/max bloom filter optimizations. Should pick the DPP.
-EXPLAIN select count(*) from srcpart_date join srcpart_small on (srcpart_date.key = srcpart_small.ds);
-select count(*) from srcpart_date join srcpart_small on (srcpart_date.key = srcpart_small.ds);
+EXPLAIN select count(*) from srcpart_date_n9 join srcpart_small_n4 on (srcpart_date_n9.key = srcpart_small_n4.ds);
+select count(*) from srcpart_date_n9 join srcpart_small_n4 on (srcpart_date_n9.key = srcpart_small_n4.ds);
set hive.tez.dynamic.semijoin.reduction=false;
--multiple sources, single key
-EXPLAIN select count(*) from srcpart_date join srcpart_small on (srcpart_date.key = srcpart_small.key1) join alltypesorc_int on (srcpart_small.key1 = alltypesorc_int.cstring);
-select count(*) from srcpart_date join srcpart_small on (srcpart_date.key = srcpart_small.key1) join alltypesorc_int on (srcpart_small.key1 = alltypesorc_int.cstring);
+EXPLAIN select count(*) from srcpart_date_n9 join srcpart_small_n4 on (srcpart_date_n9.key = srcpart_small_n4.key1) join alltypesorc_int_n2 on (srcpart_small_n4.key1 = alltypesorc_int_n2.cstring);
+select count(*) from srcpart_date_n9 join srcpart_small_n4 on (srcpart_date_n9.key = srcpart_small_n4.key1) join alltypesorc_int_n2 on (srcpart_small_n4.key1 = alltypesorc_int_n2.cstring);
set hive.tez.dynamic.semijoin.reduction=true;
-EXPLAIN select count(*) from srcpart_date join srcpart_small on (srcpart_date.key = srcpart_small.key1) join alltypesorc_int on (srcpart_small.key1 = alltypesorc_int.cstring);
-select count(*) from srcpart_date join srcpart_small on (srcpart_date.key = srcpart_small.key1) join alltypesorc_int on (srcpart_small.key1 = alltypesorc_int.cstring);
+EXPLAIN select count(*) from srcpart_date_n9 join srcpart_small_n4 on (srcpart_date_n9.key = srcpart_small_n4.key1) join alltypesorc_int_n2 on (srcpart_small_n4.key1 = alltypesorc_int_n2.cstring);
+select count(*) from srcpart_date_n9 join srcpart_small_n4 on (srcpart_date_n9.key = srcpart_small_n4.key1) join alltypesorc_int_n2 on (srcpart_small_n4.key1 = alltypesorc_int_n2.cstring);
set hive.tez.dynamic.semijoin.reduction=false;
-- single source, multiple keys
-EXPLAIN select count(*) from srcpart_date join srcpart_small on (srcpart_date.key = srcpart_small.key1 and srcpart_date.value = srcpart_small.value1);
-select count(*) from srcpart_date join srcpart_small on (srcpart_date.key = srcpart_small.key1 and srcpart_date.value = srcpart_small.value1);
+EXPLAIN select count(*) from srcpart_date_n9 join srcpart_small_n4 on (srcpart_date_n9.key = srcpart_small_n4.key1 and srcpart_date_n9.value = srcpart_small_n4.value1);
+select count(*) from srcpart_date_n9 join srcpart_small_n4 on (srcpart_date_n9.key = srcpart_small_n4.key1 and srcpart_date_n9.value = srcpart_small_n4.value1);
set hive.tez.dynamic.semijoin.reduction=true;
-EXPLAIN select count(*) from srcpart_date join srcpart_small on (srcpart_date.key = srcpart_small.key1 and srcpart_date.value = srcpart_small.value1);
-select count(*) from srcpart_date join srcpart_small on (srcpart_date.key = srcpart_small.key1 and srcpart_date.value = srcpart_small.value1);
+EXPLAIN select count(*) from srcpart_date_n9 join srcpart_small_n4 on (srcpart_date_n9.key = srcpart_small_n4.key1 and srcpart_date_n9.value = srcpart_small_n4.value1);
+select count(*) from srcpart_date_n9 join srcpart_small_n4 on (srcpart_date_n9.key = srcpart_small_n4.key1 and srcpart_date_n9.value = srcpart_small_n4.value1);
set hive.tez.dynamic.semijoin.reduction=false;
-- multiple sources, different keys
-EXPLAIN select count(*) from srcpart_date join srcpart_small on (srcpart_date.key = srcpart_small.key1) join alltypesorc_int on (srcpart_date.value = alltypesorc_int.cstring);
-select count(*) from srcpart_date join srcpart_small on (srcpart_date.key = srcpart_small.key1) join alltypesorc_int on (srcpart_date.value = alltypesorc_int.cstring);
+EXPLAIN select count(*) from srcpart_date_n9 join srcpart_small_n4 on (srcpart_date_n9.key = srcpart_small_n4.key1) join alltypesorc_int_n2 on (srcpart_date_n9.value = alltypesorc_int_n2.cstring);
+select count(*) from srcpart_date_n9 join srcpart_small_n4 on (srcpart_date_n9.key = srcpart_small_n4.key1) join alltypesorc_int_n2 on (srcpart_date_n9.value = alltypesorc_int_n2.cstring);
set hive.tez.dynamic.semijoin.reduction=true;
-EXPLAIN select count(*) from srcpart_date join srcpart_small on (srcpart_date.key = srcpart_small.key1) join alltypesorc_int on (srcpart_date.value = alltypesorc_int.cstring);
-select count(*) from srcpart_date join srcpart_small on (srcpart_date.key = srcpart_small.key1) join alltypesorc_int on (srcpart_date.value = alltypesorc_int.cstring);
+EXPLAIN select count(*) from srcpart_date_n9 join srcpart_small_n4 on (srcpart_date_n9.key = srcpart_small_n4.key1) join alltypesorc_int_n2 on (srcpart_date_n9.value = alltypesorc_int_n2.cstring);
+select count(*) from srcpart_date_n9 join srcpart_small_n4 on (srcpart_date_n9.key = srcpart_small_n4.key1) join alltypesorc_int_n2 on (srcpart_date_n9.value = alltypesorc_int_n2.cstring);
-- Explain extended to verify fast start for Reducer in semijoin branch
-EXPLAIN extended select count(*) from srcpart_date join srcpart_small on (srcpart_date.key = srcpart_small.key1);
+EXPLAIN extended select count(*) from srcpart_date_n9 join srcpart_small_n4 on (srcpart_date_n9.key = srcpart_small_n4.key1);
set hive.tez.dynamic.semijoin.reduction=false;
-- With Mapjoins.
@@ -83,28 +83,28 @@ set hive.auto.convert.join=true;
set hive.auto.convert.join.noconditionaltask=true;
set hive.auto.convert.join.noconditionaltask.size=100000000000;
-EXPLAIN select count(*) from srcpart_date join srcpart_small on (srcpart_date.key = srcpart_small.key1);
-select count(*) from srcpart_date join srcpart_small on (srcpart_date.key = srcpart_small.key1);
+EXPLAIN select count(*) from srcpart_date_n9 join srcpart_small_n4 on (srcpart_date_n9.key = srcpart_small_n4.key1);
+select count(*) from srcpart_date_n9 join srcpart_small_n4 on (srcpart_date_n9.key = srcpart_small_n4.key1);
set hive.tez.dynamic.semijoin.reduction=true;
-EXPLAIN select count(*) from srcpart_date join srcpart_small on (srcpart_date.key = srcpart_small.key1);
-select count(*) from srcpart_date join srcpart_small on (srcpart_date.key = srcpart_small.key1);
+EXPLAIN select count(*) from srcpart_date_n9 join srcpart_small_n4 on (srcpart_date_n9.key = srcpart_small_n4.key1);
+select count(*) from srcpart_date_n9 join srcpart_small_n4 on (srcpart_date_n9.key = srcpart_small_n4.key1);
set hive.tez.dynamic.semijoin.reduction=false;
-- multiple sources, different keys
-EXPLAIN select count(*) from srcpart_date join srcpart_small on (srcpart_date.key = srcpart_small.key1) join alltypesorc_int on (srcpart_date.value = alltypesorc_int.cstring);
-select count(*) from srcpart_date join srcpart_small on (srcpart_date.key = srcpart_small.key1) join alltypesorc_int on (srcpart_date.value = alltypesorc_int.cstring);
+EXPLAIN select count(*) from srcpart_date_n9 join srcpart_small_n4 on (srcpart_date_n9.key = srcpart_small_n4.key1) join alltypesorc_int_n2 on (srcpart_date_n9.value = alltypesorc_int_n2.cstring);
+select count(*) from srcpart_date_n9 join srcpart_small_n4 on (srcpart_date_n9.key = srcpart_small_n4.key1) join alltypesorc_int_n2 on (srcpart_date_n9.value = alltypesorc_int_n2.cstring);
set hive.tez.dynamic.semijoin.reduction=true;
-EXPLAIN select count(*) from srcpart_date join srcpart_small on (srcpart_date.key = srcpart_small.key1) join alltypesorc_int on (srcpart_date.value = alltypesorc_int.cstring);
-select count(*) from srcpart_date join srcpart_small on (srcpart_date.key = srcpart_small.key1) join alltypesorc_int on (srcpart_date.value = alltypesorc_int.cstring);
+EXPLAIN select count(*) from srcpart_date_n9 join srcpart_small_n4 on (srcpart_date_n9.key = srcpart_small_n4.key1) join alltypesorc_int_n2 on (srcpart_date_n9.value = alltypesorc_int_n2.cstring);
+select count(*) from srcpart_date_n9 join srcpart_small_n4 on (srcpart_date_n9.key = srcpart_small_n4.key1) join alltypesorc_int_n2 on (srcpart_date_n9.value = alltypesorc_int_n2.cstring);
--set hive.tez.dynamic.semijoin.reduction=false;
-- With unions
-explain select * from alltypesorc_int join
- (select srcpart_date.key as key from srcpart_date
+explain select * from alltypesorc_int_n2 join
+ (select srcpart_date_n9.key as key from srcpart_date_n9
union all
- select srcpart_small.key1 as key from srcpart_small) unionsrc on (alltypesorc_int.cstring = unionsrc.key);
+ select srcpart_small_n4.key1 as key from srcpart_small_n4) unionsrc on (alltypesorc_int_n2.cstring = unionsrc.key);
-drop table srcpart_date;
-drop table srcpart_small;
-drop table alltypesorc_int;
+drop table srcpart_date_n9;
+drop table srcpart_small_n4;
+drop table alltypesorc_int_n2;
http://git-wip-us.apache.org/repos/asf/hive/blob/38d3b8e1/ql/src/test/queries/clientpositive/dynpart_sort_opt_bucketing.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/dynpart_sort_opt_bucketing.q b/ql/src/test/queries/clientpositive/dynpart_sort_opt_bucketing.q
index af9853c..f2f36d1 100644
--- a/ql/src/test/queries/clientpositive/dynpart_sort_opt_bucketing.q
+++ b/ql/src/test/queries/clientpositive/dynpart_sort_opt_bucketing.q
@@ -19,9 +19,9 @@ set hive.exec.dynamic.partition.mode=nonstrict;
-drop table t1;
+drop table t1_n147;
-create table t1(
+create table t1_n147(
a string,
b int,
c int,
@@ -30,18 +30,18 @@ partitioned by (e string)
clustered by(a)
sorted by(a desc) into 10 buckets stored as textfile;
-insert overwrite table t1 partition(e) select a,b,c,d,'epart' from t1_staging;
+insert overwrite table t1_n147 partition(e) select a,b,c,d,'epart' from t1_staging;
select 'bucket_0';
-dfs -cat ${hiveconf:hive.metastore.warehouse.dir}/t1/e=epart/000000_0;
+dfs -cat ${hiveconf:hive.metastore.warehouse.dir}/t1_n147/e=epart/000000_0;
select 'bucket_2';
-dfs -cat ${hiveconf:hive.metastore.warehouse.dir}/t1/e=epart/000002_0;
+dfs -cat ${hiveconf:hive.metastore.warehouse.dir}/t1_n147/e=epart/000002_0;
select 'bucket_4';
-dfs -cat ${hiveconf:hive.metastore.warehouse.dir}/t1/e=epart/000004_0;
+dfs -cat ${hiveconf:hive.metastore.warehouse.dir}/t1_n147/e=epart/000004_0;
select 'bucket_6';
-dfs -cat ${hiveconf:hive.metastore.warehouse.dir}/t1/e=epart/000006_0;
+dfs -cat ${hiveconf:hive.metastore.warehouse.dir}/t1_n147/e=epart/000006_0;
select 'bucket_8';
-dfs -cat ${hiveconf:hive.metastore.warehouse.dir}/t1/e=epart/000008_0;
+dfs -cat ${hiveconf:hive.metastore.warehouse.dir}/t1_n147/e=epart/000008_0;
set hive.optimize.sort.dynamic.partition=false;
set hive.exec.dynamic.partition.mode=nonstrict;
@@ -49,9 +49,9 @@ set hive.exec.dynamic.partition.mode=nonstrict;
-- disable sorted dynamic partition optimization to make sure the results are correct
-drop table t1;
+drop table t1_n147;
-create table t1(
+create table t1_n147(
a string,
b int,
c int,
@@ -60,15 +60,15 @@ partitioned by (e string)
clustered by(a)
sorted by(a desc) into 10 buckets stored as textfile;
-insert overwrite table t1 partition(e) select a,b,c,d,'epart' from t1_staging;
+insert overwrite table t1_n147 partition(e) select a,b,c,d,'epart' from t1_staging;
select 'bucket_0';
-dfs -cat ${hiveconf:hive.metastore.warehouse.dir}/t1/e=epart/000000_0;
+dfs -cat ${hiveconf:hive.metastore.warehouse.dir}/t1_n147/e=epart/000000_0;
select 'bucket_2';
-dfs -cat ${hiveconf:hive.metastore.warehouse.dir}/t1/e=epart/000002_0;
+dfs -cat ${hiveconf:hive.metastore.warehouse.dir}/t1_n147/e=epart/000002_0;
select 'bucket_4';
-dfs -cat ${hiveconf:hive.metastore.warehouse.dir}/t1/e=epart/000004_0;
+dfs -cat ${hiveconf:hive.metastore.warehouse.dir}/t1_n147/e=epart/000004_0;
select 'bucket_6';
-dfs -cat ${hiveconf:hive.metastore.warehouse.dir}/t1/e=epart/000006_0;
+dfs -cat ${hiveconf:hive.metastore.warehouse.dir}/t1_n147/e=epart/000006_0;
select 'bucket_8';
-dfs -cat ${hiveconf:hive.metastore.warehouse.dir}/t1/e=epart/000008_0;
+dfs -cat ${hiveconf:hive.metastore.warehouse.dir}/t1_n147/e=epart/000008_0;
http://git-wip-us.apache.org/repos/asf/hive/blob/38d3b8e1/ql/src/test/queries/clientpositive/dynpart_sort_opt_vectorization.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/dynpart_sort_opt_vectorization.q b/ql/src/test/queries/clientpositive/dynpart_sort_opt_vectorization.q
index 6a90f83..435cdad 100644
--- a/ql/src/test/queries/clientpositive/dynpart_sort_opt_vectorization.q
+++ b/ql/src/test/queries/clientpositive/dynpart_sort_opt_vectorization.q
@@ -10,7 +10,7 @@ set hive.vectorized.execution.enabled=true;
-create table over1k(
+create table over1k_n1(
t tinyint,
si smallint,
i int,
@@ -25,11 +25,11 @@ create table over1k(
row format delimited
fields terminated by '|';
-load data local inpath '../../data/files/over1k' into table over1k;
+load data local inpath '../../data/files/over1k' into table over1k_n1;
-create table over1k_orc like over1k;
+create table over1k_orc like over1k_n1;
alter table over1k_orc set fileformat orc;
-insert overwrite table over1k_orc select * from over1k;
+insert overwrite table over1k_orc select * from over1k_n1;
create table over1k_part_orc(
si smallint,
http://git-wip-us.apache.org/repos/asf/hive/blob/38d3b8e1/ql/src/test/queries/clientpositive/dynpart_sort_optimization.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/dynpart_sort_optimization.q b/ql/src/test/queries/clientpositive/dynpart_sort_optimization.q
index 3f35b81..436c0ed 100644
--- a/ql/src/test/queries/clientpositive/dynpart_sort_optimization.q
+++ b/ql/src/test/queries/clientpositive/dynpart_sort_optimization.q
@@ -10,7 +10,7 @@ set hive.exec.dynamic.partition.mode=nonstrict;
-create table over1k(
+create table over1k_n3(
t tinyint,
si smallint,
i int,
@@ -25,7 +25,7 @@ create table over1k(
row format delimited
fields terminated by '|';
-load data local inpath '../../data/files/over1k' into table over1k;
+load data local inpath '../../data/files/over1k' into table over1k_n3;
create table over1k_part(
si smallint,
@@ -54,29 +54,29 @@ create table over1k_part_buck_sort(
sorted by (f) into 4 buckets;
-- map-only jobs converted to map-reduce job by hive.optimize.sort.dynamic.partition optimization
-explain insert overwrite table over1k_part partition(ds="foo", t) select si,i,b,f,t from over1k where t is null or t=27;
-explain insert overwrite table over1k_part_limit partition(ds="foo", t) select si,i,b,f,t from over1k where t is null or t=27 limit 10;
-explain insert overwrite table over1k_part_buck partition(t) select si,i,b,f,t from over1k where t is null or t=27;
-explain insert overwrite table over1k_part_buck_sort partition(t) select si,i,b,f,t from over1k where t is null or t=27;
+explain insert overwrite table over1k_part partition(ds="foo", t) select si,i,b,f,t from over1k_n3 where t is null or t=27;
+explain insert overwrite table over1k_part_limit partition(ds="foo", t) select si,i,b,f,t from over1k_n3 where t is null or t=27 limit 10;
+explain insert overwrite table over1k_part_buck partition(t) select si,i,b,f,t from over1k_n3 where t is null or t=27;
+explain insert overwrite table over1k_part_buck_sort partition(t) select si,i,b,f,t from over1k_n3 where t is null or t=27;
-insert overwrite table over1k_part partition(ds="foo", t) select si,i,b,f,t from over1k where t is null or t=27;
-insert overwrite table over1k_part_limit partition(ds="foo", t) select si,i,b,f,t from over1k where t is null or t=27 limit 10;
-insert overwrite table over1k_part_buck partition(t) select si,i,b,f,t from over1k where t is null or t=27;
-insert overwrite table over1k_part_buck_sort partition(t) select si,i,b,f,t from over1k where t is null or t=27;
+insert overwrite table over1k_part partition(ds="foo", t) select si,i,b,f,t from over1k_n3 where t is null or t=27;
+insert overwrite table over1k_part_limit partition(ds="foo", t) select si,i,b,f,t from over1k_n3 where t is null or t=27 limit 10;
+insert overwrite table over1k_part_buck partition(t) select si,i,b,f,t from over1k_n3 where t is null or t=27;
+insert overwrite table over1k_part_buck_sort partition(t) select si,i,b,f,t from over1k_n3 where t is null or t=27;
-- map-reduce jobs modified by hive.optimize.sort.dynamic.partition optimization
-explain insert into table over1k_part partition(ds="foo", t) select si,i,b,f,t from over1k where t is null or t=27;
-explain insert into table over1k_part_limit partition(ds="foo", t) select si,i,b,f,t from over1k where t is null or t=27 limit 10;
-explain insert into table over1k_part_buck partition(t) select si,i,b,f,t from over1k where t is null or t=27;
-explain insert into table over1k_part_buck_sort partition(t) select si,i,b,f,t from over1k where t is null or t=27;
+explain insert into table over1k_part partition(ds="foo", t) select si,i,b,f,t from over1k_n3 where t is null or t=27;
+explain insert into table over1k_part_limit partition(ds="foo", t) select si,i,b,f,t from over1k_n3 where t is null or t=27 limit 10;
+explain insert into table over1k_part_buck partition(t) select si,i,b,f,t from over1k_n3 where t is null or t=27;
+explain insert into table over1k_part_buck_sort partition(t) select si,i,b,f,t from over1k_n3 where t is null or t=27;
-insert into table over1k_part partition(ds="foo", t) select si,i,b,f,t from over1k where t is null or t=27;
-insert into table over1k_part_limit partition(ds="foo", t) select si,i,b,f,t from over1k where t is null or t=27 limit 10;
-insert into table over1k_part_buck partition(t) select si,i,b,f,t from over1k where t is null or t=27;
-insert into table over1k_part_buck_sort partition(t) select si,i,b,f,t from over1k where t is null or t=27;
+insert into table over1k_part partition(ds="foo", t) select si,i,b,f,t from over1k_n3 where t is null or t=27;
+insert into table over1k_part_limit partition(ds="foo", t) select si,i,b,f,t from over1k_n3 where t is null or t=27 limit 10;
+insert into table over1k_part_buck partition(t) select si,i,b,f,t from over1k_n3 where t is null or t=27;
+insert into table over1k_part_buck_sort partition(t) select si,i,b,f,t from over1k_n3 where t is null or t=27;
desc formatted over1k_part partition(ds="foo",t=27);
desc formatted over1k_part partition(ds="foo",t="__HIVE_DEFAULT_PARTITION__");
@@ -101,19 +101,19 @@ create table over1k_part2(
partitioned by (ds string, t tinyint);
set hive.optimize.sort.dynamic.partition=false;
-explain insert overwrite table over1k_part2 partition(ds="foo",t) select si,i,b,f,t from over1k where t is null or t=27 order by i;
+explain insert overwrite table over1k_part2 partition(ds="foo",t) select si,i,b,f,t from over1k_n3 where t is null or t=27 order by i;
set hive.optimize.sort.dynamic.partition=true;
-explain insert overwrite table over1k_part2 partition(ds="foo",t) select si,i,b,f,t from over1k where t is null or t=27 order by i;
-explain insert overwrite table over1k_part2 partition(ds="foo",t) select si,i,b,f,t from (select * from over1k order by i limit 10) tmp where t is null or t=27;
+explain insert overwrite table over1k_part2 partition(ds="foo",t) select si,i,b,f,t from over1k_n3 where t is null or t=27 order by i;
+explain insert overwrite table over1k_part2 partition(ds="foo",t) select si,i,b,f,t from (select * from over1k_n3 order by i limit 10) tmp where t is null or t=27;
set hive.optimize.sort.dynamic.partition=false;
-explain insert overwrite table over1k_part2 partition(ds="foo",t) select si,i,b,f,t from over1k where t is null or t=27 group by si,i,b,f,t;
+explain insert overwrite table over1k_part2 partition(ds="foo",t) select si,i,b,f,t from over1k_n3 where t is null or t=27 group by si,i,b,f,t;
set hive.optimize.sort.dynamic.partition=true;
-- tests for HIVE-8162, only partition column 't' should be in last RS operator
-explain insert overwrite table over1k_part2 partition(ds="foo",t) select si,i,b,f,t from over1k where t is null or t=27 group by si,i,b,f,t;
+explain insert overwrite table over1k_part2 partition(ds="foo",t) select si,i,b,f,t from over1k_n3 where t is null or t=27 group by si,i,b,f,t;
set hive.optimize.sort.dynamic.partition=false;
-insert overwrite table over1k_part2 partition(ds="foo",t) select si,i,b,f,t from over1k where t is null or t=27 order by i;
+insert overwrite table over1k_part2 partition(ds="foo",t) select si,i,b,f,t from over1k_n3 where t is null or t=27 order by i;
desc formatted over1k_part2 partition(ds="foo",t=27);
desc formatted over1k_part2 partition(ds="foo",t="__HIVE_DEFAULT_PARTITION__");
@@ -123,7 +123,7 @@ select * from over1k_part2;
select count(*) from over1k_part2;
set hive.optimize.sort.dynamic.partition=true;
-insert overwrite table over1k_part2 partition(ds="foo",t) select si,i,b,f,t from over1k where t is null or t=27 order by i;
+insert overwrite table over1k_part2 partition(ds="foo",t) select si,i,b,f,t from over1k_n3 where t is null or t=27 order by i;
desc formatted over1k_part2 partition(ds="foo",t=27);
desc formatted over1k_part2 partition(ds="foo",t="__HIVE_DEFAULT_PARTITION__");
@@ -145,12 +145,12 @@ create table over1k_part_buck_sort2(
sorted by (f) into 1 buckets;
set hive.optimize.sort.dynamic.partition=false;
-explain insert overwrite table over1k_part_buck_sort2 partition(t) select si,i,b,f,t from over1k where t is null or t=27;
+explain insert overwrite table over1k_part_buck_sort2 partition(t) select si,i,b,f,t from over1k_n3 where t is null or t=27;
set hive.optimize.sort.dynamic.partition=true;
-explain insert overwrite table over1k_part_buck_sort2 partition(t) select si,i,b,f,t from over1k where t is null or t=27;
+explain insert overwrite table over1k_part_buck_sort2 partition(t) select si,i,b,f,t from over1k_n3 where t is null or t=27;
set hive.optimize.sort.dynamic.partition=false;
-insert overwrite table over1k_part_buck_sort2 partition(t) select si,i,b,f,t from over1k where t is null or t=27;
+insert overwrite table over1k_part_buck_sort2 partition(t) select si,i,b,f,t from over1k_n3 where t is null or t=27;
desc formatted over1k_part_buck_sort2 partition(t=27);
desc formatted over1k_part_buck_sort2 partition(t="__HIVE_DEFAULT_PARTITION__");
@@ -159,7 +159,7 @@ select * from over1k_part_buck_sort2;
select count(*) from over1k_part_buck_sort2;
set hive.optimize.sort.dynamic.partition=true;
-insert overwrite table over1k_part_buck_sort2 partition(t) select si,i,b,f,t from over1k where t is null or t=27;
+insert overwrite table over1k_part_buck_sort2 partition(t) select si,i,b,f,t from over1k_n3 where t is null or t=27;
desc formatted over1k_part_buck_sort2 partition(t=27);
desc formatted over1k_part_buck_sort2 partition(t="__HIVE_DEFAULT_PARTITION__");
@@ -174,21 +174,21 @@ create table over1k_part3(
partitioned by (s string, t tinyint, i int);
set hive.optimize.sort.dynamic.partition=true;
-explain insert overwrite table over1k_part3 partition(s,t,i) select si,b,f,s,t,i from over1k where s="foo";
-explain insert overwrite table over1k_part3 partition(s,t,i) select si,b,f,s,t,i from over1k where t=27;
-explain insert overwrite table over1k_part3 partition(s,t,i) select si,b,f,s,t,i from over1k where i=100;
-explain insert overwrite table over1k_part3 partition(s,t,i) select si,b,f,s,t,i from over1k where i=100 and t=27;
-explain insert overwrite table over1k_part3 partition(s,t,i) select si,b,f,s,t,i from over1k where i=100 and s="foo";
-explain insert overwrite table over1k_part3 partition(s,t,i) select si,b,f,s,t,i from over1k where t=27 and s="foo";
-explain insert overwrite table over1k_part3 partition(s,t,i) select si,b,f,s,t,i from over1k where i=100 and t=27 and s="foo";
-
-insert overwrite table over1k_part3 partition(s,t,i) select si,b,f,s,t,i from over1k where s="foo";
-insert overwrite table over1k_part3 partition(s,t,i) select si,b,f,s,t,i from over1k where t=27;
-insert overwrite table over1k_part3 partition(s,t,i) select si,b,f,s,t,i from over1k where i=100;
-insert overwrite table over1k_part3 partition(s,t,i) select si,b,f,s,t,i from over1k where i=100 and t=27;
-insert overwrite table over1k_part3 partition(s,t,i) select si,b,f,s,t,i from over1k where i=100 and s="foo";
-insert overwrite table over1k_part3 partition(s,t,i) select si,b,f,s,t,i from over1k where t=27 and s="foo";
-insert overwrite table over1k_part3 partition(s,t,i) select si,b,f,s,t,i from over1k where i=100 and t=27 and s="foo";
+explain insert overwrite table over1k_part3 partition(s,t,i) select si,b,f,s,t,i from over1k_n3 where s="foo";
+explain insert overwrite table over1k_part3 partition(s,t,i) select si,b,f,s,t,i from over1k_n3 where t=27;
+explain insert overwrite table over1k_part3 partition(s,t,i) select si,b,f,s,t,i from over1k_n3 where i=100;
+explain insert overwrite table over1k_part3 partition(s,t,i) select si,b,f,s,t,i from over1k_n3 where i=100 and t=27;
+explain insert overwrite table over1k_part3 partition(s,t,i) select si,b,f,s,t,i from over1k_n3 where i=100 and s="foo";
+explain insert overwrite table over1k_part3 partition(s,t,i) select si,b,f,s,t,i from over1k_n3 where t=27 and s="foo";
+explain insert overwrite table over1k_part3 partition(s,t,i) select si,b,f,s,t,i from over1k_n3 where i=100 and t=27 and s="foo";
+
+insert overwrite table over1k_part3 partition(s,t,i) select si,b,f,s,t,i from over1k_n3 where s="foo";
+insert overwrite table over1k_part3 partition(s,t,i) select si,b,f,s,t,i from over1k_n3 where t=27;
+insert overwrite table over1k_part3 partition(s,t,i) select si,b,f,s,t,i from over1k_n3 where i=100;
+insert overwrite table over1k_part3 partition(s,t,i) select si,b,f,s,t,i from over1k_n3 where i=100 and t=27;
+insert overwrite table over1k_part3 partition(s,t,i) select si,b,f,s,t,i from over1k_n3 where i=100 and s="foo";
+insert overwrite table over1k_part3 partition(s,t,i) select si,b,f,s,t,i from over1k_n3 where t=27 and s="foo";
+insert overwrite table over1k_part3 partition(s,t,i) select si,b,f,s,t,i from over1k_n3 where i=100 and t=27 and s="foo";
select sum(hash(*)) from over1k_part3;
@@ -200,12 +200,12 @@ create table over1k_part3(
f float)
partitioned by (s string, t tinyint, i int);
set hive.optimize.sort.dynamic.partition=false;
-insert overwrite table over1k_part3 partition(s,t,i) select si,b,f,s,t,i from over1k where s="foo";
-insert overwrite table over1k_part3 partition(s,t,i) select si,b,f,s,t,i from over1k where t=27;
-insert overwrite table over1k_part3 partition(s,t,i) select si,b,f,s,t,i from over1k where i=100;
-insert overwrite table over1k_part3 partition(s,t,i) select si,b,f,s,t,i from over1k where i=100 and t=27;
-insert overwrite table over1k_part3 partition(s,t,i) select si,b,f,s,t,i from over1k where i=100 and s="foo";
-insert overwrite table over1k_part3 partition(s,t,i) select si,b,f,s,t,i from over1k where t=27 and s="foo";
-insert overwrite table over1k_part3 partition(s,t,i) select si,b,f,s,t,i from over1k where i=100 and t=27 and s="foo";
+insert overwrite table over1k_part3 partition(s,t,i) select si,b,f,s,t,i from over1k_n3 where s="foo";
+insert overwrite table over1k_part3 partition(s,t,i) select si,b,f,s,t,i from over1k_n3 where t=27;
+insert overwrite table over1k_part3 partition(s,t,i) select si,b,f,s,t,i from over1k_n3 where i=100;
+insert overwrite table over1k_part3 partition(s,t,i) select si,b,f,s,t,i from over1k_n3 where i=100 and t=27;
+insert overwrite table over1k_part3 partition(s,t,i) select si,b,f,s,t,i from over1k_n3 where i=100 and s="foo";
+insert overwrite table over1k_part3 partition(s,t,i) select si,b,f,s,t,i from over1k_n3 where t=27 and s="foo";
+insert overwrite table over1k_part3 partition(s,t,i) select si,b,f,s,t,i from over1k_n3 where i=100 and t=27 and s="foo";
select sum(hash(*)) from over1k_part3;
http://git-wip-us.apache.org/repos/asf/hive/blob/38d3b8e1/ql/src/test/queries/clientpositive/dynpart_sort_optimization2.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/dynpart_sort_optimization2.q b/ql/src/test/queries/clientpositive/dynpart_sort_optimization2.q
index 9a52fb4..6dfb51a 100644
--- a/ql/src/test/queries/clientpositive/dynpart_sort_optimization2.q
+++ b/ql/src/test/queries/clientpositive/dynpart_sort_optimization2.q
@@ -11,12 +11,12 @@ set hive.exec.dynamic.partition.mode=nonstrict;
-- SORT_QUERY_RESULTS
-drop table ss;
+drop table ss_n0;
drop table ss_orc;
drop table ss_part;
drop table ss_part_orc;
-create table ss (
+create table ss_n0 (
ss_sold_date_sk int,
ss_net_paid_inc_tax float,
ss_net_profit float);
@@ -26,13 +26,13 @@ ss_net_paid_inc_tax float,
ss_net_profit float)
partitioned by (ss_sold_date_sk int);
-load data local inpath '../../data/files/dynpart_test.txt' overwrite into table ss;
+load data local inpath '../../data/files/dynpart_test.txt' overwrite into table ss_n0;
explain insert overwrite table ss_part partition (ss_sold_date_sk)
select ss_net_paid_inc_tax,
ss_net_profit,
ss_sold_date_sk
- from ss
+ from ss_n0
where ss_sold_date_sk>=2452617 and ss_sold_date_sk<=2452638
group by ss_sold_date_sk,
ss_net_paid_inc_tax,
@@ -43,7 +43,7 @@ insert overwrite table ss_part partition (ss_sold_date_sk)
select ss_net_paid_inc_tax,
ss_net_profit,
ss_sold_date_sk
- from ss
+ from ss_n0
where ss_sold_date_sk>=2452617 and ss_sold_date_sk<=2452638
group by ss_sold_date_sk,
ss_net_paid_inc_tax,
@@ -60,7 +60,7 @@ explain insert overwrite table ss_part partition (ss_sold_date_sk)
select ss_net_paid_inc_tax,
ss_net_profit,
ss_sold_date_sk
- from ss
+ from ss_n0
where ss_sold_date_sk>=2452617 and ss_sold_date_sk<=2452638
distribute by ss_sold_date_sk;
@@ -68,7 +68,7 @@ insert overwrite table ss_part partition (ss_sold_date_sk)
select ss_net_paid_inc_tax,
ss_net_profit,
ss_sold_date_sk
- from ss
+ from ss_n0
where ss_sold_date_sk>=2452617 and ss_sold_date_sk<=2452638
distribute by ss_sold_date_sk;
@@ -85,7 +85,7 @@ explain insert overwrite table ss_part partition (ss_sold_date_sk)
select ss_net_paid_inc_tax,
ss_net_profit,
ss_sold_date_sk
- from ss
+ from ss_n0
where ss_sold_date_sk>=2452617 and ss_sold_date_sk<=2452638
group by ss_sold_date_sk,
ss_net_paid_inc_tax,
@@ -96,7 +96,7 @@ insert overwrite table ss_part partition (ss_sold_date_sk)
select ss_net_paid_inc_tax,
ss_net_profit,
ss_sold_date_sk
- from ss
+ from ss_n0
where ss_sold_date_sk>=2452617 and ss_sold_date_sk<=2452638
group by ss_sold_date_sk,
ss_net_paid_inc_tax,
@@ -113,7 +113,7 @@ explain insert overwrite table ss_part partition (ss_sold_date_sk)
select ss_net_paid_inc_tax,
ss_net_profit,
ss_sold_date_sk
- from ss
+ from ss_n0
where ss_sold_date_sk>=2452617 and ss_sold_date_sk<=2452638
distribute by ss_sold_date_sk;
@@ -121,7 +121,7 @@ insert overwrite table ss_part partition (ss_sold_date_sk)
select ss_net_paid_inc_tax,
ss_net_profit,
ss_sold_date_sk
- from ss
+ from ss_n0
where ss_sold_date_sk>=2452617 and ss_sold_date_sk<=2452638
distribute by ss_sold_date_sk;
@@ -144,9 +144,9 @@ ss_net_paid_inc_tax float,
ss_net_profit float)
partitioned by (ss_sold_date_sk int) stored as orc;
-insert overwrite table ss_orc select * from ss;
+insert overwrite table ss_orc select * from ss_n0;
-drop table ss;
+drop table ss_n0;
drop table ss_part;
explain insert overwrite table ss_part_orc partition (ss_sold_date_sk)
http://git-wip-us.apache.org/repos/asf/hive/blob/38d3b8e1/ql/src/test/queries/clientpositive/empty_join.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/empty_join.q b/ql/src/test/queries/clientpositive/empty_join.q
index 088b66a..cac2a6a 100644
--- a/ql/src/test/queries/clientpositive/empty_join.q
+++ b/ql/src/test/queries/clientpositive/empty_join.q
@@ -2,23 +2,23 @@ set hive.auto.convert.join=true;
set hive.mapjoin.hybridgrace.hashtable=false;
set hive.explain.user=true;
-DROP TABLE IF EXISTS test_1;
-CREATE TABLE test_1 AS SELECT 1 AS id;
+DROP TABLE IF EXISTS test_1_n2;
+CREATE TABLE test_1_n2 AS SELECT 1 AS id;
-DROP TABLE IF EXISTS test_2;
-CREATE TABLE test_2 (id INT);
+DROP TABLE IF EXISTS test_2_n2;
+CREATE TABLE test_2_n2 (id INT);
-DROP TABLE IF EXISTS test_3;
-CREATE TABLE test_3 AS SELECT 1 AS id;
+DROP TABLE IF EXISTS test_3_n0;
+CREATE TABLE test_3_n0 AS SELECT 1 AS id;
explain
SELECT t1.id, t2.id, t3.id
-FROM test_1 t1
-LEFT JOIN test_2 t2 ON t1.id = t2.id
-INNER JOIN test_3 t3 ON t1.id = t3.id;
+FROM test_1_n2 t1
+LEFT JOIN test_2_n2 t2 ON t1.id = t2.id
+INNER JOIN test_3_n0 t3 ON t1.id = t3.id;
SELECT t1.id, t2.id, t3.id
-FROM test_1 t1
-LEFT JOIN test_2 t2 ON t1.id = t2.id
-INNER JOIN test_3 t3 ON t1.id = t3.id
+FROM test_1_n2 t1
+LEFT JOIN test_2_n2 t2 ON t1.id = t2.id
+INNER JOIN test_3_n0 t3 ON t1.id = t3.id
;
http://git-wip-us.apache.org/repos/asf/hive/blob/38d3b8e1/ql/src/test/queries/clientpositive/encryption_auto_purge_tables.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/encryption_auto_purge_tables.q b/ql/src/test/queries/clientpositive/encryption_auto_purge_tables.q
index b095557..14c7f7e 100644
--- a/ql/src/test/queries/clientpositive/encryption_auto_purge_tables.q
+++ b/ql/src/test/queries/clientpositive/encryption_auto_purge_tables.q
@@ -3,37 +3,37 @@
-- we're setting this so that TestNegaiveCliDriver.vm doesn't stop processing after DROP TABLE fails;
-DROP TABLE IF EXISTS encrypted_table PURGE;
-DROP TABLE IF EXISTS encrypted_ext_table PURGE;
+DROP TABLE IF EXISTS encrypted_table_n5 PURGE;
+DROP TABLE IF EXISTS encrypted_ext_table_n0 PURGE;
-CREATE TABLE encrypted_table (key INT, value STRING) LOCATION '${hiveconf:hive.metastore.warehouse.dir}/default/encrypted_table';
+CREATE TABLE encrypted_table_n5 (key INT, value STRING) LOCATION '${hiveconf:hive.metastore.warehouse.dir}/default/encrypted_table';
CRYPTO CREATE_KEY --keyName key_128 --bitLength 128;
CRYPTO CREATE_ZONE --keyName key_128 --path ${hiveconf:hive.metastore.warehouse.dir}/default/encrypted_table;
SHOW TABLES LIKE "encrypted_*";
-ALTER TABLE encrypted_table SET TBLPROPERTIES("auto.purge"="true");
+ALTER TABLE encrypted_table_n5 SET TBLPROPERTIES("auto.purge"="true");
-INSERT OVERWRITE TABLE encrypted_table SELECT * FROM src;
-SELECT COUNT(*) from encrypted_table;
+INSERT OVERWRITE TABLE encrypted_table_n5 SELECT * FROM src;
+SELECT COUNT(*) from encrypted_table_n5;
-TRUNCATE TABLE encrypted_table;
-SELECT COUNT(*) FROM encrypted_table;
+TRUNCATE TABLE encrypted_table_n5;
+SELECT COUNT(*) FROM encrypted_table_n5;
-INSERT OVERWRITE TABLE encrypted_table SELECT * FROM src;
-SELECT COUNT(*) FROM encrypted_table;
+INSERT OVERWRITE TABLE encrypted_table_n5 SELECT * FROM src;
+SELECT COUNT(*) FROM encrypted_table_n5;
-CREATE EXTERNAL TABLE encrypted_ext_table (key INT, value STRING) LOCATION '${hiveconf:hive.metastore.warehouse.dir}/default/encrypted_table';
-ALTER TABLE encrypted_ext_table SET TBLPROPERTIES("auto.purge"="true");
+CREATE EXTERNAL TABLE encrypted_ext_table_n0 (key INT, value STRING) LOCATION '${hiveconf:hive.metastore.warehouse.dir}/default/encrypted_table';
+ALTER TABLE encrypted_ext_table_n0 SET TBLPROPERTIES("auto.purge"="true");
-INSERT OVERWRITE TABLE encrypted_ext_table SELECT * FROM src;
-SELECT COUNT(*) from encrypted_ext_table;
+INSERT OVERWRITE TABLE encrypted_ext_table_n0 SELECT * FROM src;
+SELECT COUNT(*) from encrypted_ext_table_n0;
-DROP TABLE encrypted_table;
-DROP TABLE encrypted_ext_table;
+DROP TABLE encrypted_table_n5;
+DROP TABLE encrypted_ext_table_n0;
SHOW TABLES LIKE "encrypted_*";
-- cleanup
-DROP TABLE IF EXISTS encrypted_table PURGE;
-DROP TABLE IF EXISTS encrypted_ext_table PURGE;
+DROP TABLE IF EXISTS encrypted_table_n5 PURGE;
+DROP TABLE IF EXISTS encrypted_ext_table_n0 PURGE;
CRYPTO DELETE_KEY --keyName key_128;
http://git-wip-us.apache.org/repos/asf/hive/blob/38d3b8e1/ql/src/test/queries/clientpositive/encryption_drop_table.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/encryption_drop_table.q b/ql/src/test/queries/clientpositive/encryption_drop_table.q
index 2212e7d..884e510 100644
--- a/ql/src/test/queries/clientpositive/encryption_drop_table.q
+++ b/ql/src/test/queries/clientpositive/encryption_drop_table.q
@@ -5,14 +5,14 @@
set hive.cli.errors.ignore=true;
-DROP TABLE IF EXISTS encrypted_table;
+DROP TABLE IF EXISTS encrypted_table_n2;
DROP TABLE IF EXISTS encrypted_ext_table;
-CREATE TABLE encrypted_table (key INT, value STRING) LOCATION '${hiveconf:hive.metastore.warehouse.dir}/default/encrypted_table';
+CREATE TABLE encrypted_table_n2 (key INT, value STRING) LOCATION '${hiveconf:hive.metastore.warehouse.dir}/default/encrypted_table';
CRYPTO CREATE_KEY --keyName key_128 --bitLength 128;
CRYPTO CREATE_ZONE --keyName key_128 --path ${hiveconf:hive.metastore.warehouse.dir}/default/encrypted_table;
-INSERT OVERWRITE TABLE encrypted_table SELECT * FROM src;
+INSERT OVERWRITE TABLE encrypted_table_n2 SELECT * FROM src;
CREATE EXTERNAL TABLE encrypted_ext_table (key INT, value STRING) LOCATION '${hiveconf:hive.metastore.warehouse.dir}/default/encrypted_table';
SHOW TABLES LIKE "encrypted_*";
@@ -20,7 +20,7 @@ SHOW TABLES LIKE "encrypted_*";
DROP TABLE default.encrypted_ext_table;
SHOW TABLES LIKE "encrypted_*";
-DROP TABLE default.encrypted_table;
+DROP TABLE default.encrypted_table_n2;
SHOW TABLES LIKE "encrypted_*";
DROP TABLE IF EXISTS encrypted_table1;
http://git-wip-us.apache.org/repos/asf/hive/blob/38d3b8e1/ql/src/test/queries/clientpositive/encryption_insert_partition_dynamic.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/encryption_insert_partition_dynamic.q b/ql/src/test/queries/clientpositive/encryption_insert_partition_dynamic.q
index 6cfb382..b5ea78b 100644
--- a/ql/src/test/queries/clientpositive/encryption_insert_partition_dynamic.q
+++ b/ql/src/test/queries/clientpositive/encryption_insert_partition_dynamic.q
@@ -8,40 +8,40 @@ set hive.txn.manager=org.apache.hadoop.hive.ql.lockmgr.DbTxnManager;
-- SORT_QUERY_RESULTS
-- init
-drop table IF EXISTS encryptedTable PURGE;
-drop table IF EXISTS unencryptedTable PURGE;
+drop table IF EXISTS encryptedTable_n0 PURGE;
+drop table IF EXISTS unencryptedTable_n0 PURGE;
-create table encryptedTable(value string)
+create table encryptedTable_n0(value string)
partitioned by (key string) clustered by (value) into 2 buckets stored as orc
LOCATION '${hiveconf:hive.metastore.warehouse.dir}/encryptedTable' TBLPROPERTIES ('transactional'='true');
CRYPTO CREATE_KEY --keyName key_1 --bitLength 128;
CRYPTO CREATE_ZONE --keyName key_1 --path ${hiveconf:hive.metastore.warehouse.dir}/encryptedTable;
-create table unencryptedTable(value string)
+create table unencryptedTable_n0(value string)
partitioned by (key string) clustered by (value) into 2 buckets stored as orc TBLPROPERTIES ('transactional'='true');
-- insert encrypted table from values
-insert into table encryptedTable partition (key) values
+insert into table encryptedTable_n0 partition (key) values
('val_501', '501'),
('val_502', '502');
-select * from encryptedTable order by key;
+select * from encryptedTable_n0 order by key;
-- insert encrypted table from unencrypted source
from src
-insert into table encryptedTable partition (key)
+insert into table encryptedTable_n0 partition (key)
select value, key limit 2;
-select * from encryptedTable order by key;
+select * from encryptedTable_n0 order by key;
-- insert unencrypted table from encrypted source
-from encryptedTable
-insert into table unencryptedTable partition (key)
+from encryptedTable_n0
+insert into table unencryptedTable_n0 partition (key)
select value, key;
-select * from unencryptedTable order by key;
+select * from unencryptedTable_n0 order by key;
-- clean up
-drop table encryptedTable PURGE;
+drop table encryptedTable_n0 PURGE;
CRYPTO DELETE_KEY --keyName key_1;
-drop table unencryptedTable PURGE;
+drop table unencryptedTable_n0 PURGE;
http://git-wip-us.apache.org/repos/asf/hive/blob/38d3b8e1/ql/src/test/queries/clientpositive/encryption_insert_values.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/encryption_insert_values.q b/ql/src/test/queries/clientpositive/encryption_insert_values.q
index a712f35..242515e 100644
--- a/ql/src/test/queries/clientpositive/encryption_insert_values.q
+++ b/ql/src/test/queries/clientpositive/encryption_insert_values.q
@@ -1,13 +1,13 @@
-- SORT_QUERY_RESULTS;
set hive.stats.column.autogather=false;
-DROP TABLE IF EXISTS encrypted_table PURGE;
-CREATE TABLE encrypted_table (key INT, value STRING) LOCATION '${hiveconf:hive.metastore.warehouse.dir}/default/encrypted_table';
+DROP TABLE IF EXISTS encrypted_table_n3 PURGE;
+CREATE TABLE encrypted_table_n3 (key INT, value STRING) LOCATION '${hiveconf:hive.metastore.warehouse.dir}/default/encrypted_table';
CRYPTO CREATE_KEY --keyName key_128 --bitLength 128;
CRYPTO CREATE_ZONE --keyName key_128 --path ${hiveconf:hive.metastore.warehouse.dir}/default/encrypted_table;
-INSERT INTO encrypted_table values(1,'foo'),(2,'bar');
+INSERT INTO encrypted_table_n3 values(1,'foo'),(2,'bar');
-select * from encrypted_table;
+select * from encrypted_table_n3;
CRYPTO DELETE_KEY --keyName key_128;
http://git-wip-us.apache.org/repos/asf/hive/blob/38d3b8e1/ql/src/test/queries/clientpositive/encryption_load_data_to_encrypted_tables.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/encryption_load_data_to_encrypted_tables.q b/ql/src/test/queries/clientpositive/encryption_load_data_to_encrypted_tables.q
index a96807e..834bfbc 100644
--- a/ql/src/test/queries/clientpositive/encryption_load_data_to_encrypted_tables.q
+++ b/ql/src/test/queries/clientpositive/encryption_load_data_to_encrypted_tables.q
@@ -1,20 +1,20 @@
-DROP TABLE IF EXISTS encrypted_table PURGE;
+DROP TABLE IF EXISTS encrypted_table_n0 PURGE;
-CREATE TABLE encrypted_table (key STRING, value STRING) LOCATION '${hiveconf:hive.metastore.warehouse.dir}/encrypted_table';
+CREATE TABLE encrypted_table_n0 (key STRING, value STRING) LOCATION '${hiveconf:hive.metastore.warehouse.dir}/encrypted_table';
-- Create encryption key and zone;
crypto create_key --keyName key1;
crypto create_zone --keyName key1 --path ${hiveconf:hive.metastore.warehouse.dir}/encrypted_table;
-- Test loading data from the local filesystem;
-LOAD DATA LOCAL INPATH '../../data/files/kv1.txt' OVERWRITE INTO TABLE encrypted_table;
-SELECT * FROM encrypted_table;
+LOAD DATA LOCAL INPATH '../../data/files/kv1.txt' OVERWRITE INTO TABLE encrypted_table_n0;
+SELECT * FROM encrypted_table_n0;
-- Test loading data from the hdfs filesystem;
dfs -copyFromLocal ../../data/files/kv1.txt hdfs:///tmp/kv1.txt;
-LOAD DATA INPATH '/tmp/kv1.txt' OVERWRITE INTO TABLE encrypted_table;
-SELECT * FROM encrypted_table;
+LOAD DATA INPATH '/tmp/kv1.txt' OVERWRITE INTO TABLE encrypted_table_n0;
+SELECT * FROM encrypted_table_n0;
-DROP TABLE encrypted_table PURGE;
+DROP TABLE encrypted_table_n0 PURGE;
crypto delete_key --keyName key1;
\ No newline at end of file
http://git-wip-us.apache.org/repos/asf/hive/blob/38d3b8e1/ql/src/test/queries/clientpositive/encryption_move_tbl.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/encryption_move_tbl.q b/ql/src/test/queries/clientpositive/encryption_move_tbl.q
index ff1d57c..b636076 100644
--- a/ql/src/test/queries/clientpositive/encryption_move_tbl.q
+++ b/ql/src/test/queries/clientpositive/encryption_move_tbl.q
@@ -6,11 +6,11 @@ set hive.stats.column.autogather=false;
set hive.cli.errors.ignore=true;
-DROP TABLE IF EXISTS encrypted_table PURGE;
+DROP TABLE IF EXISTS encrypted_table_n1 PURGE;
DROP DATABASE IF EXISTS encrypted_db;
--- create table default.encrypted_table in its default warehouse location ${hiveconf:hive.metastore.warehouse.dir}/encrypted_table
-CREATE TABLE encrypted_table (key INT, value STRING) LOCATION '${hiveconf:hive.metastore.warehouse.dir}/encrypted_table';
+-- create table default.encrypted_table_n1 in its default warehouse location ${hiveconf:hive.metastore.warehouse.dir}/encrypted_table
+CREATE TABLE encrypted_table_n1 (key INT, value STRING) LOCATION '${hiveconf:hive.metastore.warehouse.dir}/encrypted_table';
CRYPTO CREATE_KEY --keyName key_128 --bitLength 128;
CRYPTO CREATE_ZONE --keyName key_128 --path ${hiveconf:hive.metastore.warehouse.dir}/encrypted_table;
@@ -19,20 +19,20 @@ CREATE DATABASE encrypted_db LOCATION '${hiveconf:hive.metastore.warehouse.dir}/
CRYPTO CREATE_KEY --keyName key_128_2 --bitLength 128;
CRYPTO CREATE_ZONE --keyName key_128_2 --path ${hiveconf:hive.metastore.warehouse.dir}/encrypted_db.db;
-INSERT OVERWRITE TABLE encrypted_table SELECT * FROM src;
+INSERT OVERWRITE TABLE encrypted_table_n1 SELECT * FROM src;
SHOW TABLES LIKE "encrypted_*";
-ANALYZE TABLE encrypted_table COMPUTE STATISTICS FOR COLUMNS;
-DESCRIBE FORMATTED encrypted_table key;
-DESCRIBE FORMATTED encrypted_table value;
+ANALYZE TABLE encrypted_table_n1 COMPUTE STATISTICS FOR COLUMNS;
+DESCRIBE FORMATTED encrypted_table_n1 key;
+DESCRIBE FORMATTED encrypted_table_n1 value;
-- should fail, since they are in different encryption zones, but table columns statistics should not change
-ALTER TABLE default.encrypted_table RENAME TO encrypted_db.encrypted_table_2;
+ALTER TABLE default.encrypted_table_n1 RENAME TO encrypted_db.encrypted_table_2;
SHOW TABLES;
-DESCRIBE FORMATTED encrypted_table key;
-DESCRIBE FORMATTED encrypted_table value;
+DESCRIBE FORMATTED encrypted_table_n1 key;
+DESCRIBE FORMATTED encrypted_table_n1 value;
-- should succeed in Hadoop 2.7 but fail in 2.6 (HDFS-7530)
-ALTER TABLE default.encrypted_table RENAME TO default.plain_table;
+ALTER TABLE default.encrypted_table_n1 RENAME TO default.plain_table;
SHOW TABLES;
-- create table encrypted_table_outloc under default database but in a specified location other than the default db location in the warehouse
@@ -49,20 +49,20 @@ CRYPTO CREATE_KEY --keyName key_128_4 --bitLength 128;
CRYPTO CREATE_ZONE --keyName key_128_4 --path ${hiveconf:hive.metastore.warehouse.dir}/../specified_db_location;
USE encrypted_db_outloc;
-CREATE TABLE encrypted_table (key INT, value STRING);
-INSERT OVERWRITE TABLE encrypted_table SELECT * FROM default.src;
-ALTER TABLE encrypted_table RENAME TO renamed_encrypted_table;
+CREATE TABLE encrypted_table_n1 (key INT, value STRING);
+INSERT OVERWRITE TABLE encrypted_table_n1 SELECT * FROM default.src;
+ALTER TABLE encrypted_table_n1 RENAME TO renamed_encrypted_table_n1;
-- should succeed since data moves within specified_db_location
SHOW TABLES;
-- should fail, since they are in different encryption zones
-ALTER TABLE encrypted_db_outloc.renamed_encrypted_table RENAME TO default.plain_table_2;
+ALTER TABLE encrypted_db_outloc.renamed_encrypted_table_n1 RENAME TO default.plain_table_2;
SHOW TABLES;
-DROP TABLE default.encrypted_table PURGE;
+DROP TABLE default.encrypted_table_n1 PURGE;
DROP TABLE default.plain_table PURGE;
DROP TABLE default.renamed_encrypted_table_outloc PURGE;
DROP DATABASE encrypted_db;
-DROP TABLE encrypted_db_outloc.renamed_encrypted_table PURGE;
+DROP TABLE encrypted_db_outloc.renamed_encrypted_table_n1 PURGE;
DROP DATABASE encrypted_db_outloc;
CRYPTO DELETE_KEY --keyName key_128;
CRYPTO DELETE_KEY --keyName key_128_2;
http://git-wip-us.apache.org/repos/asf/hive/blob/38d3b8e1/ql/src/test/queries/clientpositive/encryption_select_read_only_encrypted_tbl.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/encryption_select_read_only_encrypted_tbl.q b/ql/src/test/queries/clientpositive/encryption_select_read_only_encrypted_tbl.q
index c9ac035..94052a5 100644
--- a/ql/src/test/queries/clientpositive/encryption_select_read_only_encrypted_tbl.q
+++ b/ql/src/test/queries/clientpositive/encryption_select_read_only_encrypted_tbl.q
@@ -1,16 +1,16 @@
-- SORT_QUERY_RESULTS
-DROP TABLE IF EXISTS encrypted_table PURGE;
-CREATE TABLE encrypted_table (key INT, value STRING) LOCATION '${hiveconf:hive.metastore.warehouse.dir}/default/encrypted_table';
+DROP TABLE IF EXISTS encrypted_table_n4 PURGE;
+CREATE TABLE encrypted_table_n4 (key INT, value STRING) LOCATION '${hiveconf:hive.metastore.warehouse.dir}/default/encrypted_table';
CRYPTO CREATE_KEY --keyName key_128 --bitLength 128;
CRYPTO CREATE_ZONE --keyName key_128 --path ${hiveconf:hive.metastore.warehouse.dir}/default/encrypted_table;
-LOAD DATA LOCAL INPATH '../../data/files/kv1.txt' INTO TABLE encrypted_table;
+LOAD DATA LOCAL INPATH '../../data/files/kv1.txt' INTO TABLE encrypted_table_n4;
dfs -chmod -R 555 ${hiveconf:hive.metastore.warehouse.dir}/default/encrypted_table;
-SELECT count(*) FROM encrypted_table;
+SELECT count(*) FROM encrypted_table_n4;
-drop table encrypted_table PURGE;
+drop table encrypted_table_n4 PURGE;
CRYPTO DELETE_KEY --keyName key_128;