You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hive.apache.org by jc...@apache.org on 2018/05/21 20:25:01 UTC
[24/51] [abbrv] [partial] hive git commit: HIVE-19617: Rename test
tables to avoid collisions during execution in batches (Jesus Camacho
Rodriguez, reviewed by Gunther Hagleitner)
http://git-wip-us.apache.org/repos/asf/hive/blob/38d3b8e1/ql/src/test/queries/clientpositive/metadata_only_queries_with_filters.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/metadata_only_queries_with_filters.q b/ql/src/test/queries/clientpositive/metadata_only_queries_with_filters.q
index 692c414..ed75c57 100644
--- a/ql/src/test/queries/clientpositive/metadata_only_queries_with_filters.q
+++ b/ql/src/test/queries/clientpositive/metadata_only_queries_with_filters.q
@@ -2,7 +2,7 @@ set hive.stats.column.autogather=false;
set hive.stats.dbclass=fs;
set hive.compute.query.using.stats=true;
set hive.explain.user=false;
-create table over10k(
+create table over10k_n23(
t tinyint,
si smallint,
i int,
@@ -17,9 +17,9 @@ create table over10k(
row format delimited
fields terminated by '|';
-load data local inpath '../../data/files/over10k' into table over10k;
+load data local inpath '../../data/files/over10k' into table over10k_n23;
-create table stats_tbl_part(
+create table stats_tbl_part_n0(
t tinyint,
si smallint,
i int,
@@ -33,22 +33,22 @@ create table stats_tbl_part(
bin binary) partitioned by (dt int);
-from over10k
-insert overwrite table stats_tbl_part partition (dt=2010) select t,si,i,b,f,d,bo,s,ts,`dec`,bin where t>0 and t<30
-insert overwrite table stats_tbl_part partition (dt=2014) select t,si,i,b,f,d,bo,s,ts,`dec`,bin where t > 30 and t<60;
+from over10k_n23
+insert overwrite table stats_tbl_part_n0 partition (dt=2010) select t,si,i,b,f,d,bo,s,ts,`dec`,bin where t>0 and t<30
+insert overwrite table stats_tbl_part_n0 partition (dt=2014) select t,si,i,b,f,d,bo,s,ts,`dec`,bin where t > 30 and t<60;
-analyze table stats_tbl_part partition(dt) compute statistics;
-analyze table stats_tbl_part partition(dt=2010) compute statistics for columns t,si,i,b,f,d,bo,s,bin;
-analyze table stats_tbl_part partition(dt=2014) compute statistics for columns t,si,i,b,f,d,bo,s,bin;
+analyze table stats_tbl_part_n0 partition(dt) compute statistics;
+analyze table stats_tbl_part_n0 partition(dt=2010) compute statistics for columns t,si,i,b,f,d,bo,s,bin;
+analyze table stats_tbl_part_n0 partition(dt=2014) compute statistics for columns t,si,i,b,f,d,bo,s,bin;
explain
-select count(*), count(1), sum(1), count(s), count(bo), count(bin), count(si), max(i), min(b), max(f), min(d) from stats_tbl_part where dt = 2010;
-select count(*), count(1), sum(1), count(s), count(bo), count(bin), count(si), max(i), min(b), max(f), min(d) from stats_tbl_part where dt = 2010;
+select count(*), count(1), sum(1), count(s), count(bo), count(bin), count(si), max(i), min(b), max(f), min(d) from stats_tbl_part_n0 where dt = 2010;
+select count(*), count(1), sum(1), count(s), count(bo), count(bin), count(si), max(i), min(b), max(f), min(d) from stats_tbl_part_n0 where dt = 2010;
explain
-select count(*), count(1), sum(1), sum(2), count(s), count(bo), count(bin), count(si), max(i), min(b), max(f), min(d) from stats_tbl_part where dt > 2010;
-select count(*), count(1), sum(1), sum(2), count(s), count(bo), count(bin), count(si), max(i), min(b), max(f), min(d) from stats_tbl_part where dt > 2010;
+select count(*), count(1), sum(1), sum(2), count(s), count(bo), count(bin), count(si), max(i), min(b), max(f), min(d) from stats_tbl_part_n0 where dt > 2010;
+select count(*), count(1), sum(1), sum(2), count(s), count(bo), count(bin), count(si), max(i), min(b), max(f), min(d) from stats_tbl_part_n0 where dt > 2010;
-select count(*) from stats_tbl_part;
-select count(*)/2 from stats_tbl_part;
-drop table stats_tbl_part;
+select count(*) from stats_tbl_part_n0;
+select count(*)/2 from stats_tbl_part_n0;
+drop table stats_tbl_part_n0;
set hive.compute.query.using.stats=false;
http://git-wip-us.apache.org/repos/asf/hive/blob/38d3b8e1/ql/src/test/queries/clientpositive/metadataonly1.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/metadataonly1.q b/ql/src/test/queries/clientpositive/metadataonly1.q
index e3bf819..e247c6a 100644
--- a/ql/src/test/queries/clientpositive/metadataonly1.q
+++ b/ql/src/test/queries/clientpositive/metadataonly1.q
@@ -1,48 +1,48 @@
--! qt:dataset:srcpart
set hive.mapred.mode=nonstrict;
set hive.optimize.metadataonly=true;
-CREATE TABLE TEST1(A INT, B DOUBLE) partitioned by (ds string);
-explain extended select max(ds) from TEST1;
-select max(ds) from TEST1;
+CREATE TABLE TEST1_n12(A INT, B DOUBLE) partitioned by (ds string);
+explain extended select max(ds) from TEST1_n12;
+select max(ds) from TEST1_n12;
-alter table TEST1 add partition (ds='1');
-explain extended select max(ds) from TEST1;
-select max(ds) from TEST1;
+alter table TEST1_n12 add partition (ds='1');
+explain extended select max(ds) from TEST1_n12;
+select max(ds) from TEST1_n12;
-explain extended select count(distinct ds) from TEST1;
-select count(distinct ds) from TEST1;
+explain extended select count(distinct ds) from TEST1_n12;
+select count(distinct ds) from TEST1_n12;
-explain extended select count(ds) from TEST1;
-select count(ds) from TEST1;
+explain extended select count(ds) from TEST1_n12;
+select count(ds) from TEST1_n12;
-alter table TEST1 add partition (ds='2');
+alter table TEST1_n12 add partition (ds='2');
explain extended
-select count(*) from TEST1 a2 join (select max(ds) m from TEST1) b on a2.ds=b.m;
-select count(*) from TEST1 a2 join (select max(ds) m from TEST1) b on a2.ds=b.m;
+select count(*) from TEST1_n12 a2 join (select max(ds) m from TEST1_n12) b on a2.ds=b.m;
+select count(*) from TEST1_n12 a2 join (select max(ds) m from TEST1_n12) b on a2.ds=b.m;
-CREATE TABLE TEST2(A INT, B DOUBLE) partitioned by (ds string, hr string);
-alter table TEST2 add partition (ds='1', hr='1');
-alter table TEST2 add partition (ds='1', hr='2');
-alter table TEST2 add partition (ds='1', hr='3');
+CREATE TABLE TEST2_n8(A INT, B DOUBLE) partitioned by (ds string, hr string);
+alter table TEST2_n8 add partition (ds='1', hr='1');
+alter table TEST2_n8 add partition (ds='1', hr='2');
+alter table TEST2_n8 add partition (ds='1', hr='3');
-explain extended select ds, count(distinct hr) from TEST2 group by ds;
-select ds, count(distinct hr) from TEST2 group by ds;
+explain extended select ds, count(distinct hr) from TEST2_n8 group by ds;
+select ds, count(distinct hr) from TEST2_n8 group by ds;
-explain extended select ds, count(hr) from TEST2 group by ds;
-select ds, count(hr) from TEST2 group by ds;
+explain extended select ds, count(hr) from TEST2_n8 group by ds;
+select ds, count(hr) from TEST2_n8 group by ds;
set hive.input.format=org.apache.hadoop.hive.ql.io.CombineHiveInputFormat;
-explain extended select max(ds) from TEST1;
-select max(ds) from TEST1;
+explain extended select max(ds) from TEST1_n12;
+select max(ds) from TEST1_n12;
select distinct ds from srcpart;
select min(ds),max(ds) from srcpart;
-- HIVE-3594 URI encoding for temporary path
-alter table TEST2 add partition (ds='01:10:10', hr='01');
-alter table TEST2 add partition (ds='01:10:20', hr='02');
+alter table TEST2_n8 add partition (ds='01:10:10', hr='01');
+alter table TEST2_n8 add partition (ds='01:10:20', hr='02');
-explain extended select ds, count(distinct hr) from TEST2 group by ds;
-select ds, count(distinct hr) from TEST2 group by ds;
+explain extended select ds, count(distinct hr) from TEST2_n8 group by ds;
+select ds, count(distinct hr) from TEST2_n8 group by ds;
http://git-wip-us.apache.org/repos/asf/hive/blob/38d3b8e1/ql/src/test/queries/clientpositive/misc_json.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/misc_json.q b/ql/src/test/queries/clientpositive/misc_json.q
index 541e369..95a3e5b 100644
--- a/ql/src/test/queries/clientpositive/misc_json.q
+++ b/ql/src/test/queries/clientpositive/misc_json.q
@@ -1,10 +1,10 @@
set hive.ddl.output.format=json;
-CREATE TABLE IF NOT EXISTS jsontable (key INT, value STRING) COMMENT 'json table' STORED AS TEXTFILE;
+CREATE TABLE IF NOT EXISTS jsontable_n0 (key INT, value STRING) COMMENT 'json table' STORED AS TEXTFILE;
-ALTER TABLE jsontable ADD COLUMNS (name STRING COMMENT 'a new column');
+ALTER TABLE jsontable_n0 ADD COLUMNS (name STRING COMMENT 'a new column');
-ALTER TABLE jsontable RENAME TO jsontable2;
+ALTER TABLE jsontable_n0 RENAME TO jsontable2;
SHOW TABLE EXTENDED LIKE jsontable2;
http://git-wip-us.apache.org/repos/asf/hive/blob/38d3b8e1/ql/src/test/queries/clientpositive/mm_all.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/mm_all.q b/ql/src/test/queries/clientpositive/mm_all.q
index 7d0955b..61dd3e7 100644
--- a/ql/src/test/queries/clientpositive/mm_all.q
+++ b/ql/src/test/queries/clientpositive/mm_all.q
@@ -13,32 +13,32 @@ set hive.support.concurrency=true;
set hive.txn.manager=org.apache.hadoop.hive.ql.lockmgr.DbTxnManager;
-- Force multiple writers when reading
-drop table intermediate;
-create table intermediate(key int) partitioned by (p int) stored as orc;
-insert into table intermediate partition(p='455') select distinct key from src where key >= 0 order by key desc limit 2;
-insert into table intermediate partition(p='456') select distinct key from src where key is not null order by key asc limit 2;
-insert into table intermediate partition(p='457') select distinct key from src where key >= 100 order by key asc limit 2;
-
-
-drop table part_mm;
-create table part_mm(key int) partitioned by (key_mm int) stored as orc tblproperties ("transactional"="true", "transactional_properties"="insert_only");
-explain insert into table part_mm partition(key_mm=455) select key from intermediate;
-insert into table part_mm partition(key_mm=455) select key from intermediate;
-insert into table part_mm partition(key_mm=456) select key from intermediate;
-insert into table part_mm partition(key_mm=455) select key from intermediate;
-select * from part_mm order by key, key_mm;
-
--- TODO: doesn't work truncate table part_mm partition(key_mm=455);
-select * from part_mm order by key, key_mm;
-truncate table part_mm;
-select * from part_mm order by key, key_mm;
-drop table part_mm;
+drop table intermediate_n0;
+create table intermediate_n0(key int) partitioned by (p int) stored as orc;
+insert into table intermediate_n0 partition(p='455') select distinct key from src where key >= 0 order by key desc limit 2;
+insert into table intermediate_n0 partition(p='456') select distinct key from src where key is not null order by key asc limit 2;
+insert into table intermediate_n0 partition(p='457') select distinct key from src where key >= 100 order by key asc limit 2;
+
+
+drop table part_mm_n0;
+create table part_mm_n0(key int) partitioned by (key_mm int) stored as orc tblproperties ("transactional"="true", "transactional_properties"="insert_only");
+explain insert into table part_mm_n0 partition(key_mm=455) select key from intermediate_n0;
+insert into table part_mm_n0 partition(key_mm=455) select key from intermediate_n0;
+insert into table part_mm_n0 partition(key_mm=456) select key from intermediate_n0;
+insert into table part_mm_n0 partition(key_mm=455) select key from intermediate_n0;
+select * from part_mm_n0 order by key, key_mm;
+
+-- TODO: doesn't work truncate table part_mm_n0 partition(key_mm=455);
+select * from part_mm_n0 order by key, key_mm;
+truncate table part_mm_n0;
+select * from part_mm_n0 order by key, key_mm;
+drop table part_mm_n0;
drop table simple_mm;
create table simple_mm(key int) stored as orc tblproperties ("transactional"="true", "transactional_properties"="insert_only");
-insert into table simple_mm select key from intermediate;
+insert into table simple_mm select key from intermediate_n0;
select * from simple_mm order by key;
-insert into table simple_mm select key from intermediate;
+insert into table simple_mm select key from intermediate_n0;
select * from simple_mm order by key;
truncate table simple_mm;
select * from simple_mm;
@@ -57,7 +57,7 @@ set hive.merge.tezfiles=false;
create table dp_mm (key int) partitioned by (key1 string, key2 int) stored as orc
tblproperties ("transactional"="true", "transactional_properties"="insert_only");
-insert into table dp_mm partition (key1='123', key2) select key, key from intermediate;
+insert into table dp_mm partition (key1='123', key2) select key, key from intermediate_n0;
select * from dp_mm order by key;
@@ -69,35 +69,35 @@ drop table dp_mm;
create table union_mm(id int) tblproperties ("transactional"="true", "transactional_properties"="insert_only");
insert into table union_mm
select temps.p from (
-select key as p from intermediate
+select key as p from intermediate_n0
union all
-select key + 1 as p from intermediate ) temps;
+select key + 1 as p from intermediate_n0 ) temps;
select * from union_mm order by id;
insert into table union_mm
select p from
(
-select key + 1 as p from intermediate
+select key + 1 as p from intermediate_n0
union all
-select key from intermediate
+select key from intermediate_n0
) tab group by p
union all
-select key + 2 as p from intermediate;
+select key + 2 as p from intermediate_n0;
select * from union_mm order by id;
insert into table union_mm
SELECT p FROM
(
- SELECT key + 1 as p FROM intermediate
+ SELECT key + 1 as p FROM intermediate_n0
UNION ALL
SELECT key as p FROM (
SELECT distinct key FROM (
SELECT key FROM (
- SELECT key + 2 as key FROM intermediate
+ SELECT key + 2 as key FROM intermediate_n0
UNION ALL
- SELECT key FROM intermediate
+ SELECT key FROM intermediate_n0
)t1
group by key)t2
)t3
@@ -112,9 +112,9 @@ drop table union_mm;
create table partunion_mm(id int) partitioned by (key int) tblproperties ("transactional"="true", "transactional_properties"="insert_only");
insert into table partunion_mm partition(key)
select temps.* from (
-select key as p, key from intermediate
+select key as p, key from intermediate_n0
union all
-select key + 1 as p, key + 1 from intermediate ) temps;
+select key + 1 as p, key + 1 from intermediate_n0 ) temps;
select * from partunion_mm order by id;
drop table partunion_mm;
@@ -127,7 +127,7 @@ create table skew_mm(k1 int, k2 int, k4 int) skewed by (k1, k4) on ((0,0),(1,1),
stored as directories tblproperties ("transactional"="true", "transactional_properties"="insert_only");
insert into table skew_mm
-select key, key, key from intermediate;
+select key, key, key from intermediate_n0;
select * from skew_mm order by k2, k1, k4;
drop table skew_mm;
@@ -137,9 +137,9 @@ create table skew_dp_union_mm(k1 int, k2 int, k4 int) partitioned by (k3 int)
skewed by (k1, k4) on ((0,0),(1,1),(2,2),(3,3)) stored as directories tblproperties ("transactional"="true", "transactional_properties"="insert_only");
insert into table skew_dp_union_mm partition (k3)
-select key as i, key as j, key as k, key as l from intermediate
+select key as i, key as j, key as k, key as l from intermediate_n0
union all
-select key +1 as i, key +2 as j, key +3 as k, key +4 as l from intermediate;
+select key +1 as i, key +2 as j, key +3 as k, key +4 as l from intermediate_n0;
select * from skew_dp_union_mm order by k2, k1, k4;
@@ -155,11 +155,11 @@ set hive.merge.mapredfiles=true;
create table merge0_mm (id int) stored as orc tblproperties("transactional"="true", "transactional_properties"="insert_only");
-insert into table merge0_mm select key from intermediate;
+insert into table merge0_mm select key from intermediate_n0;
select * from merge0_mm;
set tez.grouping.split-count=1;
-insert into table merge0_mm select key from intermediate;
+insert into table merge0_mm select key from intermediate_n0;
set tez.grouping.split-count=0;
select * from merge0_mm;
@@ -168,11 +168,11 @@ drop table merge0_mm;
create table merge2_mm (id int) tblproperties("transactional"="true", "transactional_properties"="insert_only");
-insert into table merge2_mm select key from intermediate;
+insert into table merge2_mm select key from intermediate_n0;
select * from merge2_mm;
set tez.grouping.split-count=1;
-insert into table merge2_mm select key from intermediate;
+insert into table merge2_mm select key from intermediate_n0;
set tez.grouping.split-count=0;
select * from merge2_mm;
@@ -181,11 +181,11 @@ drop table merge2_mm;
create table merge1_mm (id int) partitioned by (key int) stored as orc tblproperties("transactional"="true", "transactional_properties"="insert_only");
-insert into table merge1_mm partition (key) select key, key from intermediate;
+insert into table merge1_mm partition (key) select key, key from intermediate_n0;
select * from merge1_mm order by id, key;
set tez.grouping.split-count=1;
-insert into table merge1_mm partition (key) select key, key from intermediate;
+insert into table merge1_mm partition (key) select key, key from intermediate_n0;
set tez.grouping.split-count=0;
select * from merge1_mm order by id, key;
@@ -199,13 +199,13 @@ set hive.merge.mapredfiles=false;
drop table ctas0_mm;
-create table ctas0_mm tblproperties ("transactional"="true", "transactional_properties"="insert_only") as select * from intermediate;
+create table ctas0_mm tblproperties ("transactional"="true", "transactional_properties"="insert_only") as select * from intermediate_n0;
select * from ctas0_mm;
drop table ctas0_mm;
drop table ctas1_mm;
create table ctas1_mm tblproperties ("transactional"="true", "transactional_properties"="insert_only") as
- select * from intermediate union all select * from intermediate;
+ select * from intermediate_n0 union all select * from intermediate_n0;
select * from ctas1_mm;
drop table ctas1_mm;
@@ -215,7 +215,7 @@ drop table multi0_2_mm;
create table multi0_1_mm (key int, key2 int) tblproperties("transactional"="true", "transactional_properties"="insert_only");
create table multi0_2_mm (key int, key2 int) tblproperties("transactional"="true", "transactional_properties"="insert_only");
-from intermediate
+from intermediate_n0
insert overwrite table multi0_1_mm select key, p
insert overwrite table multi0_2_mm select p, key;
@@ -226,7 +226,7 @@ set hive.merge.mapredfiles=true;
set hive.merge.sparkfiles=true;
set hive.merge.tezfiles=true;
-from intermediate
+from intermediate_n0
insert into table multi0_1_mm select p, key
insert overwrite table multi0_2_mm select key, p;
@@ -243,25 +243,25 @@ drop table multi0_2_mm;
drop table multi1_mm;
create table multi1_mm (key int, key2 int) partitioned by (p int) tblproperties("transactional"="true", "transactional_properties"="insert_only");
-from intermediate
+from intermediate_n0
insert into table multi1_mm partition(p=1) select p, key
insert into table multi1_mm partition(p=2) select key, p;
select * from multi1_mm order by key, key2, p;
-from intermediate
+from intermediate_n0
insert into table multi1_mm partition(p=2) select p, key
insert overwrite table multi1_mm partition(p=1) select key, p;
select * from multi1_mm order by key, key2, p;
-from intermediate
+from intermediate_n0
insert into table multi1_mm partition(p) select p, key, p
insert into table multi1_mm partition(p=1) select key, p;
select key, key2, p from multi1_mm order by key, key2, p;
-from intermediate
+from intermediate_n0
insert into table multi1_mm partition(p) select p, key, 1
insert into table multi1_mm partition(p=1) select key, p;
@@ -277,11 +277,11 @@ set hive.stats.autogather=true;
drop table stats_mm;
create table stats_mm(key int) tblproperties("transactional"="true", "transactional_properties"="insert_only");
---insert overwrite table stats_mm select key from intermediate;
-insert into table stats_mm select key from intermediate;
+--insert overwrite table stats_mm select key from intermediate_n0;
+insert into table stats_mm select key from intermediate_n0;
desc formatted stats_mm;
-insert into table stats_mm select key from intermediate;
+insert into table stats_mm select key from intermediate_n0;
desc formatted stats_mm;
drop table stats_mm;
@@ -321,12 +321,12 @@ set hive.auto.convert.join=true;
DROP TABLE IF EXISTS temp1;
CREATE TEMPORARY TABLE temp1 (a int) TBLPROPERTIES ("transactional"="true", "transactional_properties"="insert_only");
-INSERT INTO temp1 SELECT key FROM intermediate;
+INSERT INTO temp1 SELECT key FROM intermediate_n0;
DESC EXTENDED temp1;
SELECT * FROM temp1;
-drop table intermediate;
+drop table intermediate_n0;
http://git-wip-us.apache.org/repos/asf/hive/blob/38d3b8e1/ql/src/test/queries/clientpositive/mm_buckets.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/mm_buckets.q b/ql/src/test/queries/clientpositive/mm_buckets.q
index 3e91cdd..12dc4fe 100644
--- a/ql/src/test/queries/clientpositive/mm_buckets.q
+++ b/ql/src/test/queries/clientpositive/mm_buckets.q
@@ -12,11 +12,11 @@ set hive.txn.manager=org.apache.hadoop.hive.ql.lockmgr.DbTxnManager;
-- Bucketing tests are slow and some tablesample ones don't work w/o MM
-- Force multiple writers when reading
-drop table intermediate;
-create table intermediate(key int) partitioned by (p int) stored as orc;
-insert into table intermediate partition(p='455') select distinct key from src where key >= 0 order by key desc limit 2;
-insert into table intermediate partition(p='456') select distinct key from src where key is not null order by key asc limit 2;
-insert into table intermediate partition(p='457') select distinct key from src where key >= 100 order by key asc limit 2;
+drop table intermediate_n2;
+create table intermediate_n2(key int) partitioned by (p int) stored as orc;
+insert into table intermediate_n2 partition(p='455') select distinct key from src where key >= 0 order by key desc limit 2;
+insert into table intermediate_n2 partition(p='456') select distinct key from src where key is not null order by key asc limit 2;
+insert into table intermediate_n2 partition(p='457') select distinct key from src where key >= 100 order by key asc limit 2;
@@ -24,11 +24,11 @@ drop table bucket0_mm;
create table bucket0_mm(key int, id int)
clustered by (key) into 2 buckets
tblproperties("transactional"="true", "transactional_properties"="insert_only");
-insert into table bucket0_mm select key, key from intermediate;
+insert into table bucket0_mm select key, key from intermediate_n2;
select * from bucket0_mm order by key, id;
select * from bucket0_mm tablesample (bucket 1 out of 2) s;
select * from bucket0_mm tablesample (bucket 2 out of 2) s;
-insert into table bucket0_mm select key, key from intermediate;
+insert into table bucket0_mm select key, key from intermediate_n2;
select * from bucket0_mm order by key, id;
select * from bucket0_mm tablesample (bucket 1 out of 2) s;
select * from bucket0_mm tablesample (bucket 2 out of 2) s;
@@ -40,9 +40,9 @@ create table bucket1_mm(key int, id int) partitioned by (key2 int)
clustered by (key) sorted by (key) into 2 buckets
tblproperties("transactional"="true", "transactional_properties"="insert_only");
insert into table bucket1_mm partition (key2)
-select key + 1, key, key - 1 from intermediate
+select key + 1, key, key - 1 from intermediate_n2
union all
-select key - 1, key, key + 1 from intermediate;
+select key - 1, key, key + 1 from intermediate_n2;
select * from bucket1_mm order by key, id;
select * from bucket1_mm tablesample (bucket 1 out of 2) s order by key, id;
select * from bucket1_mm tablesample (bucket 2 out of 2) s order by key, id;
@@ -54,14 +54,14 @@ drop table bucket2_mm;
create table bucket2_mm(key int, id int)
clustered by (key) into 10 buckets
tblproperties("transactional"="true", "transactional_properties"="insert_only");
-insert into table bucket2_mm select key, key from intermediate where key == 0;
+insert into table bucket2_mm select key, key from intermediate_n2 where key == 0;
select * from bucket2_mm order by key, id;
select * from bucket2_mm tablesample (bucket 1 out of 10) s order by key, id;
select * from bucket2_mm tablesample (bucket 4 out of 10) s order by key, id;
-insert into table bucket2_mm select key, key from intermediate where key in (0, 103);
+insert into table bucket2_mm select key, key from intermediate_n2 where key in (0, 103);
select * from bucket2_mm;
select * from bucket2_mm tablesample (bucket 1 out of 10) s order by key, id;
select * from bucket2_mm tablesample (bucket 4 out of 10) s order by key, id;
drop table bucket2_mm;
-drop table intermediate;
\ No newline at end of file
+drop table intermediate_n2;
\ No newline at end of file
http://git-wip-us.apache.org/repos/asf/hive/blob/38d3b8e1/ql/src/test/queries/clientpositive/mm_cttas.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/mm_cttas.q b/ql/src/test/queries/clientpositive/mm_cttas.q
index da8b84d..8f1274c 100644
--- a/ql/src/test/queries/clientpositive/mm_cttas.q
+++ b/ql/src/test/queries/clientpositive/mm_cttas.q
@@ -3,19 +3,19 @@
set hive.support.concurrency=true;
set hive.txn.manager=org.apache.hadoop.hive.ql.lockmgr.DbTxnManager;
-drop table intermediate;
-create table intermediate(key int) partitioned by (p int) stored as orc;
-insert into table intermediate partition(p='455') select distinct key from src where key >= 0 order by key desc limit 2;
-insert into table intermediate partition(p='456') select distinct key from src where key is not null order by key asc limit 2;
-insert into table intermediate partition(p='457') select distinct key from src where key >= 100 order by key asc limit 2;
+drop table intermediate_n1;
+create table intermediate_n1(key int) partitioned by (p int) stored as orc;
+insert into table intermediate_n1 partition(p='455') select distinct key from src where key >= 0 order by key desc limit 2;
+insert into table intermediate_n1 partition(p='456') select distinct key from src where key is not null order by key asc limit 2;
+insert into table intermediate_n1 partition(p='457') select distinct key from src where key >= 100 order by key asc limit 2;
drop table cttas1_mm;
-create temporary table cttas1_mm tblproperties ("transactional"="true", "transactional_properties"="insert_only") as select * from intermediate;
+create temporary table cttas1_mm tblproperties ("transactional"="true", "transactional_properties"="insert_only") as select * from intermediate_n1;
select * from cttas1_mm;
drop table cttas1_mm;
-drop table intermediate;
+drop table intermediate_n1;
http://git-wip-us.apache.org/repos/asf/hive/blob/38d3b8e1/ql/src/test/queries/clientpositive/mm_default.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/mm_default.q b/ql/src/test/queries/clientpositive/mm_default.q
index 0fa46a9..8e34bf3 100644
--- a/ql/src/test/queries/clientpositive/mm_default.q
+++ b/ql/src/test/queries/clientpositive/mm_default.q
@@ -14,10 +14,10 @@ create table acid0 (key string) stored as ORC tblproperties("transactional"="tr
set hive.create.as.insert.only=true;
create table mm1 like non_mm0;
create table mm2 like mm0;
-create table acid1 like acid0;
+create table acid1_n0 like acid0;
create table mm3 as select key from src limit 1;
create table mm4 (key string);
-create table acid2 (key string) stored as ORC tblproperties("transactional"="true");
+create table acid2_n0 (key string) stored as ORC tblproperties("transactional"="true");
create table non_mm1 tblproperties("transactional"="false")
as select key from src limit 1;
@@ -28,8 +28,8 @@ desc formatted mm2;
desc formatted mm3;
desc formatted mm4;
desc formatted non_mm1;
-desc formatted acid1;
-desc formatted acid2;
+desc formatted acid1_n0;
+desc formatted acid2_n0;
drop table non_mm0;
@@ -40,8 +40,8 @@ drop table mm2;
drop table mm3;
drop table mm4;
drop table acid0;
-drop table acid1;
-drop table acid2;
+drop table acid1_n0;
+drop table acid2_n0;
http://git-wip-us.apache.org/repos/asf/hive/blob/38d3b8e1/ql/src/test/queries/clientpositive/mm_exim.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/mm_exim.q b/ql/src/test/queries/clientpositive/mm_exim.q
index a2b6e08..9870bf4 100644
--- a/ql/src/test/queries/clientpositive/mm_exim.q
+++ b/ql/src/test/queries/clientpositive/mm_exim.q
@@ -10,27 +10,27 @@ set hive.support.concurrency=true;
set hive.txn.manager=org.apache.hadoop.hive.ql.lockmgr.DbTxnManager;
-drop table intermediate;
-create table intermediate(key int) partitioned by (p int) stored as orc tblproperties("transactional"="false");
-insert into table intermediate partition(p='455') select distinct key from src where key >= 0 order by key desc limit 2;
-insert into table intermediate partition(p='456') select distinct key from src where key is not null order by key asc limit 2;
-insert into table intermediate partition(p='457') select distinct key from src where key >= 100 order by key asc limit 2;
+drop table intermediate_n4;
+create table intermediate_n4(key int) partitioned by (p int) stored as orc tblproperties("transactional"="false");
+insert into table intermediate_n4 partition(p='455') select distinct key from src where key >= 0 order by key desc limit 2;
+insert into table intermediate_n4 partition(p='456') select distinct key from src where key is not null order by key asc limit 2;
+insert into table intermediate_n4 partition(p='457') select distinct key from src where key >= 100 order by key asc limit 2;
drop table intermediate_nonpart;
drop table intermmediate_part;
drop table intermmediate_nonpart;
create table intermediate_nonpart(key int, p int) tblproperties("transactional"="false");
-insert into intermediate_nonpart select * from intermediate;
+insert into intermediate_nonpart select * from intermediate_n4;
create table intermmediate_nonpart(key int, p int) tblproperties("transactional"="true", "transactional_properties"="insert_only");
-insert into intermmediate_nonpart select * from intermediate;
+insert into intermmediate_nonpart select * from intermediate_n4;
create table intermmediate(key int) partitioned by (p int) tblproperties("transactional"="true", "transactional_properties"="insert_only");
-insert into table intermmediate partition(p) select key, p from intermediate;
+insert into table intermmediate partition(p) select key, p from intermediate_n4;
set hive.exim.test.mode=true;
export table intermediate_nonpart to 'ql/test/data/exports/intermediate_nonpart';
export table intermmediate_nonpart to 'ql/test/data/exports/intermmediate_nonpart';
-export table intermediate to 'ql/test/data/exports/intermediate_part';
+export table intermediate_n4 to 'ql/test/data/exports/intermediate_part';
export table intermmediate to 'ql/test/data/exports/intermmediate_part';
drop table intermediate_nonpart;
http://git-wip-us.apache.org/repos/asf/hive/blob/38d3b8e1/ql/src/test/queries/clientpositive/msck_repair_0.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/msck_repair_0.q b/ql/src/test/queries/clientpositive/msck_repair_0.q
index cb291fe..aeb4820 100644
--- a/ql/src/test/queries/clientpositive/msck_repair_0.q
+++ b/ql/src/test/queries/clientpositive/msck_repair_0.q
@@ -1,36 +1,36 @@
set hive.msck.repair.batch.size=1;
set hive.mv.files.thread=0;
-DROP TABLE IF EXISTS repairtable;
+DROP TABLE IF EXISTS repairtable_n5;
-CREATE TABLE repairtable(col STRING) PARTITIONED BY (p1 STRING, p2 STRING);
+CREATE TABLE repairtable_n5(col STRING) PARTITIONED BY (p1 STRING, p2 STRING);
-MSCK TABLE repairtable;
+MSCK TABLE repairtable_n5;
-show partitions repairtable;
+show partitions repairtable_n5;
dfs ${system:test.dfs.mkdir} ${system:test.warehouse.dir}/repairtable/p1=c/p2=a/p3=b;
dfs -touchz ${system:test.warehouse.dir}/repairtable/p1=c/p2=a/p3=b/datafile;
-MSCK TABLE default.repairtable;
+MSCK TABLE default.repairtable_n5;
-show partitions default.repairtable;
+show partitions default.repairtable_n5;
-MSCK REPAIR TABLE default.repairtable;
+MSCK REPAIR TABLE default.repairtable_n5;
-show partitions default.repairtable;
+show partitions default.repairtable_n5;
-MSCK TABLE repairtable;
+MSCK TABLE repairtable_n5;
-show partitions repairtable;
+show partitions repairtable_n5;
set hive.mapred.mode=strict;
dfs ${system:test.dfs.mkdir} ${system:test.warehouse.dir}/repairtable/p1=e/p2=f/p3=g;
dfs -touchz ${system:test.warehouse.dir}/repairtable/p1=e/p2=f/p3=g/datafile;
-MSCK REPAIR TABLE default.repairtable;
+MSCK REPAIR TABLE default.repairtable_n5;
-show partitions default.repairtable;
+show partitions default.repairtable_n5;
-DROP TABLE default.repairtable;
+DROP TABLE default.repairtable_n5;
http://git-wip-us.apache.org/repos/asf/hive/blob/38d3b8e1/ql/src/test/queries/clientpositive/msck_repair_2.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/msck_repair_2.q b/ql/src/test/queries/clientpositive/msck_repair_2.q
index 77785e9..be745b2 100644
--- a/ql/src/test/queries/clientpositive/msck_repair_2.q
+++ b/ql/src/test/queries/clientpositive/msck_repair_2.q
@@ -1,25 +1,25 @@
set hive.msck.repair.batch.size=1;
set hive.msck.path.validation=skip;
-DROP TABLE IF EXISTS repairtable;
+DROP TABLE IF EXISTS repairtable_n2;
-CREATE TABLE repairtable(col STRING) PARTITIONED BY (p1 STRING, p2 STRING);
+CREATE TABLE repairtable_n2(col STRING) PARTITIONED BY (p1 STRING, p2 STRING);
-MSCK TABLE repairtable;
+MSCK TABLE repairtable_n2;
-show partitions repairtable;
+show partitions repairtable_n2;
dfs ${system:test.dfs.mkdir} ${system:test.warehouse.dir}/repairtable/p1=c/p2=a/p3=b;
dfs -touchz ${system:test.warehouse.dir}/repairtable/p1=c/p2=a/p3=b/datafile;
dfs -touchz ${system:test.warehouse.dir}/repairtable/p1=c/datafile;
-MSCK TABLE default.repairtable;
-show partitions repairtable;
+MSCK TABLE default.repairtable_n2;
+show partitions repairtable_n2;
-MSCK REPAIR TABLE default.repairtable;
-show partitions repairtable;
+MSCK REPAIR TABLE default.repairtable_n2;
+show partitions repairtable_n2;
-MSCK TABLE repairtable;
-show partitions repairtable;
+MSCK TABLE repairtable_n2;
+show partitions repairtable_n2;
-DROP TABLE default.repairtable;
+DROP TABLE default.repairtable_n2;
http://git-wip-us.apache.org/repos/asf/hive/blob/38d3b8e1/ql/src/test/queries/clientpositive/msck_repair_3.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/msck_repair_3.q b/ql/src/test/queries/clientpositive/msck_repair_3.q
index f42443f..140a690 100644
--- a/ql/src/test/queries/clientpositive/msck_repair_3.q
+++ b/ql/src/test/queries/clientpositive/msck_repair_3.q
@@ -1,21 +1,21 @@
set hive.msck.repair.batch.size=1;
-DROP TABLE IF EXISTS repairtable;
+DROP TABLE IF EXISTS repairtable_n3;
-CREATE TABLE repairtable(col STRING) PARTITIONED BY (p1 STRING, p2 STRING);
+CREATE TABLE repairtable_n3(col STRING) PARTITIONED BY (p1 STRING, p2 STRING);
-MSCK TABLE repairtable;
-show partitions repairtable;
+MSCK TABLE repairtable_n3;
+show partitions repairtable_n3;
dfs ${system:test.dfs.mkdir} ${system:test.warehouse.dir}/repairtable/p1=c/p2=a/p3=b;
-MSCK TABLE default.repairtable;
-show partitions repairtable;
+MSCK TABLE default.repairtable_n3;
+show partitions repairtable_n3;
-MSCK REPAIR TABLE default.repairtable;
-show partitions repairtable;
+MSCK REPAIR TABLE default.repairtable_n3;
+show partitions repairtable_n3;
-MSCK TABLE repairtable;
-show partitions repairtable;
+MSCK TABLE repairtable_n3;
+show partitions repairtable_n3;
-DROP TABLE default.repairtable;
+DROP TABLE default.repairtable_n3;
http://git-wip-us.apache.org/repos/asf/hive/blob/38d3b8e1/ql/src/test/queries/clientpositive/msck_repair_batchsize.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/msck_repair_batchsize.q b/ql/src/test/queries/clientpositive/msck_repair_batchsize.q
index a44c00e..5a7afcc 100644
--- a/ql/src/test/queries/clientpositive/msck_repair_batchsize.q
+++ b/ql/src/test/queries/clientpositive/msck_repair_batchsize.q
@@ -1,10 +1,10 @@
set hive.msck.repair.batch.size=2;
-DROP TABLE IF EXISTS repairtable;
+DROP TABLE IF EXISTS repairtable_n0;
-CREATE TABLE repairtable(col STRING) PARTITIONED BY (p1 STRING, p2 STRING);
+CREATE TABLE repairtable_n0(col STRING) PARTITIONED BY (p1 STRING, p2 STRING);
-MSCK TABLE repairtable;
+MSCK TABLE repairtable_n0;
dfs ${system:test.dfs.mkdir} ${system:test.warehouse.dir}/repairtable/p1=a/p2=a;
dfs ${system:test.dfs.mkdir} ${system:test.warehouse.dir}/repairtable/p1=b/p2=a;
@@ -13,24 +13,24 @@ dfs -touchz ${system:test.warehouse.dir}/repairtable/p1=a/p2=a/datafile;
dfs -touchz ${system:test.warehouse.dir}/repairtable/p1=b/p2=a/datafile;
dfs -touchz ${system:test.warehouse.dir}/repairtable/p1=c/p2=a/datafile;
-MSCK TABLE default.repairtable;
-show partitions default.repairtable;
+MSCK TABLE default.repairtable_n0;
+show partitions default.repairtable_n0;
-MSCK REPAIR TABLE default.repairtable;
-show partitions default.repairtable;
+MSCK REPAIR TABLE default.repairtable_n0;
+show partitions default.repairtable_n0;
-MSCK TABLE repairtable;
-show partitions repairtable;
+MSCK TABLE repairtable_n0;
+show partitions repairtable_n0;
-DROP TABLE default.repairtable;
+DROP TABLE default.repairtable_n0;
dfs ${system:test.dfs.mkdir} -p ${system:test.tmp.dir}/apps/hive/warehouse/test.db/repairtable/p1=c/p2=a/p3=b;
-CREATE TABLE `repairtable`( `col` string) PARTITIONED BY ( `p1` string, `p2` string) location '${system:test.tmp.dir}/apps/hive/warehouse/test.db/repairtable/';
+CREATE TABLE `repairtable_n0`( `col` string) PARTITIONED BY ( `p1` string, `p2` string) location '${system:test.tmp.dir}/apps/hive/warehouse/test.db/repairtable/';
dfs -touchz ${system:test.tmp.dir}/apps/hive/warehouse/test.db/repairtable/p1=c/p2=a/p3=b/datafile;
set hive.mv.files.thread=1;
-MSCK TABLE repairtable;
-show partitions repairtable;
+MSCK TABLE repairtable_n0;
+show partitions repairtable_n0;
-DROP TABLE default.repairtable;
+DROP TABLE default.repairtable_n0;
http://git-wip-us.apache.org/repos/asf/hive/blob/38d3b8e1/ql/src/test/queries/clientpositive/msck_repair_drop.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/msck_repair_drop.q b/ql/src/test/queries/clientpositive/msck_repair_drop.q
index bc14d98..9923fb5 100644
--- a/ql/src/test/queries/clientpositive/msck_repair_drop.q
+++ b/ql/src/test/queries/clientpositive/msck_repair_drop.q
@@ -1,9 +1,9 @@
set hive.mv.files.thread=0;
-DROP TABLE IF EXISTS repairtable;
+DROP TABLE IF EXISTS repairtable_n1;
-CREATE TABLE repairtable(col STRING) PARTITIONED BY (p1 STRING, p2 STRING);
+CREATE TABLE repairtable_n1(col STRING) PARTITIONED BY (p1 STRING, p2 STRING);
--- repairtable will have partitions created with part keys p1=1, p1=2, p1=3, p1=4 and p1=5
+-- repairtable_n1 will have partitions created with part keys p1=1, p1=2, p1=3, p1=4 and p1=5
-- p1=2 will be used to test drop in 3 tests
-- 1. each partition is dropped individually: set hive.msck.repair.batch.size=1;
-- 2. partition are dropped in groups of 3: set hive.msck.repair.batch.size=3;
@@ -37,20 +37,20 @@ dfs -touchz ${system:test.warehouse.dir}/repairtable/p1=2/p2=29/p3=291/datafile;
dfs ${system:test.dfs.mkdir} ${system:test.warehouse.dir}/repairtable/p1=2/p2=210/p3=2101;
dfs -touchz ${system:test.warehouse.dir}/repairtable/p1=2/p2=210/p3=2101/datafile;
-MSCK TABLE default.repairtable;
-MSCK REPAIR TABLE default.repairtable;
+MSCK TABLE default.repairtable_n1;
+MSCK REPAIR TABLE default.repairtable_n1;
-- Now all 12 partitions are in
-show partitions default.repairtable;
+show partitions default.repairtable_n1;
-- Remove all p1=2 partitions from file system
dfs -rmr ${system:test.warehouse.dir}/repairtable/p1=2;
-- test 1: each partition is dropped individually
set hive.msck.repair.batch.size=1;
-MSCK TABLE default.repairtable DROP PARTITIONS;
-MSCK REPAIR TABLE default.repairtable DROP PARTITIONS;
-show partitions default.repairtable;
+MSCK TABLE default.repairtable_n1 DROP PARTITIONS;
+MSCK REPAIR TABLE default.repairtable_n1 DROP PARTITIONS;
+show partitions default.repairtable_n1;
-- Recreate p1=2 partitions
dfs ${system:test.dfs.mkdir} ${system:test.warehouse.dir}/repairtable/p1=2/p2=21/p3=211;
@@ -74,20 +74,20 @@ dfs -touchz ${system:test.warehouse.dir}/repairtable/p1=2/p2=29/p3=291/datafile;
dfs ${system:test.dfs.mkdir} ${system:test.warehouse.dir}/repairtable/p1=2/p2=210/p3=2101;
dfs -touchz ${system:test.warehouse.dir}/repairtable/p1=2/p2=210/p3=2101/datafile;
-MSCK TABLE default.repairtable;
-MSCK REPAIR TABLE default.repairtable;
+MSCK TABLE default.repairtable_n1;
+MSCK REPAIR TABLE default.repairtable_n1;
-- Now all 12 partitions are in
-show partitions default.repairtable;
+show partitions default.repairtable_n1;
-- Remove all p1=2 partitions from file system
dfs -rmr ${system:test.warehouse.dir}/repairtable/p1=2;
-- test 2: partition are dropped in groups of 3
set hive.msck.repair.batch.size=3;
-MSCK TABLE default.repairtable DROP PARTITIONS;
-MSCK REPAIR TABLE default.repairtable DROP PARTITIONS;
-show partitions default.repairtable;
+MSCK TABLE default.repairtable_n1 DROP PARTITIONS;
+MSCK REPAIR TABLE default.repairtable_n1 DROP PARTITIONS;
+show partitions default.repairtable_n1;
-- Recreate p1=2 partitions
dfs ${system:test.dfs.mkdir} ${system:test.warehouse.dir}/repairtable/p1=2/p2=21/p3=211;
@@ -111,20 +111,20 @@ dfs -touchz ${system:test.warehouse.dir}/repairtable/p1=2/p2=29/p3=291/datafile;
dfs ${system:test.dfs.mkdir} ${system:test.warehouse.dir}/repairtable/p1=2/p2=210/p3=2101;
dfs -touchz ${system:test.warehouse.dir}/repairtable/p1=2/p2=210/p3=2101/datafile;
-MSCK TABLE default.repairtable;
-MSCK REPAIR TABLE default.repairtable;
+MSCK TABLE default.repairtable_n1;
+MSCK REPAIR TABLE default.repairtable_n1;
-- Now all 12 partitions are in
-show partitions default.repairtable;
+show partitions default.repairtable_n1;
-- Remove all p1=2 partitions from file system
dfs -rmr ${system:test.warehouse.dir}/repairtable/p1=2;
-- test 3. all partitions are dropped in 1 shot
set hive.msck.repair.batch.size=0;
-MSCK TABLE default.repairtable DROP PARTITIONS;
-MSCK REPAIR TABLE default.repairtable DROP PARTITIONS;
-show partitions default.repairtable;
+MSCK TABLE default.repairtable_n1 DROP PARTITIONS;
+MSCK REPAIR TABLE default.repairtable_n1 DROP PARTITIONS;
+show partitions default.repairtable_n1;
-- test add parition keyword: begin
dfs ${system:test.dfs.mkdir} ${system:test.warehouse.dir}/repairtable/p1=3/p2=31/p3=311;
@@ -132,9 +132,9 @@ dfs -touchz ${system:test.warehouse.dir}/repairtable/p1=3/p2=31/p3=311/datafile;
dfs ${system:test.dfs.mkdir} ${system:test.warehouse.dir}/repairtable/p1=3/p2=32/p3=321;
dfs -touchz ${system:test.warehouse.dir}/repairtable/p1=3/p2=32/p3=321/datafile;
-MSCK TABLE default.repairtable;
-MSCK REPAIR TABLE default.repairtable;
-show partitions default.repairtable;
+MSCK TABLE default.repairtable_n1;
+MSCK REPAIR TABLE default.repairtable_n1;
+show partitions default.repairtable_n1;
-- Create p1=4 in filesystem
dfs ${system:test.dfs.mkdir} ${system:test.warehouse.dir}/repairtable/p1=4/p2=41/p3=411;
@@ -148,9 +148,9 @@ dfs -rmr ${system:test.warehouse.dir}/repairtable/p1=3;
-- Status: p1=3 dropped from filesystem, but exists in metastore
-- p1=4 exists in filesystem but not in metastore
-- test add partition: only brings in p1=4 and doesn't remove p1=3
-MSCK TABLE default.repairtable ADD PARTITIONS;
-MSCK REPAIR TABLE default.repairtable ADD PARTITIONS;
-show partitions default.repairtable;
+MSCK TABLE default.repairtable_n1 ADD PARTITIONS;
+MSCK REPAIR TABLE default.repairtable_n1 ADD PARTITIONS;
+show partitions default.repairtable_n1;
-- test add partition keyword: end
-- test drop partition keyword: begin
@@ -162,9 +162,9 @@ dfs -touchz ${system:test.warehouse.dir}/repairtable/p1=5/p2=52/p3=521/datafile;
-- Status: p1=3 removed from filesystem, but exists in metastore (as part of add test)
-- p1=5 exists in filesystem but not in metastore
-- test drop partition: only removes p1=3 from metastore but doesn't update metadata for p1=5
-MSCK TABLE default.repairtable DROP PARTITIONS;
-MSCK REPAIR TABLE default.repairtable DROP PARTITIONS;
-show partitions default.repairtable;
+MSCK TABLE default.repairtable_n1 DROP PARTITIONS;
+MSCK REPAIR TABLE default.repairtable_n1 DROP PARTITIONS;
+show partitions default.repairtable_n1;
-- test drop partition keyword: end
-- test sync partition keyword: begin
@@ -174,7 +174,7 @@ dfs -rmr ${system:test.warehouse.dir}/repairtable/p1=4;
-- Status: p1=4 dropped from filesystem, but exists in metastore
-- p1=5 exists in filesystem but not in metastore (as part of drop test)
-- test sync partition: removes p1=4 from metastore and updates metadata for p1=5
-MSCK TABLE default.repairtable SYNC PARTITIONS;
-MSCK REPAIR TABLE default.repairtable SYNC PARTITIONS;
-show partitions default.repairtable;
+MSCK TABLE default.repairtable_n1 SYNC PARTITIONS;
+MSCK REPAIR TABLE default.repairtable_n1 SYNC PARTITIONS;
+show partitions default.repairtable_n1;
-- test sync partition keyword: end
http://git-wip-us.apache.org/repos/asf/hive/blob/38d3b8e1/ql/src/test/queries/clientpositive/multiMapJoin2.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/multiMapJoin2.q b/ql/src/test/queries/clientpositive/multiMapJoin2.q
index efaadcb..166cb09 100644
--- a/ql/src/test/queries/clientpositive/multiMapJoin2.q
+++ b/ql/src/test/queries/clientpositive/multiMapJoin2.q
@@ -179,18 +179,18 @@ GROUP BY tmp1.key
ORDER BY key, cnt;
-- Check if we can correctly handle partitioned table.
-CREATE TABLE part_table(key string, value string) PARTITIONED BY (partitionId int);
-INSERT OVERWRITE TABLE part_table PARTITION (partitionId=1)
+CREATE TABLE part_table_n0(key string, value string) PARTITIONED BY (partitionId int);
+INSERT OVERWRITE TABLE part_table_n0 PARTITION (partitionId=1)
SELECT key, value FROM src ORDER BY key, value LIMIT 100;
-INSERT OVERWRITE TABLE part_table PARTITION (partitionId=2)
+INSERT OVERWRITE TABLE part_table_n0 PARTITION (partitionId=2)
SELECT key, value FROM src1 ORDER BY key, value;
EXPLAIN
SELECT count(*)
-FROM part_table x JOIN src1 y ON (x.key = y.key);
+FROM part_table_n0 x JOIN src1 y ON (x.key = y.key);
SELECT count(*)
-FROM part_table x JOIN src1 y ON (x.key = y.key);
+FROM part_table_n0 x JOIN src1 y ON (x.key = y.key);
set hive.auto.convert.join.noconditionaltask.size=10000000;
set hive.optimize.correlation=false;
http://git-wip-us.apache.org/repos/asf/hive/blob/38d3b8e1/ql/src/test/queries/clientpositive/multi_column_in.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/multi_column_in.q b/ql/src/test/queries/clientpositive/multi_column_in.q
index 288406c..44fb1d9 100644
--- a/ql/src/test/queries/clientpositive/multi_column_in.q
+++ b/ql/src/test/queries/clientpositive/multi_column_in.q
@@ -1,72 +1,72 @@
set hive.mapred.mode=nonstrict;
-drop table emps;
+drop table emps_n1;
-create table emps (empno int, deptno int, empname string);
+create table emps_n1 (empno int, deptno int, empname string);
-insert into table emps values (1,2,"11"),(1,2,"11"),(3,4,"33"),(1,3,"11"),(2,5,"22"),(2,5,"22");
+insert into table emps_n1 values (1,2,"11"),(1,2,"11"),(3,4,"33"),(1,3,"11"),(2,5,"22"),(2,5,"22");
-select * from emps;
+select * from emps_n1;
-select * from emps where (int(empno+deptno/2), int(deptno/3)) in ((2,0),(3,2));
+select * from emps_n1 where (int(empno+deptno/2), int(deptno/3)) in ((2,0),(3,2));
-select * from emps where (int(empno+deptno/2), int(deptno/3)) not in ((2,0),(3,2));
+select * from emps_n1 where (int(empno+deptno/2), int(deptno/3)) not in ((2,0),(3,2));
-select * from emps where (empno,deptno) in ((1,2),(3,2));
+select * from emps_n1 where (empno,deptno) in ((1,2),(3,2));
-select * from emps where (empno,deptno) not in ((1,2),(3,2));
+select * from emps_n1 where (empno,deptno) not in ((1,2),(3,2));
-select * from emps where (empno,deptno) in ((1,2),(1,3));
+select * from emps_n1 where (empno,deptno) in ((1,2),(1,3));
-select * from emps where (empno,deptno) not in ((1,2),(1,3));
+select * from emps_n1 where (empno,deptno) not in ((1,2),(1,3));
explain
-select * from emps where (empno+1,deptno) in ((1,2),(3,2));
+select * from emps_n1 where (empno+1,deptno) in ((1,2),(3,2));
explain
-select * from emps where (empno+1,deptno) not in ((1,2),(3,2));
+select * from emps_n1 where (empno+1,deptno) not in ((1,2),(3,2));
-select * from emps where empno in (1,2);
+select * from emps_n1 where empno in (1,2);
-select * from emps where empno in (1,2) and deptno > 2;
+select * from emps_n1 where empno in (1,2) and deptno > 2;
-select * from emps where (empno) in (1,2) and deptno > 2;
+select * from emps_n1 where (empno) in (1,2) and deptno > 2;
-select * from emps where ((empno) in (1,2) and deptno > 2);
+select * from emps_n1 where ((empno) in (1,2) and deptno > 2);
-explain select * from emps where ((empno*2)|1,deptno) in ((empno+1,2),(empno+2,2));
+explain select * from emps_n1 where ((empno*2)|1,deptno) in ((empno+1,2),(empno+2,2));
-select * from emps where ((empno*2)|1,deptno) in ((empno+1,2),(empno+2,2));
+select * from emps_n1 where ((empno*2)|1,deptno) in ((empno+1,2),(empno+2,2));
-select (empno*2)|1,substr(empname,1,1) from emps;
+select (empno*2)|1,substr(empname,1,1) from emps_n1;
-select * from emps where ((empno*2)|1,substr(empname,1,1)) in ((empno+1,'2'),(empno+2,'2'));
+select * from emps_n1 where ((empno*2)|1,substr(empname,1,1)) in ((empno+1,'2'),(empno+2,'2'));
-select * from emps where ((empno*2)|1,substr(empname,1,1)) not in ((empno+1,'2'),(empno+2,'2'));
+select * from emps_n1 where ((empno*2)|1,substr(empname,1,1)) not in ((empno+1,'2'),(empno+2,'2'));
-select * from emps where ((empno*2)|1,substr(empname,1,1)) in ((empno+1,'2'),(empno+3,'2'));
+select * from emps_n1 where ((empno*2)|1,substr(empname,1,1)) in ((empno+1,'2'),(empno+3,'2'));
-select * from emps where ((empno*2)|1,substr(empname,1,1)) not in ((empno+1,'2'),(empno+3,'2'));
+select * from emps_n1 where ((empno*2)|1,substr(empname,1,1)) not in ((empno+1,'2'),(empno+3,'2'));
-select sum(empno), empname from emps where ((empno*2)|1,substr(empname,1,1)) in ((empno+1,'2'),(empno+3,'2'))
+select sum(empno), empname from emps_n1 where ((empno*2)|1,substr(empname,1,1)) in ((empno+1,'2'),(empno+3,'2'))
group by empname;
-select * from emps where ((empno*2)|1,substr(empname,1,1)) in ((empno+1,'2'),(empno+3,'2'))
+select * from emps_n1 where ((empno*2)|1,substr(empname,1,1)) in ((empno+1,'2'),(empno+3,'2'))
union
-select * from emps where (empno,deptno) in ((1,2),(3,2));
+select * from emps_n1 where (empno,deptno) in ((1,2),(3,2));
-drop view v;
+drop view v_n2;
-create view v as
+create view v_n2 as
select * from(
-select * from emps where ((empno*2)|1,substr(empname,1,1)) in ((empno+1,'2'),(empno+3,'2'))
+select * from emps_n1 where ((empno*2)|1,substr(empname,1,1)) in ((empno+1,'2'),(empno+3,'2'))
union
-select * from emps where (empno,deptno) in ((1,2),(3,2)))subq order by empno desc;
+select * from emps_n1 where (empno,deptno) in ((1,2),(3,2)))subq order by empno desc;
-select * from v;
+select * from v_n2;
select subq.e1 from
-(select (empno*2)|1 as e1, substr(empname,1,1) as n1 from emps)subq
+(select (empno*2)|1 as e1, substr(empname,1,1) as n1 from emps_n1)subq
join
-(select empno as e2 from emps where ((empno*2)|1,substr(empname,1,1)) in ((empno+1,'2'),(empno+3,'2')))subq2
+(select empno as e2 from emps_n1 where ((empno*2)|1,substr(empname,1,1)) in ((empno+1,'2'),(empno+3,'2')))subq2
on e1=e2+1;
http://git-wip-us.apache.org/repos/asf/hive/blob/38d3b8e1/ql/src/test/queries/clientpositive/multi_column_in_single.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/multi_column_in_single.q b/ql/src/test/queries/clientpositive/multi_column_in_single.q
index 7151fc4..32a2167 100644
--- a/ql/src/test/queries/clientpositive/multi_column_in_single.q
+++ b/ql/src/test/queries/clientpositive/multi_column_in_single.q
@@ -3,66 +3,66 @@ set hive.mapred.mode=nonstrict;
select * from src where (key, value) in (('238','val_238'));
-drop table emps;
+drop table emps_n7;
-create table emps (empno int, deptno int, empname string);
+create table emps_n7 (empno int, deptno int, empname string);
-insert into table emps values (1,2,"11"),(1,2,"11"),(3,4,"33"),(1,3,"11"),(2,5,"22"),(2,5,"22");
+insert into table emps_n7 values (1,2,"11"),(1,2,"11"),(3,4,"33"),(1,3,"11"),(2,5,"22"),(2,5,"22");
-select * from emps;
+select * from emps_n7;
-select * from emps where (int(empno+deptno/2), int(deptno/3)) in ((3,2));
+select * from emps_n7 where (int(empno+deptno/2), int(deptno/3)) in ((3,2));
-select * from emps where (int(empno+deptno/2), int(deptno/3)) not in ((3,2));
+select * from emps_n7 where (int(empno+deptno/2), int(deptno/3)) not in ((3,2));
-select * from emps where (empno,deptno) in ((3,2));
+select * from emps_n7 where (empno,deptno) in ((3,2));
-select * from emps where (empno,deptno) not in ((3,2));
+select * from emps_n7 where (empno,deptno) not in ((3,2));
-select * from emps where (empno,deptno) in ((1,3));
+select * from emps_n7 where (empno,deptno) in ((1,3));
-select * from emps where (empno,deptno) not in ((1,3));
+select * from emps_n7 where (empno,deptno) not in ((1,3));
explain
-select * from emps where (empno+1,deptno) in ((3,2));
+select * from emps_n7 where (empno+1,deptno) in ((3,2));
explain
-select * from emps where (empno+1,deptno) not in ((3,2));
+select * from emps_n7 where (empno+1,deptno) not in ((3,2));
-explain select * from emps where ((empno*2)|1,deptno) in ((empno+2,2));
+explain select * from emps_n7 where ((empno*2)|1,deptno) in ((empno+2,2));
-select * from emps where ((empno*2)|1,deptno) in ((empno+2,2));
+select * from emps_n7 where ((empno*2)|1,deptno) in ((empno+2,2));
-select (empno*2)|1,substr(empname,1,1) from emps;
+select (empno*2)|1,substr(empname,1,1) from emps_n7;
-select * from emps where ((empno*2)|1,substr(empname,1,1)) in ((empno+2,'2'));
+select * from emps_n7 where ((empno*2)|1,substr(empname,1,1)) in ((empno+2,'2'));
-select * from emps where ((empno*2)|1,substr(empname,1,1)) not in ((empno+2,'2'));
+select * from emps_n7 where ((empno*2)|1,substr(empname,1,1)) not in ((empno+2,'2'));
-select * from emps where ((empno*2)|1,substr(empname,1,1)) in ((empno+3,'2'));
+select * from emps_n7 where ((empno*2)|1,substr(empname,1,1)) in ((empno+3,'2'));
-select * from emps where ((empno*2)|1,substr(empname,1,1)) not in ((empno+3,'2'));
+select * from emps_n7 where ((empno*2)|1,substr(empname,1,1)) not in ((empno+3,'2'));
-select sum(empno), empname from emps where ((empno*2)|1,substr(empname,1,1)) in ((empno+3,'2'))
+select sum(empno), empname from emps_n7 where ((empno*2)|1,substr(empname,1,1)) in ((empno+3,'2'))
group by empname;
-select * from emps where ((empno*2)|1,substr(empname,1,1)) in ((empno+3,'2'))
+select * from emps_n7 where ((empno*2)|1,substr(empname,1,1)) in ((empno+3,'2'))
union
-select * from emps where (empno,deptno) in ((3,2));
+select * from emps_n7 where (empno,deptno) in ((3,2));
-drop view v;
+drop view v_n11;
-create view v as
+create view v_n11 as
select * from(
-select * from emps where ((empno*2)|1,substr(empname,1,1)) in ((empno+3,'2'))
+select * from emps_n7 where ((empno*2)|1,substr(empname,1,1)) in ((empno+3,'2'))
union
-select * from emps where (empno,deptno) in ((3,2)))subq order by empno desc;
+select * from emps_n7 where (empno,deptno) in ((3,2)))subq order by empno desc;
-select * from v;
+select * from v_n11;
select subq.e1 from
-(select (empno*2)|1 as e1, substr(empname,1,1) as n1 from emps)subq
+(select (empno*2)|1 as e1, substr(empname,1,1) as n1 from emps_n7)subq
join
-(select empno as e2 from emps where ((empno*2)|1,substr(empname,1,1)) in ((empno+3,'2')))subq2
+(select empno as e2 from emps_n7 where ((empno*2)|1,substr(empname,1,1)) in ((empno+3,'2')))subq2
on e1=e2+1;
http://git-wip-us.apache.org/repos/asf/hive/blob/38d3b8e1/ql/src/test/queries/clientpositive/multi_count_distinct.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/multi_count_distinct.q b/ql/src/test/queries/clientpositive/multi_count_distinct.q
index 038be3e..6a9cbc9 100644
--- a/ql/src/test/queries/clientpositive/multi_count_distinct.q
+++ b/ql/src/test/queries/clientpositive/multi_count_distinct.q
@@ -1,36 +1,36 @@
SET hive.vectorized.execution.enabled=false;
set hive.mapred.mode=nonstrict;
-drop table employee;
+drop table employee_n1;
-create table employee (department_id int, gender varchar(10), education_level int);
+create table employee_n1 (department_id int, gender varchar(10), education_level int);
-insert into employee values (1, 'M', 1),(1, 'M', 1),(2, 'F', 1),(1, 'F', 3),(1, 'M', 2),(4, 'M', 1),(2, 'F', 1),(2, 'F', 3),(3, 'M', 2);
+insert into employee_n1 values (1, 'M', 1),(1, 'M', 1),(2, 'F', 1),(1, 'F', 3),(1, 'M', 2),(4, 'M', 1),(2, 'F', 1),(2, 'F', 3),(3, 'M', 2);
-explain select count(distinct department_id), count(distinct gender), count(distinct education_level) from employee;
+explain select count(distinct department_id), count(distinct gender), count(distinct education_level) from employee_n1;
-select count(distinct department_id), count(distinct gender), count(distinct education_level) from employee;
+select count(distinct department_id), count(distinct gender), count(distinct education_level) from employee_n1;
-select count(distinct department_id), count(distinct gender), count(distinct education_level), count(distinct education_level) from employee;
+select count(distinct department_id), count(distinct gender), count(distinct education_level), count(distinct education_level) from employee_n1;
select count(distinct department_id), count(distinct gender), count(distinct education_level),
-count(distinct education_level, department_id) from employee;
+count(distinct education_level, department_id) from employee_n1;
select count(distinct gender), count(distinct department_id), count(distinct gender), count(distinct education_level),
-count(distinct education_level, department_id), count(distinct department_id, education_level) from employee;
+count(distinct education_level, department_id), count(distinct department_id, education_level) from employee_n1;
explain select count(distinct gender), count(distinct department_id), count(distinct gender), count(distinct education_level),
-count(distinct education_level, department_id), count(distinct department_id, education_level), count(distinct department_id, education_level, gender) from employee;
+count(distinct education_level, department_id), count(distinct department_id, education_level), count(distinct department_id, education_level, gender) from employee_n1;
select count(distinct gender), count(distinct department_id), count(distinct gender), count(distinct education_level),
-count(distinct education_level, department_id), count(distinct department_id, education_level), count(distinct department_id, education_level, gender) from employee;
+count(distinct education_level, department_id), count(distinct department_id, education_level), count(distinct department_id, education_level, gender) from employee_n1;
select count(case i when 3 then 1 else null end) as c0, count(case i when 5 then 1 else null end) as c1,
count(case i when 6 then 1 else null end) as c2 from (select grouping__id as i, department_id, gender,
-education_level from employee group by department_id, gender, education_level grouping sets
+education_level from employee_n1 group by department_id, gender, education_level grouping sets
(department_id, gender, education_level))subq;
-select grouping__id as i, department_id, gender, education_level from employee
+select grouping__id as i, department_id, gender, education_level from employee_n1
group by department_id, gender, education_level grouping sets
(department_id, gender, education_level, (education_level, department_id));
http://git-wip-us.apache.org/repos/asf/hive/blob/38d3b8e1/ql/src/test/queries/clientpositive/multi_insert.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/multi_insert.q b/ql/src/test/queries/clientpositive/multi_insert.q
index 9e33a84..634bdb4 100644
--- a/ql/src/test/queries/clientpositive/multi_insert.q
+++ b/ql/src/test/queries/clientpositive/multi_insert.q
@@ -1,23 +1,23 @@
--! qt:dataset:src
-- SORT_QUERY_RESULTS
-create table src_multi1 like src;
-create table src_multi2 like src;
+create table src_multi1_n5 like src;
+create table src_multi2_n6 like src;
set hive.merge.mapfiles=false;
set hive.merge.mapredfiles=false;
set hive.stats.dbclass=fs;
explain
from src
-insert overwrite table src_multi1 select * where key < 10
-insert overwrite table src_multi2 select * where key > 10 and key < 20;
+insert overwrite table src_multi1_n5 select * where key < 10
+insert overwrite table src_multi2_n6 select * where key > 10 and key < 20;
from src
-insert overwrite table src_multi1 select * where key < 10
-insert overwrite table src_multi2 select * where key > 10 and key < 20;
+insert overwrite table src_multi1_n5 select * where key < 10
+insert overwrite table src_multi2_n6 select * where key > 10 and key < 20;
-select * from src_multi1;
-select * from src_multi2;
+select * from src_multi1_n5;
+select * from src_multi2_n6;
set hive.merge.mapfiles=true;
@@ -25,45 +25,45 @@ set hive.merge.mapredfiles=false;
explain
from src
-insert overwrite table src_multi1 select * where key < 10
-insert overwrite table src_multi2 select * where key > 10 and key < 20;
+insert overwrite table src_multi1_n5 select * where key < 10
+insert overwrite table src_multi2_n6 select * where key > 10 and key < 20;
from src
-insert overwrite table src_multi1 select * where key < 10
-insert overwrite table src_multi2 select * where key > 10 and key < 20;
+insert overwrite table src_multi1_n5 select * where key < 10
+insert overwrite table src_multi2_n6 select * where key > 10 and key < 20;
-select * from src_multi1;
-select * from src_multi2;
+select * from src_multi1_n5;
+select * from src_multi2_n6;
set hive.merge.mapfiles=false;
set hive.merge.mapredfiles=true;
explain
from src
-insert overwrite table src_multi1 select * where key < 10
-insert overwrite table src_multi2 select * where key > 10 and key < 20;
+insert overwrite table src_multi1_n5 select * where key < 10
+insert overwrite table src_multi2_n6 select * where key > 10 and key < 20;
from src
-insert overwrite table src_multi1 select * where key < 10
-insert overwrite table src_multi2 select * where key > 10 and key < 20;
+insert overwrite table src_multi1_n5 select * where key < 10
+insert overwrite table src_multi2_n6 select * where key > 10 and key < 20;
-select * from src_multi1;
-select * from src_multi2;
+select * from src_multi1_n5;
+select * from src_multi2_n6;
set hive.merge.mapfiles=true;
set hive.merge.mapredfiles=true;
explain
from src
-insert overwrite table src_multi1 select * where key < 10
-insert overwrite table src_multi2 select * where key > 10 and key < 20;
+insert overwrite table src_multi1_n5 select * where key < 10
+insert overwrite table src_multi2_n6 select * where key > 10 and key < 20;
from src
-insert overwrite table src_multi1 select * where key < 10
-insert overwrite table src_multi2 select * where key > 10 and key < 20;
+insert overwrite table src_multi1_n5 select * where key < 10
+insert overwrite table src_multi2_n6 select * where key > 10 and key < 20;
-select * from src_multi1;
-select * from src_multi2;
+select * from src_multi1_n5;
+select * from src_multi2_n6;
@@ -72,15 +72,15 @@ set hive.merge.mapredfiles=false;
explain
from src
-insert overwrite table src_multi1 select * where key < 10 group by key, value
-insert overwrite table src_multi2 select * where key > 10 and key < 20 group by key, value;
+insert overwrite table src_multi1_n5 select * where key < 10 group by key, value
+insert overwrite table src_multi2_n6 select * where key > 10 and key < 20 group by key, value;
from src
-insert overwrite table src_multi1 select * where key < 10 group by key, value
-insert overwrite table src_multi2 select * where key > 10 and key < 20 group by key, value;
+insert overwrite table src_multi1_n5 select * where key < 10 group by key, value
+insert overwrite table src_multi2_n6 select * where key > 10 and key < 20 group by key, value;
-select * from src_multi1;
-select * from src_multi2;
+select * from src_multi1_n5;
+select * from src_multi2_n6;
set hive.merge.mapfiles=false;
@@ -88,30 +88,30 @@ set hive.merge.mapredfiles=true;
explain
from src
-insert overwrite table src_multi1 select * where key < 10 group by key, value
-insert overwrite table src_multi2 select * where key > 10 and key < 20 group by key, value;
+insert overwrite table src_multi1_n5 select * where key < 10 group by key, value
+insert overwrite table src_multi2_n6 select * where key > 10 and key < 20 group by key, value;
from src
-insert overwrite table src_multi1 select * where key < 10 group by key, value
-insert overwrite table src_multi2 select * where key > 10 and key < 20 group by key, value;
+insert overwrite table src_multi1_n5 select * where key < 10 group by key, value
+insert overwrite table src_multi2_n6 select * where key > 10 and key < 20 group by key, value;
-select * from src_multi1;
-select * from src_multi2;
+select * from src_multi1_n5;
+select * from src_multi2_n6;
set hive.merge.mapfiles=true;
set hive.merge.mapredfiles=false;
explain
from src
-insert overwrite table src_multi1 select * where key < 10 group by key, value
-insert overwrite table src_multi2 select * where key > 10 and key < 20 group by key, value;
+insert overwrite table src_multi1_n5 select * where key < 10 group by key, value
+insert overwrite table src_multi2_n6 select * where key > 10 and key < 20 group by key, value;
from src
-insert overwrite table src_multi1 select * where key < 10 group by key, value
-insert overwrite table src_multi2 select * where key > 10 and key < 20 group by key, value;
+insert overwrite table src_multi1_n5 select * where key < 10 group by key, value
+insert overwrite table src_multi2_n6 select * where key > 10 and key < 20 group by key, value;
-select * from src_multi1;
-select * from src_multi2;
+select * from src_multi1_n5;
+select * from src_multi2_n6;
set hive.merge.mapfiles=true;
@@ -119,15 +119,15 @@ set hive.merge.mapredfiles=true;
explain
from src
-insert overwrite table src_multi1 select * where key < 10 group by key, value
-insert overwrite table src_multi2 select * where key > 10 and key < 20 group by key, value;
+insert overwrite table src_multi1_n5 select * where key < 10 group by key, value
+insert overwrite table src_multi2_n6 select * where key > 10 and key < 20 group by key, value;
from src
-insert overwrite table src_multi1 select * where key < 10 group by key, value
-insert overwrite table src_multi2 select * where key > 10 and key < 20 group by key, value;
+insert overwrite table src_multi1_n5 select * where key < 10 group by key, value
+insert overwrite table src_multi2_n6 select * where key > 10 and key < 20 group by key, value;
-select * from src_multi1;
-select * from src_multi2;
+select * from src_multi1_n5;
+select * from src_multi2_n6;
@@ -137,60 +137,60 @@ set hive.merge.mapredfiles=false;
explain
from (select * from src union all select * from src) s
-insert overwrite table src_multi1 select * where key < 10
-insert overwrite table src_multi2 select * where key > 10 and key < 20;
+insert overwrite table src_multi1_n5 select * where key < 10
+insert overwrite table src_multi2_n6 select * where key > 10 and key < 20;
from (select * from src union all select * from src) s
-insert overwrite table src_multi1 select * where key < 10
-insert overwrite table src_multi2 select * where key > 10 and key < 20;
+insert overwrite table src_multi1_n5 select * where key < 10
+insert overwrite table src_multi2_n6 select * where key > 10 and key < 20;
-select * from src_multi1;
-select * from src_multi2;
+select * from src_multi1_n5;
+select * from src_multi2_n6;
set hive.merge.mapfiles=true;
set hive.merge.mapredfiles=false;
explain
from (select * from src union all select * from src) s
-insert overwrite table src_multi1 select * where key < 10
-insert overwrite table src_multi2 select * where key > 10 and key < 20;
+insert overwrite table src_multi1_n5 select * where key < 10
+insert overwrite table src_multi2_n6 select * where key > 10 and key < 20;
from (select * from src union all select * from src) s
-insert overwrite table src_multi1 select * where key < 10
-insert overwrite table src_multi2 select * where key > 10 and key < 20;
+insert overwrite table src_multi1_n5 select * where key < 10
+insert overwrite table src_multi2_n6 select * where key > 10 and key < 20;
-select * from src_multi1;
-select * from src_multi2;
+select * from src_multi1_n5;
+select * from src_multi2_n6;
set hive.merge.mapfiles=false;
set hive.merge.mapredfiles=true;
explain
from (select * from src union all select * from src) s
-insert overwrite table src_multi1 select * where key < 10
-insert overwrite table src_multi2 select * where key > 10 and key < 20;
+insert overwrite table src_multi1_n5 select * where key < 10
+insert overwrite table src_multi2_n6 select * where key > 10 and key < 20;
from (select * from src union all select * from src) s
-insert overwrite table src_multi1 select * where key < 10
-insert overwrite table src_multi2 select * where key > 10 and key < 20;
+insert overwrite table src_multi1_n5 select * where key < 10
+insert overwrite table src_multi2_n6 select * where key > 10 and key < 20;
-select * from src_multi1;
-select * from src_multi2;
+select * from src_multi1_n5;
+select * from src_multi2_n6;
set hive.merge.mapfiles=true;
set hive.merge.mapredfiles=true;
explain
from (select * from src union all select * from src) s
-insert overwrite table src_multi1 select * where key < 10
-insert overwrite table src_multi2 select * where key > 10 and key < 20;
+insert overwrite table src_multi1_n5 select * where key < 10
+insert overwrite table src_multi2_n6 select * where key > 10 and key < 20;
from (select * from src union all select * from src) s
-insert overwrite table src_multi1 select * where key < 10
-insert overwrite table src_multi2 select * where key > 10 and key < 20;
+insert overwrite table src_multi1_n5 select * where key < 10
+insert overwrite table src_multi2_n6 select * where key > 10 and key < 20;
-select * from src_multi1;
-select * from src_multi2;
+select * from src_multi1_n5;
+select * from src_multi2_n6;
http://git-wip-us.apache.org/repos/asf/hive/blob/38d3b8e1/ql/src/test/queries/clientpositive/multi_insert_gby.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/multi_insert_gby.q b/ql/src/test/queries/clientpositive/multi_insert_gby.q
index 7c9e8de..0ec2b35 100644
--- a/ql/src/test/queries/clientpositive/multi_insert_gby.q
+++ b/ql/src/test/queries/clientpositive/multi_insert_gby.q
@@ -2,35 +2,35 @@
-- SORT_QUERY_RESULTS
--HIVE-3699 Multiple insert overwrite into multiple tables query stores same results in all tables
-create table e1 (key string, count int);
-create table e2 (key string, count int);
+create table e1_n0 (key string, count int);
+create table e2_n1 (key string, count int);
explain FROM src
-INSERT OVERWRITE TABLE e1
+INSERT OVERWRITE TABLE e1_n0
SELECT key, COUNT(*) WHERE key>450 GROUP BY key
-INSERT OVERWRITE TABLE e2
+INSERT OVERWRITE TABLE e2_n1
SELECT key, COUNT(*) WHERE key>500 GROUP BY key;
FROM src
-INSERT OVERWRITE TABLE e1
+INSERT OVERWRITE TABLE e1_n0
SELECT key, COUNT(*) WHERE key>450 GROUP BY key
-INSERT OVERWRITE TABLE e2
+INSERT OVERWRITE TABLE e2_n1
SELECT key, COUNT(*) WHERE key>500 GROUP BY key;
-select * from e1;
-select * from e2;
+select * from e1_n0;
+select * from e2_n1;
explain FROM src
-INSERT OVERWRITE TABLE e1
+INSERT OVERWRITE TABLE e1_n0
SELECT key, COUNT(*) WHERE key>450 GROUP BY key
-INSERT OVERWRITE TABLE e2
+INSERT OVERWRITE TABLE e2_n1
SELECT key, COUNT(*) GROUP BY key;
FROM src
-INSERT OVERWRITE TABLE e1
+INSERT OVERWRITE TABLE e1_n0
SELECT key, COUNT(*) WHERE key>450 GROUP BY key
-INSERT OVERWRITE TABLE e2
+INSERT OVERWRITE TABLE e2_n1
SELECT key, COUNT(*) GROUP BY key;
-select * from e1;
-select * from e2;
+select * from e1_n0;
+select * from e2_n1;
http://git-wip-us.apache.org/repos/asf/hive/blob/38d3b8e1/ql/src/test/queries/clientpositive/multi_insert_gby2.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/multi_insert_gby2.q b/ql/src/test/queries/clientpositive/multi_insert_gby2.q
index a477f83..67077ba 100644
--- a/ql/src/test/queries/clientpositive/multi_insert_gby2.q
+++ b/ql/src/test/queries/clientpositive/multi_insert_gby2.q
@@ -2,20 +2,20 @@
set hive.mapred.mode=nonstrict;
--HIVE-3699 Multiple insert overwrite into multiple tables query stores same results in all tables
create table e1 (count int);
-create table e2 (percentile double);
+create table e2_n0 (percentile double);
set hive.stats.dbclass=fs;
explain
FROM (select key, cast(key as double) as value from src order by key) a
INSERT OVERWRITE TABLE e1
SELECT COUNT(*)
-INSERT OVERWRITE TABLE e2
+INSERT OVERWRITE TABLE e2_n0
SELECT percentile_approx(value, 0.5);
FROM (select key, cast(key as double) as value from src order by key) a
INSERT OVERWRITE TABLE e1
SELECT COUNT(*)
-INSERT OVERWRITE TABLE e2
+INSERT OVERWRITE TABLE e2_n0
SELECT percentile_approx(value, 0.5);
select * from e1;
-select * from e2;
+select * from e2_n0;
http://git-wip-us.apache.org/repos/asf/hive/blob/38d3b8e1/ql/src/test/queries/clientpositive/multi_insert_gby3.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/multi_insert_gby3.q b/ql/src/test/queries/clientpositive/multi_insert_gby3.q
index 6bed476..de60fa7 100644
--- a/ql/src/test/queries/clientpositive/multi_insert_gby3.q
+++ b/ql/src/test/queries/clientpositive/multi_insert_gby3.q
@@ -1,45 +1,45 @@
--! qt:dataset:src
set hive.mapred.mode=nonstrict;
-- SORT_QUERY_RESULTS
-create table e1 (key string, keyD double);
-create table e2 (key string, keyD double, value string);
+create table e1_n2 (key string, keyD double);
+create table e2_n3 (key string, keyD double, value string);
create table e3 (key string, keyD double);
set hive.stats.dbclass=fs;
explain
FROM (select key, cast(key as double) as keyD, value from src order by key) a
-INSERT OVERWRITE TABLE e1
+INSERT OVERWRITE TABLE e1_n2
SELECT key, COUNT(distinct value) group by key
-INSERT OVERWRITE TABLE e2
+INSERT OVERWRITE TABLE e2_n3
SELECT key, sum(keyD), value group by key, value;
explain
FROM (select key, cast(key as double) as keyD, value from src order by key) a
-INSERT OVERWRITE TABLE e2
+INSERT OVERWRITE TABLE e2_n3
SELECT key, sum(keyD), value group by key, value
-INSERT OVERWRITE TABLE e1
+INSERT OVERWRITE TABLE e1_n2
SELECT key, COUNT(distinct value) group by key;
FROM (select key, cast(key as double) as keyD, value from src order by key) a
-INSERT OVERWRITE TABLE e1
+INSERT OVERWRITE TABLE e1_n2
SELECT key, COUNT(distinct value) group by key
-INSERT OVERWRITE TABLE e2
+INSERT OVERWRITE TABLE e2_n3
SELECT key, sum(keyD), value group by key, value;
-select * from e1;
-select * from e2;
+select * from e1_n2;
+select * from e2_n3;
FROM (select key, cast(key as double) as keyD, value from src order by key) a
-INSERT OVERWRITE TABLE e2
+INSERT OVERWRITE TABLE e2_n3
SELECT key, sum(keyD), value group by key, value
-INSERT OVERWRITE TABLE e1
+INSERT OVERWRITE TABLE e1_n2
SELECT key, COUNT(distinct value) group by key;
-select * from e1;
-select * from e2;
+select * from e1_n2;
+select * from e2_n3;
explain
from src
-insert overwrite table e1
+insert overwrite table e1_n2
select key, count(distinct value) group by key
insert overwrite table e3
select value, count(distinct key) group by value;
@@ -47,9 +47,9 @@ select value, count(distinct key) group by value;
explain
FROM (select key, cast(key as double) as keyD, value from src order by key) a
-INSERT OVERWRITE TABLE e1
+INSERT OVERWRITE TABLE e1_n2
SELECT key, COUNT(distinct value) group by key
-INSERT OVERWRITE TABLE e2
+INSERT OVERWRITE TABLE e2_n3
SELECT key, sum(keyD), value group by key, value
INSERT overwrite table e3
SELECT key, COUNT(distinct keyD) group by key, keyD, value;
http://git-wip-us.apache.org/repos/asf/hive/blob/38d3b8e1/ql/src/test/queries/clientpositive/multi_insert_gby4.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/multi_insert_gby4.q b/ql/src/test/queries/clientpositive/multi_insert_gby4.q
index d485813..b89d674 100644
--- a/ql/src/test/queries/clientpositive/multi_insert_gby4.q
+++ b/ql/src/test/queries/clientpositive/multi_insert_gby4.q
@@ -1,27 +1,27 @@
--! qt:dataset:src
-- SORT_QUERY_RESULTS
-create table e1 (key string, count int);
-create table e2 (key string, count int);
-create table e3 (key string, count int);
+create table e1_n4 (key string, count int);
+create table e2_n5 (key string, count int);
+create table e3_n0 (key string, count int);
explain
FROM (SELECT key, value FROM src) a
-INSERT OVERWRITE TABLE e1
+INSERT OVERWRITE TABLE e1_n4
SELECT key, COUNT(*) WHERE key>450 GROUP BY key
-INSERT OVERWRITE TABLE e2
+INSERT OVERWRITE TABLE e2_n5
SELECT key, COUNT(*) WHERE key>500 GROUP BY key
-INSERT OVERWRITE TABLE e3
+INSERT OVERWRITE TABLE e3_n0
SELECT key, COUNT(*) WHERE key>490 GROUP BY key;
FROM (SELECT key, value FROM src) a
-INSERT OVERWRITE TABLE e1
+INSERT OVERWRITE TABLE e1_n4
SELECT key, COUNT(*) WHERE key>450 GROUP BY key
-INSERT OVERWRITE TABLE e2
+INSERT OVERWRITE TABLE e2_n5
SELECT key, COUNT(*) WHERE key>500 GROUP BY key
-INSERT OVERWRITE TABLE e3
+INSERT OVERWRITE TABLE e3_n0
SELECT key, COUNT(*) WHERE key>490 GROUP BY key;
-select * from e1;
-select * from e2;
-select * from e3;
+select * from e1_n4;
+select * from e2_n5;
+select * from e3_n0;
http://git-wip-us.apache.org/repos/asf/hive/blob/38d3b8e1/ql/src/test/queries/clientpositive/multi_insert_lateral_view.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/multi_insert_lateral_view.q b/ql/src/test/queries/clientpositive/multi_insert_lateral_view.q
index 481bdf2..3442e82 100644
--- a/ql/src/test/queries/clientpositive/multi_insert_lateral_view.q
+++ b/ql/src/test/queries/clientpositive/multi_insert_lateral_view.q
@@ -2,7 +2,7 @@
set hive.stats.dbclass=fs;
-- SORT_QUERY_RESULTS
-create table src_10 as select * from src limit 10;
+create table src_10_n0 as select * from src limit 10;
create table src_lv1 (key string, value string);
create table src_lv2 (key string, value string);
@@ -14,11 +14,11 @@ create table src_lv3 (key string, value string);
-- -LVF[6]-SEL[7]-LVJ[10]-SEL[13]-FS[14]
-- -SEL[8]-UDTF[9]-LVJ[10]
explain
-from src_10
+from src_10_n0
insert overwrite table src_lv1 select key, C lateral view explode(array(key+1, key+2)) A as C
insert overwrite table src_lv2 select key, C lateral view explode(array(key+3, key+4)) A as C;
-from src_10
+from src_10_n0
insert overwrite table src_lv1 select key, C lateral view explode(array(key+1, key+2)) A as C
insert overwrite table src_lv2 select key, C lateral view explode(array(key+3, key+4)) A as C;
@@ -31,11 +31,11 @@ select * from src_lv2;
-- -LVF[6]-SEL[7]-LVJ[10]-SEL[17]-GBY[18]-RS[19]-GBY[20]-SEL[21]-FS[22]
-- -SEL[8]-UDTF[9]-LVJ[10]
explain
-from src_10
+from src_10_n0
insert overwrite table src_lv1 select key, sum(C) lateral view explode(array(key+1, key+2)) A as C group by key
insert overwrite table src_lv2 select key, sum(C) lateral view explode(array(key+3, key+4)) A as C group by key;
-from src_10
+from src_10_n0
insert overwrite table src_lv1 select key, sum(C) lateral view explode(array(key+1, key+2)) A as C group by key
insert overwrite table src_lv2 select key, sum(C) lateral view explode(array(key+3, key+4)) A as C group by key;
@@ -48,12 +48,12 @@ select * from src_lv2;
-- -FIL[12]-SEL[13]-RS[14]-FOR[15]-FIL[16]-GBY[17]-SEL[18]-FS[19]
-- -FIL[20]-GBY[21]-SEL[22]-FS[23]
explain
-from src_10
+from src_10_n0
insert overwrite table src_lv1 select key, sum(C) lateral view explode(array(key+1, key+2)) A as C group by key
insert overwrite table src_lv2 select key, count(value) where key > 200 group by key
insert overwrite table src_lv3 select key, count(value) where key < 200 group by key;
-from src_10
+from src_10_n0
insert overwrite table src_lv1 select key, sum(C) lateral view explode(array(key+1, key+2)) A as C group by key
insert overwrite table src_lv2 select key, count(value) where key > 200 group by key
insert overwrite table src_lv3 select key, count(value) where key < 200 group by key;
@@ -70,12 +70,12 @@ select * from src_lv3;
-- -SEL[8]-UDTF[9]-LVJ[10]
-- -SEL[23]-GBY[24]-RS[25]-GBY[26]-SEL[27]-FS[28]
explain
-from src_10
+from src_10_n0
insert overwrite table src_lv1 select C, sum(distinct key) lateral view explode(array(key+1, key+2)) A as C group by C
insert overwrite table src_lv2 select C, sum(distinct key) lateral view explode(array(key+3, key+4)) A as C group by C
insert overwrite table src_lv3 select value, sum(distinct key) group by value;
-from src_10
+from src_10_n0
insert overwrite table src_lv1 select C, sum(distinct key) lateral view explode(array(key+1, key+2)) A as C group by C
insert overwrite table src_lv2 select C, sum(distinct key) lateral view explode(array(key+3, key+4)) A as C group by C
insert overwrite table src_lv3 select value, sum(distinct key) group by value;
@@ -88,13 +88,13 @@ create table src_lv4 (key string, value string);
-- Common distincts optimization works across non-lateral view queries, but not across lateral view multi inserts
explain
-from src_10
+from src_10_n0
insert overwrite table src_lv1 select key, sum(distinct C) lateral view explode(array(key+1, key+2)) A as C group by key
insert overwrite table src_lv2 select key, sum(distinct C) lateral view explode(array(key+3, key+4)) A as C group by key
insert overwrite table src_lv3 select value, sum(distinct key) where key > 200 group by value
insert overwrite table src_lv4 select value, sum(distinct key) where key < 200 group by value;
-from src_10
+from src_10_n0
insert overwrite table src_lv1 select key, sum(distinct C) lateral view explode(array(key+1, key+2)) A as C group by key
insert overwrite table src_lv2 select key, sum(distinct C) lateral view explode(array(key+3, key+4)) A as C group by key
insert overwrite table src_lv3 select value, sum(distinct key) where key > 200 group by value
http://git-wip-us.apache.org/repos/asf/hive/blob/38d3b8e1/ql/src/test/queries/clientpositive/multi_insert_mixed.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/multi_insert_mixed.q b/ql/src/test/queries/clientpositive/multi_insert_mixed.q
index a5c1e87..9c8fb2e 100644
--- a/ql/src/test/queries/clientpositive/multi_insert_mixed.q
+++ b/ql/src/test/queries/clientpositive/multi_insert_mixed.q
@@ -1,7 +1,7 @@
--! qt:dataset:src
set hive.mapred.mode=nonstrict;
-create table src_multi1 like src;
-create table src_multi2 like src;
+create table src_multi1_n2 like src;
+create table src_multi2_n3 like src;
create table src_multi3 like src;
set hive.stats.dbclass=fs;
-- Testing the case where a map work contains both shuffling (ReduceSinkOperator)
@@ -9,15 +9,15 @@ set hive.stats.dbclass=fs;
explain
from src
-insert overwrite table src_multi1 select key, count(1) group by key order by key
-insert overwrite table src_multi2 select value, count(1) group by value order by value
+insert overwrite table src_multi1_n2 select key, count(1) group by key order by key
+insert overwrite table src_multi2_n3 select value, count(1) group by value order by value
insert overwrite table src_multi3 select * where key < 10;
from src
-insert overwrite table src_multi1 select key, count(1) group by key order by key
-insert overwrite table src_multi2 select value, count(1) group by value order by value
+insert overwrite table src_multi1_n2 select key, count(1) group by key order by key
+insert overwrite table src_multi2_n3 select value, count(1) group by value order by value
insert overwrite table src_multi3 select * where key < 10;
-select * from src_multi1;
-select * from src_multi2;
+select * from src_multi1_n2;
+select * from src_multi2_n3;
select * from src_multi3;