You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hive.apache.org by jc...@apache.org on 2018/05/21 20:25:01 UTC

[24/51] [abbrv] [partial] hive git commit: HIVE-19617: Rename test tables to avoid collisions during execution in batches (Jesus Camacho Rodriguez, reviewed by Gunther Hagleitner)

http://git-wip-us.apache.org/repos/asf/hive/blob/38d3b8e1/ql/src/test/queries/clientpositive/metadata_only_queries_with_filters.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/metadata_only_queries_with_filters.q b/ql/src/test/queries/clientpositive/metadata_only_queries_with_filters.q
index 692c414..ed75c57 100644
--- a/ql/src/test/queries/clientpositive/metadata_only_queries_with_filters.q
+++ b/ql/src/test/queries/clientpositive/metadata_only_queries_with_filters.q
@@ -2,7 +2,7 @@ set hive.stats.column.autogather=false;
 set hive.stats.dbclass=fs;
 set hive.compute.query.using.stats=true;
 set hive.explain.user=false;
-create table over10k(
+create table over10k_n23(
            t tinyint,
            si smallint,
            i int,
@@ -17,9 +17,9 @@ create table over10k(
        row format delimited
        fields terminated by '|';
 
-load data local inpath '../../data/files/over10k' into table over10k;
+load data local inpath '../../data/files/over10k' into table over10k_n23;
 
-create table stats_tbl_part(
+create table stats_tbl_part_n0(
            t tinyint,
            si smallint,
            i int,
@@ -33,22 +33,22 @@ create table stats_tbl_part(
            bin binary) partitioned by (dt int);
 
 
-from over10k 
-insert overwrite table stats_tbl_part partition (dt=2010) select t,si,i,b,f,d,bo,s,ts,`dec`,bin where t>0 and t<30 
-insert overwrite table stats_tbl_part partition (dt=2014) select t,si,i,b,f,d,bo,s,ts,`dec`,bin where t > 30 and t<60;
+from over10k_n23 
+insert overwrite table stats_tbl_part_n0 partition (dt=2010) select t,si,i,b,f,d,bo,s,ts,`dec`,bin where t>0 and t<30 
+insert overwrite table stats_tbl_part_n0 partition (dt=2014) select t,si,i,b,f,d,bo,s,ts,`dec`,bin where t > 30 and t<60;
 
-analyze table stats_tbl_part partition(dt) compute statistics;
-analyze table stats_tbl_part partition(dt=2010) compute statistics for columns t,si,i,b,f,d,bo,s,bin;
-analyze table stats_tbl_part partition(dt=2014) compute statistics for columns t,si,i,b,f,d,bo,s,bin;
+analyze table stats_tbl_part_n0 partition(dt) compute statistics;
+analyze table stats_tbl_part_n0 partition(dt=2010) compute statistics for columns t,si,i,b,f,d,bo,s,bin;
+analyze table stats_tbl_part_n0 partition(dt=2014) compute statistics for columns t,si,i,b,f,d,bo,s,bin;
 
 explain 
-select count(*), count(1), sum(1), count(s), count(bo), count(bin), count(si), max(i), min(b), max(f), min(d) from stats_tbl_part where dt = 2010;
-select count(*), count(1), sum(1), count(s), count(bo), count(bin), count(si), max(i), min(b), max(f), min(d) from stats_tbl_part where dt = 2010;
+select count(*), count(1), sum(1), count(s), count(bo), count(bin), count(si), max(i), min(b), max(f), min(d) from stats_tbl_part_n0 where dt = 2010;
+select count(*), count(1), sum(1), count(s), count(bo), count(bin), count(si), max(i), min(b), max(f), min(d) from stats_tbl_part_n0 where dt = 2010;
 explain 
-select count(*), count(1), sum(1), sum(2), count(s), count(bo), count(bin), count(si), max(i), min(b), max(f), min(d) from stats_tbl_part where dt > 2010;
-select count(*), count(1), sum(1), sum(2), count(s), count(bo), count(bin), count(si), max(i), min(b), max(f), min(d) from stats_tbl_part where dt > 2010;
+select count(*), count(1), sum(1), sum(2), count(s), count(bo), count(bin), count(si), max(i), min(b), max(f), min(d) from stats_tbl_part_n0 where dt > 2010;
+select count(*), count(1), sum(1), sum(2), count(s), count(bo), count(bin), count(si), max(i), min(b), max(f), min(d) from stats_tbl_part_n0 where dt > 2010;
 
-select count(*) from stats_tbl_part;
-select count(*)/2 from stats_tbl_part;
-drop table stats_tbl_part;
+select count(*) from stats_tbl_part_n0;
+select count(*)/2 from stats_tbl_part_n0;
+drop table stats_tbl_part_n0;
 set hive.compute.query.using.stats=false;

http://git-wip-us.apache.org/repos/asf/hive/blob/38d3b8e1/ql/src/test/queries/clientpositive/metadataonly1.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/metadataonly1.q b/ql/src/test/queries/clientpositive/metadataonly1.q
index e3bf819..e247c6a 100644
--- a/ql/src/test/queries/clientpositive/metadataonly1.q
+++ b/ql/src/test/queries/clientpositive/metadataonly1.q
@@ -1,48 +1,48 @@
 --! qt:dataset:srcpart
 set hive.mapred.mode=nonstrict;
 set hive.optimize.metadataonly=true;
-CREATE TABLE TEST1(A INT, B DOUBLE) partitioned by (ds string);
-explain extended select max(ds) from TEST1;
-select max(ds) from TEST1;
+CREATE TABLE TEST1_n12(A INT, B DOUBLE) partitioned by (ds string);
+explain extended select max(ds) from TEST1_n12;
+select max(ds) from TEST1_n12;
 
-alter table TEST1 add partition (ds='1');
-explain extended select max(ds) from TEST1;
-select max(ds) from TEST1;
+alter table TEST1_n12 add partition (ds='1');
+explain extended select max(ds) from TEST1_n12;
+select max(ds) from TEST1_n12;
 
-explain extended select count(distinct ds) from TEST1;
-select count(distinct ds) from TEST1;
+explain extended select count(distinct ds) from TEST1_n12;
+select count(distinct ds) from TEST1_n12;
 
-explain extended select count(ds) from TEST1;
-select count(ds) from TEST1;
+explain extended select count(ds) from TEST1_n12;
+select count(ds) from TEST1_n12;
 
-alter table TEST1 add partition (ds='2');
+alter table TEST1_n12 add partition (ds='2');
 explain extended 
-select count(*) from TEST1 a2 join (select max(ds) m from TEST1) b on a2.ds=b.m;
-select count(*) from TEST1 a2 join (select max(ds) m from TEST1) b on a2.ds=b.m;
+select count(*) from TEST1_n12 a2 join (select max(ds) m from TEST1_n12) b on a2.ds=b.m;
+select count(*) from TEST1_n12 a2 join (select max(ds) m from TEST1_n12) b on a2.ds=b.m;
 
 
-CREATE TABLE TEST2(A INT, B DOUBLE) partitioned by (ds string, hr string);
-alter table TEST2 add partition (ds='1', hr='1');
-alter table TEST2 add partition (ds='1', hr='2');
-alter table TEST2 add partition (ds='1', hr='3');
+CREATE TABLE TEST2_n8(A INT, B DOUBLE) partitioned by (ds string, hr string);
+alter table TEST2_n8 add partition (ds='1', hr='1');
+alter table TEST2_n8 add partition (ds='1', hr='2');
+alter table TEST2_n8 add partition (ds='1', hr='3');
 
-explain extended select ds, count(distinct hr) from TEST2 group by ds;
-select ds, count(distinct hr) from TEST2 group by ds;
+explain extended select ds, count(distinct hr) from TEST2_n8 group by ds;
+select ds, count(distinct hr) from TEST2_n8 group by ds;
 
-explain extended select ds, count(hr) from TEST2 group by ds;
-select ds, count(hr) from TEST2 group by ds;
+explain extended select ds, count(hr) from TEST2_n8 group by ds;
+select ds, count(hr) from TEST2_n8 group by ds;
 
 set hive.input.format=org.apache.hadoop.hive.ql.io.CombineHiveInputFormat;
 
-explain extended select max(ds) from TEST1;
-select max(ds) from TEST1;
+explain extended select max(ds) from TEST1_n12;
+select max(ds) from TEST1_n12;
 
 select distinct ds from srcpart;
 select min(ds),max(ds) from srcpart;
 
 -- HIVE-3594 URI encoding for temporary path
-alter table TEST2 add partition (ds='01:10:10', hr='01');
-alter table TEST2 add partition (ds='01:10:20', hr='02');
+alter table TEST2_n8 add partition (ds='01:10:10', hr='01');
+alter table TEST2_n8 add partition (ds='01:10:20', hr='02');
 
-explain extended select ds, count(distinct hr) from TEST2 group by ds;
-select ds, count(distinct hr) from TEST2 group by ds;
+explain extended select ds, count(distinct hr) from TEST2_n8 group by ds;
+select ds, count(distinct hr) from TEST2_n8 group by ds;

http://git-wip-us.apache.org/repos/asf/hive/blob/38d3b8e1/ql/src/test/queries/clientpositive/misc_json.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/misc_json.q b/ql/src/test/queries/clientpositive/misc_json.q
index 541e369..95a3e5b 100644
--- a/ql/src/test/queries/clientpositive/misc_json.q
+++ b/ql/src/test/queries/clientpositive/misc_json.q
@@ -1,10 +1,10 @@
 set hive.ddl.output.format=json;
 
-CREATE TABLE IF NOT EXISTS jsontable (key INT, value STRING) COMMENT 'json table' STORED AS TEXTFILE;
+CREATE TABLE IF NOT EXISTS jsontable_n0 (key INT, value STRING) COMMENT 'json table' STORED AS TEXTFILE;
 
-ALTER TABLE jsontable ADD COLUMNS (name STRING COMMENT 'a new column'); 
+ALTER TABLE jsontable_n0 ADD COLUMNS (name STRING COMMENT 'a new column'); 
 
-ALTER TABLE jsontable RENAME TO jsontable2;
+ALTER TABLE jsontable_n0 RENAME TO jsontable2;
 
 SHOW TABLE EXTENDED LIKE jsontable2;
 

http://git-wip-us.apache.org/repos/asf/hive/blob/38d3b8e1/ql/src/test/queries/clientpositive/mm_all.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/mm_all.q b/ql/src/test/queries/clientpositive/mm_all.q
index 7d0955b..61dd3e7 100644
--- a/ql/src/test/queries/clientpositive/mm_all.q
+++ b/ql/src/test/queries/clientpositive/mm_all.q
@@ -13,32 +13,32 @@ set hive.support.concurrency=true;
 set hive.txn.manager=org.apache.hadoop.hive.ql.lockmgr.DbTxnManager;
 
 -- Force multiple writers when reading
-drop table intermediate;
-create table intermediate(key int) partitioned by (p int) stored as orc;
-insert into table intermediate partition(p='455') select distinct key from src where key >= 0 order by key desc limit 2;
-insert into table intermediate partition(p='456') select distinct key from src where key is not null order by key asc limit 2;
-insert into table intermediate partition(p='457') select distinct key from src where key >= 100 order by key asc limit 2;
-
-
-drop table part_mm;
-create table part_mm(key int) partitioned by (key_mm int) stored as orc tblproperties ("transactional"="true", "transactional_properties"="insert_only");
-explain insert into table part_mm partition(key_mm=455) select key from intermediate;
-insert into table part_mm partition(key_mm=455) select key from intermediate;
-insert into table part_mm partition(key_mm=456) select key from intermediate;
-insert into table part_mm partition(key_mm=455) select key from intermediate;
-select * from part_mm order by key, key_mm;
-
--- TODO: doesn't work truncate table part_mm partition(key_mm=455);
-select * from part_mm order by key, key_mm;
-truncate table part_mm;
-select * from part_mm order by key, key_mm;
-drop table part_mm;
+drop table intermediate_n0;
+create table intermediate_n0(key int) partitioned by (p int) stored as orc;
+insert into table intermediate_n0 partition(p='455') select distinct key from src where key >= 0 order by key desc limit 2;
+insert into table intermediate_n0 partition(p='456') select distinct key from src where key is not null order by key asc limit 2;
+insert into table intermediate_n0 partition(p='457') select distinct key from src where key >= 100 order by key asc limit 2;
+
+
+drop table part_mm_n0;
+create table part_mm_n0(key int) partitioned by (key_mm int) stored as orc tblproperties ("transactional"="true", "transactional_properties"="insert_only");
+explain insert into table part_mm_n0 partition(key_mm=455) select key from intermediate_n0;
+insert into table part_mm_n0 partition(key_mm=455) select key from intermediate_n0;
+insert into table part_mm_n0 partition(key_mm=456) select key from intermediate_n0;
+insert into table part_mm_n0 partition(key_mm=455) select key from intermediate_n0;
+select * from part_mm_n0 order by key, key_mm;
+
+-- TODO: doesn't work truncate table part_mm_n0 partition(key_mm=455);
+select * from part_mm_n0 order by key, key_mm;
+truncate table part_mm_n0;
+select * from part_mm_n0 order by key, key_mm;
+drop table part_mm_n0;
 
 drop table simple_mm;
 create table simple_mm(key int) stored as orc tblproperties ("transactional"="true", "transactional_properties"="insert_only");
-insert into table simple_mm select key from intermediate;
+insert into table simple_mm select key from intermediate_n0;
 select * from simple_mm order by key;
-insert into table simple_mm select key from intermediate;
+insert into table simple_mm select key from intermediate_n0;
 select * from simple_mm order by key;
 truncate table simple_mm;
 select * from simple_mm;
@@ -57,7 +57,7 @@ set hive.merge.tezfiles=false;
 create table dp_mm (key int) partitioned by (key1 string, key2 int) stored as orc
   tblproperties ("transactional"="true", "transactional_properties"="insert_only");
 
-insert into table dp_mm partition (key1='123', key2) select key, key from intermediate;
+insert into table dp_mm partition (key1='123', key2) select key, key from intermediate_n0;
 
 select * from dp_mm order by key;
 
@@ -69,35 +69,35 @@ drop table dp_mm;
 create table union_mm(id int)  tblproperties ("transactional"="true", "transactional_properties"="insert_only");
 insert into table union_mm 
 select temps.p from (
-select key as p from intermediate 
+select key as p from intermediate_n0 
 union all 
-select key + 1 as p from intermediate ) temps;
+select key + 1 as p from intermediate_n0 ) temps;
 
 select * from union_mm order by id;
 
 insert into table union_mm 
 select p from
 (
-select key + 1 as p from intermediate
+select key + 1 as p from intermediate_n0
 union all
-select key from intermediate
+select key from intermediate_n0
 ) tab group by p
 union all
-select key + 2 as p from intermediate;
+select key + 2 as p from intermediate_n0;
 
 select * from union_mm order by id;
 
 insert into table union_mm
 SELECT p FROM
 (
-  SELECT key + 1 as p FROM intermediate
+  SELECT key + 1 as p FROM intermediate_n0
   UNION ALL
   SELECT key as p FROM ( 
     SELECT distinct key FROM (
       SELECT key FROM (
-        SELECT key + 2 as key FROM intermediate
+        SELECT key + 2 as key FROM intermediate_n0
         UNION ALL
-        SELECT key FROM intermediate
+        SELECT key FROM intermediate_n0
       )t1 
     group by key)t2
   )t3
@@ -112,9 +112,9 @@ drop table union_mm;
 create table partunion_mm(id int) partitioned by (key int) tblproperties ("transactional"="true", "transactional_properties"="insert_only");
 insert into table partunion_mm partition(key)
 select temps.* from (
-select key as p, key from intermediate 
+select key as p, key from intermediate_n0 
 union all 
-select key + 1 as p, key + 1 from intermediate ) temps;
+select key + 1 as p, key + 1 from intermediate_n0 ) temps;
 
 select * from partunion_mm order by id;
 drop table partunion_mm;
@@ -127,7 +127,7 @@ create table skew_mm(k1 int, k2 int, k4 int) skewed by (k1, k4) on ((0,0),(1,1),
  stored as directories tblproperties ("transactional"="true", "transactional_properties"="insert_only");
 
 insert into table skew_mm 
-select key, key, key from intermediate;
+select key, key, key from intermediate_n0;
 
 select * from skew_mm order by k2, k1, k4;
 drop table skew_mm;
@@ -137,9 +137,9 @@ create table skew_dp_union_mm(k1 int, k2 int, k4 int) partitioned by (k3 int)
 skewed by (k1, k4) on ((0,0),(1,1),(2,2),(3,3)) stored as directories tblproperties ("transactional"="true", "transactional_properties"="insert_only");
 
 insert into table skew_dp_union_mm partition (k3)
-select key as i, key as j, key as k, key as l from intermediate
+select key as i, key as j, key as k, key as l from intermediate_n0
 union all 
-select key +1 as i, key +2 as j, key +3 as k, key +4 as l from intermediate;
+select key +1 as i, key +2 as j, key +3 as k, key +4 as l from intermediate_n0;
 
 
 select * from skew_dp_union_mm order by k2, k1, k4;
@@ -155,11 +155,11 @@ set hive.merge.mapredfiles=true;
 
 create table merge0_mm (id int) stored as orc tblproperties("transactional"="true", "transactional_properties"="insert_only");
 
-insert into table merge0_mm select key from intermediate;
+insert into table merge0_mm select key from intermediate_n0;
 select * from merge0_mm;
 
 set tez.grouping.split-count=1;
-insert into table merge0_mm select key from intermediate;
+insert into table merge0_mm select key from intermediate_n0;
 set tez.grouping.split-count=0;
 select * from merge0_mm;
 
@@ -168,11 +168,11 @@ drop table merge0_mm;
 
 create table merge2_mm (id int) tblproperties("transactional"="true", "transactional_properties"="insert_only");
 
-insert into table merge2_mm select key from intermediate;
+insert into table merge2_mm select key from intermediate_n0;
 select * from merge2_mm;
 
 set tez.grouping.split-count=1;
-insert into table merge2_mm select key from intermediate;
+insert into table merge2_mm select key from intermediate_n0;
 set tez.grouping.split-count=0;
 select * from merge2_mm;
 
@@ -181,11 +181,11 @@ drop table merge2_mm;
 
 create table merge1_mm (id int) partitioned by (key int) stored as orc tblproperties("transactional"="true", "transactional_properties"="insert_only");
 
-insert into table merge1_mm partition (key) select key, key from intermediate;
+insert into table merge1_mm partition (key) select key, key from intermediate_n0;
 select * from merge1_mm order by id, key;
 
 set tez.grouping.split-count=1;
-insert into table merge1_mm partition (key) select key, key from intermediate;
+insert into table merge1_mm partition (key) select key, key from intermediate_n0;
 set tez.grouping.split-count=0;
 select * from merge1_mm order by id, key;
 
@@ -199,13 +199,13 @@ set hive.merge.mapredfiles=false;
 
 
 drop table ctas0_mm;
-create table ctas0_mm tblproperties ("transactional"="true", "transactional_properties"="insert_only") as select * from intermediate;
+create table ctas0_mm tblproperties ("transactional"="true", "transactional_properties"="insert_only") as select * from intermediate_n0;
 select * from ctas0_mm;
 drop table ctas0_mm;
 
 drop table ctas1_mm;
 create table ctas1_mm tblproperties ("transactional"="true", "transactional_properties"="insert_only") as
-  select * from intermediate union all select * from intermediate;
+  select * from intermediate_n0 union all select * from intermediate_n0;
 select * from ctas1_mm;
 drop table ctas1_mm;
 
@@ -215,7 +215,7 @@ drop table multi0_2_mm;
 create table multi0_1_mm (key int, key2 int)  tblproperties("transactional"="true", "transactional_properties"="insert_only");
 create table multi0_2_mm (key int, key2 int)  tblproperties("transactional"="true", "transactional_properties"="insert_only");
 
-from intermediate
+from intermediate_n0
 insert overwrite table multi0_1_mm select key, p
 insert overwrite table multi0_2_mm select p, key;
 
@@ -226,7 +226,7 @@ set hive.merge.mapredfiles=true;
 set hive.merge.sparkfiles=true;
 set hive.merge.tezfiles=true;
 
-from intermediate
+from intermediate_n0
 insert into table multi0_1_mm select p, key
 insert overwrite table multi0_2_mm select key, p;
 
@@ -243,25 +243,25 @@ drop table multi0_2_mm;
 
 drop table multi1_mm;
 create table multi1_mm (key int, key2 int) partitioned by (p int) tblproperties("transactional"="true", "transactional_properties"="insert_only");
-from intermediate
+from intermediate_n0
 insert into table multi1_mm partition(p=1) select p, key
 insert into table multi1_mm partition(p=2) select key, p;
 
 select * from multi1_mm order by key, key2, p;
 
-from intermediate
+from intermediate_n0
 insert into table multi1_mm partition(p=2) select p, key
 insert overwrite table multi1_mm partition(p=1) select key, p;
 
 select * from multi1_mm order by key, key2, p;
 
-from intermediate
+from intermediate_n0
 insert into table multi1_mm partition(p) select p, key, p
 insert into table multi1_mm partition(p=1) select key, p;
 
 select key, key2, p from multi1_mm order by key, key2, p;
 
-from intermediate
+from intermediate_n0
 insert into table multi1_mm partition(p) select p, key, 1
 insert into table multi1_mm partition(p=1) select key, p;
 
@@ -277,11 +277,11 @@ set hive.stats.autogather=true;
 
 drop table stats_mm;
 create table stats_mm(key int)  tblproperties("transactional"="true", "transactional_properties"="insert_only");
---insert overwrite table stats_mm  select key from intermediate;
-insert into table stats_mm  select key from intermediate;
+--insert overwrite table stats_mm  select key from intermediate_n0;
+insert into table stats_mm  select key from intermediate_n0;
 desc formatted stats_mm;
 
-insert into table stats_mm  select key from intermediate;
+insert into table stats_mm  select key from intermediate_n0;
 desc formatted stats_mm;
 drop table stats_mm;
 
@@ -321,12 +321,12 @@ set hive.auto.convert.join=true;
 
 DROP TABLE IF EXISTS temp1;
 CREATE TEMPORARY TABLE temp1 (a int) TBLPROPERTIES ("transactional"="true", "transactional_properties"="insert_only");
-INSERT INTO temp1 SELECT key FROM intermediate;
+INSERT INTO temp1 SELECT key FROM intermediate_n0;
 DESC EXTENDED temp1;
 SELECT * FROM temp1;
 
 
-drop table intermediate;
+drop table intermediate_n0;
 
 
 

http://git-wip-us.apache.org/repos/asf/hive/blob/38d3b8e1/ql/src/test/queries/clientpositive/mm_buckets.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/mm_buckets.q b/ql/src/test/queries/clientpositive/mm_buckets.q
index 3e91cdd..12dc4fe 100644
--- a/ql/src/test/queries/clientpositive/mm_buckets.q
+++ b/ql/src/test/queries/clientpositive/mm_buckets.q
@@ -12,11 +12,11 @@ set hive.txn.manager=org.apache.hadoop.hive.ql.lockmgr.DbTxnManager;
 -- Bucketing tests are slow and some tablesample ones don't work w/o MM
 
 -- Force multiple writers when reading
-drop table intermediate;
-create table intermediate(key int) partitioned by (p int) stored as orc;
-insert into table intermediate partition(p='455') select distinct key from src where key >= 0 order by key desc limit 2;
-insert into table intermediate partition(p='456') select distinct key from src where key is not null order by key asc limit 2;
-insert into table intermediate partition(p='457') select distinct key from src where key >= 100 order by key asc limit 2;
+drop table intermediate_n2;
+create table intermediate_n2(key int) partitioned by (p int) stored as orc;
+insert into table intermediate_n2 partition(p='455') select distinct key from src where key >= 0 order by key desc limit 2;
+insert into table intermediate_n2 partition(p='456') select distinct key from src where key is not null order by key asc limit 2;
+insert into table intermediate_n2 partition(p='457') select distinct key from src where key >= 100 order by key asc limit 2;
 
 
 
@@ -24,11 +24,11 @@ drop table bucket0_mm;
 create table bucket0_mm(key int, id int)
 clustered by (key) into 2 buckets
 tblproperties("transactional"="true", "transactional_properties"="insert_only");
-insert into table bucket0_mm select key, key from intermediate;
+insert into table bucket0_mm select key, key from intermediate_n2;
 select * from bucket0_mm order by key, id;
 select * from bucket0_mm tablesample (bucket 1 out of 2) s;
 select * from bucket0_mm tablesample (bucket 2 out of 2) s;
-insert into table bucket0_mm select key, key from intermediate;
+insert into table bucket0_mm select key, key from intermediate_n2;
 select * from bucket0_mm order by key, id;
 select * from bucket0_mm tablesample (bucket 1 out of 2) s;
 select * from bucket0_mm tablesample (bucket 2 out of 2) s;
@@ -40,9 +40,9 @@ create table bucket1_mm(key int, id int) partitioned by (key2 int)
 clustered by (key) sorted by (key) into 2 buckets
 tblproperties("transactional"="true", "transactional_properties"="insert_only");
 insert into table bucket1_mm partition (key2)
-select key + 1, key, key - 1 from intermediate
+select key + 1, key, key - 1 from intermediate_n2
 union all 
-select key - 1, key, key + 1 from intermediate;
+select key - 1, key, key + 1 from intermediate_n2;
 select * from bucket1_mm order by key, id;
 select * from bucket1_mm tablesample (bucket 1 out of 2) s  order by key, id;
 select * from bucket1_mm tablesample (bucket 2 out of 2) s  order by key, id;
@@ -54,14 +54,14 @@ drop table bucket2_mm;
 create table bucket2_mm(key int, id int)
 clustered by (key) into 10 buckets
 tblproperties("transactional"="true", "transactional_properties"="insert_only");
-insert into table bucket2_mm select key, key from intermediate where key == 0;
+insert into table bucket2_mm select key, key from intermediate_n2 where key == 0;
 select * from bucket2_mm order by key, id;
 select * from bucket2_mm tablesample (bucket 1 out of 10) s order by key, id;
 select * from bucket2_mm tablesample (bucket 4 out of 10) s order by key, id;
-insert into table bucket2_mm select key, key from intermediate where key in (0, 103);
+insert into table bucket2_mm select key, key from intermediate_n2 where key in (0, 103);
 select * from bucket2_mm;
 select * from bucket2_mm tablesample (bucket 1 out of 10) s order by key, id;
 select * from bucket2_mm tablesample (bucket 4 out of 10) s order by key, id;
 drop table bucket2_mm;
 
-drop table intermediate;
\ No newline at end of file
+drop table intermediate_n2;
\ No newline at end of file

http://git-wip-us.apache.org/repos/asf/hive/blob/38d3b8e1/ql/src/test/queries/clientpositive/mm_cttas.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/mm_cttas.q b/ql/src/test/queries/clientpositive/mm_cttas.q
index da8b84d..8f1274c 100644
--- a/ql/src/test/queries/clientpositive/mm_cttas.q
+++ b/ql/src/test/queries/clientpositive/mm_cttas.q
@@ -3,19 +3,19 @@
 set hive.support.concurrency=true;
 set hive.txn.manager=org.apache.hadoop.hive.ql.lockmgr.DbTxnManager;
 
-drop table intermediate;
-create table intermediate(key int) partitioned by (p int) stored as orc;
-insert into table intermediate partition(p='455') select distinct key from src where key >= 0 order by key desc limit 2;
-insert into table intermediate partition(p='456') select distinct key from src where key is not null order by key asc limit 2;
-insert into table intermediate partition(p='457') select distinct key from src where key >= 100 order by key asc limit 2;
+drop table intermediate_n1;
+create table intermediate_n1(key int) partitioned by (p int) stored as orc;
+insert into table intermediate_n1 partition(p='455') select distinct key from src where key >= 0 order by key desc limit 2;
+insert into table intermediate_n1 partition(p='456') select distinct key from src where key is not null order by key asc limit 2;
+insert into table intermediate_n1 partition(p='457') select distinct key from src where key >= 100 order by key asc limit 2;
 
 drop table cttas1_mm;
 
-create temporary table cttas1_mm tblproperties ("transactional"="true", "transactional_properties"="insert_only") as select * from intermediate;
+create temporary table cttas1_mm tblproperties ("transactional"="true", "transactional_properties"="insert_only") as select * from intermediate_n1;
 
 select * from cttas1_mm;
 
 drop table cttas1_mm;
 
-drop table intermediate;
+drop table intermediate_n1;
 

http://git-wip-us.apache.org/repos/asf/hive/blob/38d3b8e1/ql/src/test/queries/clientpositive/mm_default.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/mm_default.q b/ql/src/test/queries/clientpositive/mm_default.q
index 0fa46a9..8e34bf3 100644
--- a/ql/src/test/queries/clientpositive/mm_default.q
+++ b/ql/src/test/queries/clientpositive/mm_default.q
@@ -14,10 +14,10 @@ create table acid0 (key string) stored as ORC  tblproperties("transactional"="tr
 set hive.create.as.insert.only=true;
 create table mm1 like non_mm0;
 create table mm2 like mm0;
-create table acid1 like acid0;
+create table acid1_n0 like acid0;
 create table mm3 as select key from src limit 1;
 create table mm4 (key string);
-create table acid2 (key string) stored as ORC  tblproperties("transactional"="true");
+create table acid2_n0 (key string) stored as ORC  tblproperties("transactional"="true");
 create table non_mm1 tblproperties("transactional"="false")
  as select key from src limit 1;
 
@@ -28,8 +28,8 @@ desc formatted mm2;
 desc formatted mm3;
 desc formatted mm4;
 desc formatted non_mm1;
-desc formatted acid1;
-desc formatted acid2;
+desc formatted acid1_n0;
+desc formatted acid2_n0;
 
 
 drop table non_mm0;
@@ -40,8 +40,8 @@ drop table mm2;
 drop table mm3;
 drop table mm4;
 drop table acid0;
-drop table acid1;
-drop table acid2;
+drop table acid1_n0;
+drop table acid2_n0;
 
 
 

http://git-wip-us.apache.org/repos/asf/hive/blob/38d3b8e1/ql/src/test/queries/clientpositive/mm_exim.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/mm_exim.q b/ql/src/test/queries/clientpositive/mm_exim.q
index a2b6e08..9870bf4 100644
--- a/ql/src/test/queries/clientpositive/mm_exim.q
+++ b/ql/src/test/queries/clientpositive/mm_exim.q
@@ -10,27 +10,27 @@ set hive.support.concurrency=true;
 set hive.txn.manager=org.apache.hadoop.hive.ql.lockmgr.DbTxnManager;
 
 
-drop table intermediate;
-create table intermediate(key int) partitioned by (p int) stored as orc tblproperties("transactional"="false");
-insert into table intermediate partition(p='455') select distinct key from src where key >= 0 order by key desc limit 2;
-insert into table intermediate partition(p='456') select distinct key from src where key is not null order by key asc limit 2;
-insert into table intermediate partition(p='457') select distinct key from src where key >= 100 order by key asc limit 2;
+drop table intermediate_n4;
+create table intermediate_n4(key int) partitioned by (p int) stored as orc tblproperties("transactional"="false");
+insert into table intermediate_n4 partition(p='455') select distinct key from src where key >= 0 order by key desc limit 2;
+insert into table intermediate_n4 partition(p='456') select distinct key from src where key is not null order by key asc limit 2;
+insert into table intermediate_n4 partition(p='457') select distinct key from src where key >= 100 order by key asc limit 2;
 
 drop table intermediate_nonpart;
 drop table intermmediate_part;
 drop table intermmediate_nonpart;
 create table intermediate_nonpart(key int, p int) tblproperties("transactional"="false");
-insert into intermediate_nonpart select * from intermediate;
+insert into intermediate_nonpart select * from intermediate_n4;
 create table intermmediate_nonpart(key int, p int) tblproperties("transactional"="true", "transactional_properties"="insert_only");
-insert into intermmediate_nonpart select * from intermediate;
+insert into intermmediate_nonpart select * from intermediate_n4;
 create table intermmediate(key int) partitioned by (p int) tblproperties("transactional"="true", "transactional_properties"="insert_only");
-insert into table intermmediate partition(p) select key, p from intermediate;
+insert into table intermmediate partition(p) select key, p from intermediate_n4;
 
 set hive.exim.test.mode=true;
 
 export table intermediate_nonpart to 'ql/test/data/exports/intermediate_nonpart';
 export table intermmediate_nonpart to 'ql/test/data/exports/intermmediate_nonpart';
-export table intermediate to 'ql/test/data/exports/intermediate_part';
+export table intermediate_n4 to 'ql/test/data/exports/intermediate_part';
 export table intermmediate to 'ql/test/data/exports/intermmediate_part';
 
 drop table intermediate_nonpart;

http://git-wip-us.apache.org/repos/asf/hive/blob/38d3b8e1/ql/src/test/queries/clientpositive/msck_repair_0.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/msck_repair_0.q b/ql/src/test/queries/clientpositive/msck_repair_0.q
index cb291fe..aeb4820 100644
--- a/ql/src/test/queries/clientpositive/msck_repair_0.q
+++ b/ql/src/test/queries/clientpositive/msck_repair_0.q
@@ -1,36 +1,36 @@
 set hive.msck.repair.batch.size=1;
 set hive.mv.files.thread=0;
 
-DROP TABLE IF EXISTS repairtable;
+DROP TABLE IF EXISTS repairtable_n5;
 
-CREATE TABLE repairtable(col STRING) PARTITIONED BY (p1 STRING, p2 STRING);
+CREATE TABLE repairtable_n5(col STRING) PARTITIONED BY (p1 STRING, p2 STRING);
 
-MSCK TABLE repairtable;
+MSCK TABLE repairtable_n5;
 
-show partitions repairtable;
+show partitions repairtable_n5;
 
 dfs ${system:test.dfs.mkdir} ${system:test.warehouse.dir}/repairtable/p1=c/p2=a/p3=b;
 dfs -touchz ${system:test.warehouse.dir}/repairtable/p1=c/p2=a/p3=b/datafile;
 
-MSCK TABLE default.repairtable;
+MSCK TABLE default.repairtable_n5;
 
-show partitions default.repairtable;
+show partitions default.repairtable_n5;
 
-MSCK REPAIR TABLE default.repairtable;
+MSCK REPAIR TABLE default.repairtable_n5;
 
-show partitions default.repairtable;
+show partitions default.repairtable_n5;
 
-MSCK TABLE repairtable;
+MSCK TABLE repairtable_n5;
 
-show partitions repairtable;
+show partitions repairtable_n5;
 
 set hive.mapred.mode=strict;
 
 dfs ${system:test.dfs.mkdir} ${system:test.warehouse.dir}/repairtable/p1=e/p2=f/p3=g;
 dfs -touchz ${system:test.warehouse.dir}/repairtable/p1=e/p2=f/p3=g/datafile;
 
-MSCK REPAIR TABLE default.repairtable;
+MSCK REPAIR TABLE default.repairtable_n5;
 
-show partitions default.repairtable;
+show partitions default.repairtable_n5;
 
-DROP TABLE default.repairtable;
+DROP TABLE default.repairtable_n5;

http://git-wip-us.apache.org/repos/asf/hive/blob/38d3b8e1/ql/src/test/queries/clientpositive/msck_repair_2.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/msck_repair_2.q b/ql/src/test/queries/clientpositive/msck_repair_2.q
index 77785e9..be745b2 100644
--- a/ql/src/test/queries/clientpositive/msck_repair_2.q
+++ b/ql/src/test/queries/clientpositive/msck_repair_2.q
@@ -1,25 +1,25 @@
 set hive.msck.repair.batch.size=1;
 set hive.msck.path.validation=skip;
 
-DROP TABLE IF EXISTS repairtable;
+DROP TABLE IF EXISTS repairtable_n2;
 
-CREATE TABLE repairtable(col STRING) PARTITIONED BY (p1 STRING, p2 STRING);
+CREATE TABLE repairtable_n2(col STRING) PARTITIONED BY (p1 STRING, p2 STRING);
 
-MSCK TABLE repairtable;
+MSCK TABLE repairtable_n2;
 
-show partitions repairtable;
+show partitions repairtable_n2;
 
 dfs ${system:test.dfs.mkdir} ${system:test.warehouse.dir}/repairtable/p1=c/p2=a/p3=b;
 dfs -touchz ${system:test.warehouse.dir}/repairtable/p1=c/p2=a/p3=b/datafile;
 dfs -touchz ${system:test.warehouse.dir}/repairtable/p1=c/datafile;
 
-MSCK TABLE default.repairtable;
-show partitions repairtable;
+MSCK TABLE default.repairtable_n2;
+show partitions repairtable_n2;
 
-MSCK REPAIR TABLE default.repairtable;
-show partitions repairtable;
+MSCK REPAIR TABLE default.repairtable_n2;
+show partitions repairtable_n2;
 
-MSCK TABLE repairtable;
-show partitions repairtable;
+MSCK TABLE repairtable_n2;
+show partitions repairtable_n2;
 
-DROP TABLE default.repairtable;
+DROP TABLE default.repairtable_n2;

http://git-wip-us.apache.org/repos/asf/hive/blob/38d3b8e1/ql/src/test/queries/clientpositive/msck_repair_3.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/msck_repair_3.q b/ql/src/test/queries/clientpositive/msck_repair_3.q
index f42443f..140a690 100644
--- a/ql/src/test/queries/clientpositive/msck_repair_3.q
+++ b/ql/src/test/queries/clientpositive/msck_repair_3.q
@@ -1,21 +1,21 @@
 set hive.msck.repair.batch.size=1;
 
-DROP TABLE IF EXISTS repairtable;
+DROP TABLE IF EXISTS repairtable_n3;
 
-CREATE TABLE repairtable(col STRING) PARTITIONED BY (p1 STRING, p2 STRING);
+CREATE TABLE repairtable_n3(col STRING) PARTITIONED BY (p1 STRING, p2 STRING);
 
-MSCK TABLE repairtable;
-show partitions repairtable;
+MSCK TABLE repairtable_n3;
+show partitions repairtable_n3;
 
 dfs ${system:test.dfs.mkdir} ${system:test.warehouse.dir}/repairtable/p1=c/p2=a/p3=b;
 
-MSCK TABLE default.repairtable;
-show partitions repairtable;
+MSCK TABLE default.repairtable_n3;
+show partitions repairtable_n3;
 
-MSCK REPAIR TABLE default.repairtable;
-show partitions repairtable;
+MSCK REPAIR TABLE default.repairtable_n3;
+show partitions repairtable_n3;
 
-MSCK TABLE repairtable;
-show partitions repairtable;
+MSCK TABLE repairtable_n3;
+show partitions repairtable_n3;
 
-DROP TABLE default.repairtable;
+DROP TABLE default.repairtable_n3;

http://git-wip-us.apache.org/repos/asf/hive/blob/38d3b8e1/ql/src/test/queries/clientpositive/msck_repair_batchsize.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/msck_repair_batchsize.q b/ql/src/test/queries/clientpositive/msck_repair_batchsize.q
index a44c00e..5a7afcc 100644
--- a/ql/src/test/queries/clientpositive/msck_repair_batchsize.q
+++ b/ql/src/test/queries/clientpositive/msck_repair_batchsize.q
@@ -1,10 +1,10 @@
 set hive.msck.repair.batch.size=2;
 
-DROP TABLE IF EXISTS repairtable;
+DROP TABLE IF EXISTS repairtable_n0;
 
-CREATE TABLE repairtable(col STRING) PARTITIONED BY (p1 STRING, p2 STRING);
+CREATE TABLE repairtable_n0(col STRING) PARTITIONED BY (p1 STRING, p2 STRING);
 
-MSCK TABLE repairtable;
+MSCK TABLE repairtable_n0;
 
 dfs ${system:test.dfs.mkdir} ${system:test.warehouse.dir}/repairtable/p1=a/p2=a;
 dfs ${system:test.dfs.mkdir} ${system:test.warehouse.dir}/repairtable/p1=b/p2=a;
@@ -13,24 +13,24 @@ dfs -touchz ${system:test.warehouse.dir}/repairtable/p1=a/p2=a/datafile;
 dfs -touchz ${system:test.warehouse.dir}/repairtable/p1=b/p2=a/datafile;
 dfs -touchz ${system:test.warehouse.dir}/repairtable/p1=c/p2=a/datafile;
 
-MSCK TABLE default.repairtable;
-show partitions default.repairtable;
+MSCK TABLE default.repairtable_n0;
+show partitions default.repairtable_n0;
 
-MSCK REPAIR TABLE default.repairtable;
-show partitions default.repairtable;
+MSCK REPAIR TABLE default.repairtable_n0;
+show partitions default.repairtable_n0;
 
-MSCK TABLE repairtable;
-show partitions repairtable;
+MSCK TABLE repairtable_n0;
+show partitions repairtable_n0;
 
-DROP TABLE default.repairtable;
+DROP TABLE default.repairtable_n0;
 
 
 dfs  ${system:test.dfs.mkdir} -p ${system:test.tmp.dir}/apps/hive/warehouse/test.db/repairtable/p1=c/p2=a/p3=b;
-CREATE TABLE `repairtable`( `col` string) PARTITIONED BY (  `p1` string,  `p2` string) location '${system:test.tmp.dir}/apps/hive/warehouse/test.db/repairtable/';
+CREATE TABLE `repairtable_n0`( `col` string) PARTITIONED BY (  `p1` string,  `p2` string) location '${system:test.tmp.dir}/apps/hive/warehouse/test.db/repairtable/';
 
 dfs -touchz ${system:test.tmp.dir}/apps/hive/warehouse/test.db/repairtable/p1=c/p2=a/p3=b/datafile;
 set hive.mv.files.thread=1;
-MSCK TABLE repairtable;
-show partitions repairtable;
+MSCK TABLE repairtable_n0;
+show partitions repairtable_n0;
 
-DROP TABLE default.repairtable;
+DROP TABLE default.repairtable_n0;

http://git-wip-us.apache.org/repos/asf/hive/blob/38d3b8e1/ql/src/test/queries/clientpositive/msck_repair_drop.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/msck_repair_drop.q b/ql/src/test/queries/clientpositive/msck_repair_drop.q
index bc14d98..9923fb5 100644
--- a/ql/src/test/queries/clientpositive/msck_repair_drop.q
+++ b/ql/src/test/queries/clientpositive/msck_repair_drop.q
@@ -1,9 +1,9 @@
 set hive.mv.files.thread=0;
-DROP TABLE IF EXISTS repairtable;
+DROP TABLE IF EXISTS repairtable_n1;
 
-CREATE TABLE repairtable(col STRING) PARTITIONED BY (p1 STRING, p2 STRING);
+CREATE TABLE repairtable_n1(col STRING) PARTITIONED BY (p1 STRING, p2 STRING);
 
--- repairtable will have partitions created with part keys p1=1, p1=2, p1=3, p1=4 and p1=5
+-- repairtable_n1 will have partitions created with part keys p1=1, p1=2, p1=3, p1=4 and p1=5
 -- p1=2 will be used to test drop in 3 tests
 --   1.  each partition is dropped individually: set hive.msck.repair.batch.size=1;
 --   2.  partition are dropped in groups of 3:  set hive.msck.repair.batch.size=3;
@@ -37,20 +37,20 @@ dfs -touchz ${system:test.warehouse.dir}/repairtable/p1=2/p2=29/p3=291/datafile;
 dfs ${system:test.dfs.mkdir} ${system:test.warehouse.dir}/repairtable/p1=2/p2=210/p3=2101;
 dfs -touchz ${system:test.warehouse.dir}/repairtable/p1=2/p2=210/p3=2101/datafile;
 
-MSCK TABLE default.repairtable;
-MSCK REPAIR TABLE default.repairtable;
+MSCK TABLE default.repairtable_n1;
+MSCK REPAIR TABLE default.repairtable_n1;
 
 -- Now all 12 partitions are in
-show partitions default.repairtable;
+show partitions default.repairtable_n1;
 
 -- Remove all p1=2 partitions from file system
 dfs -rmr ${system:test.warehouse.dir}/repairtable/p1=2;
 
 -- test 1: each partition is dropped individually
 set hive.msck.repair.batch.size=1;
-MSCK TABLE default.repairtable DROP PARTITIONS;
-MSCK REPAIR TABLE default.repairtable DROP PARTITIONS;
-show partitions default.repairtable;
+MSCK TABLE default.repairtable_n1 DROP PARTITIONS;
+MSCK REPAIR TABLE default.repairtable_n1 DROP PARTITIONS;
+show partitions default.repairtable_n1;
 
 -- Recreate p1=2 partitions
 dfs ${system:test.dfs.mkdir} ${system:test.warehouse.dir}/repairtable/p1=2/p2=21/p3=211;
@@ -74,20 +74,20 @@ dfs -touchz ${system:test.warehouse.dir}/repairtable/p1=2/p2=29/p3=291/datafile;
 dfs ${system:test.dfs.mkdir} ${system:test.warehouse.dir}/repairtable/p1=2/p2=210/p3=2101;
 dfs -touchz ${system:test.warehouse.dir}/repairtable/p1=2/p2=210/p3=2101/datafile;
 
-MSCK TABLE default.repairtable;
-MSCK REPAIR TABLE default.repairtable;
+MSCK TABLE default.repairtable_n1;
+MSCK REPAIR TABLE default.repairtable_n1;
 
 -- Now all 12 partitions are in
-show partitions default.repairtable;
+show partitions default.repairtable_n1;
 
 -- Remove all p1=2 partitions from file system
 dfs -rmr ${system:test.warehouse.dir}/repairtable/p1=2;
 
 -- test 2: partition are dropped in groups of 3
 set hive.msck.repair.batch.size=3;
-MSCK TABLE default.repairtable DROP PARTITIONS;
-MSCK REPAIR TABLE default.repairtable DROP PARTITIONS;
-show partitions default.repairtable;
+MSCK TABLE default.repairtable_n1 DROP PARTITIONS;
+MSCK REPAIR TABLE default.repairtable_n1 DROP PARTITIONS;
+show partitions default.repairtable_n1;
 
 -- Recreate p1=2 partitions
 dfs ${system:test.dfs.mkdir} ${system:test.warehouse.dir}/repairtable/p1=2/p2=21/p3=211;
@@ -111,20 +111,20 @@ dfs -touchz ${system:test.warehouse.dir}/repairtable/p1=2/p2=29/p3=291/datafile;
 dfs ${system:test.dfs.mkdir} ${system:test.warehouse.dir}/repairtable/p1=2/p2=210/p3=2101;
 dfs -touchz ${system:test.warehouse.dir}/repairtable/p1=2/p2=210/p3=2101/datafile;
 
-MSCK TABLE default.repairtable;
-MSCK REPAIR TABLE default.repairtable;
+MSCK TABLE default.repairtable_n1;
+MSCK REPAIR TABLE default.repairtable_n1;
 
 -- Now all 12 partitions are in
-show partitions default.repairtable;
+show partitions default.repairtable_n1;
 
 -- Remove all p1=2 partitions from file system
 dfs -rmr ${system:test.warehouse.dir}/repairtable/p1=2;
 
 --  test 3.  all partitions are dropped in 1 shot
 set hive.msck.repair.batch.size=0;
-MSCK TABLE default.repairtable DROP PARTITIONS;
-MSCK REPAIR TABLE default.repairtable DROP PARTITIONS;
-show partitions default.repairtable;
+MSCK TABLE default.repairtable_n1 DROP PARTITIONS;
+MSCK REPAIR TABLE default.repairtable_n1 DROP PARTITIONS;
+show partitions default.repairtable_n1;
 
 -- test add parition keyword: begin
 dfs ${system:test.dfs.mkdir} ${system:test.warehouse.dir}/repairtable/p1=3/p2=31/p3=311;
@@ -132,9 +132,9 @@ dfs -touchz ${system:test.warehouse.dir}/repairtable/p1=3/p2=31/p3=311/datafile;
 dfs ${system:test.dfs.mkdir} ${system:test.warehouse.dir}/repairtable/p1=3/p2=32/p3=321;
 dfs -touchz ${system:test.warehouse.dir}/repairtable/p1=3/p2=32/p3=321/datafile;
 
-MSCK TABLE default.repairtable;
-MSCK REPAIR TABLE default.repairtable;
-show partitions default.repairtable;
+MSCK TABLE default.repairtable_n1;
+MSCK REPAIR TABLE default.repairtable_n1;
+show partitions default.repairtable_n1;
 
 -- Create p1=4 in filesystem
 dfs ${system:test.dfs.mkdir} ${system:test.warehouse.dir}/repairtable/p1=4/p2=41/p3=411;
@@ -148,9 +148,9 @@ dfs -rmr ${system:test.warehouse.dir}/repairtable/p1=3;
 -- Status: p1=3 dropped from filesystem, but exists in metastore
 --         p1=4 exists in filesystem but not in metastore
 -- test add partition: only brings in p1=4 and doesn't remove p1=3
-MSCK TABLE default.repairtable ADD PARTITIONS;
-MSCK REPAIR TABLE default.repairtable ADD PARTITIONS;
-show partitions default.repairtable;
+MSCK TABLE default.repairtable_n1 ADD PARTITIONS;
+MSCK REPAIR TABLE default.repairtable_n1 ADD PARTITIONS;
+show partitions default.repairtable_n1;
 -- test add partition keyword: end
 
 -- test drop partition keyword: begin
@@ -162,9 +162,9 @@ dfs -touchz ${system:test.warehouse.dir}/repairtable/p1=5/p2=52/p3=521/datafile;
 -- Status: p1=3 removed from filesystem, but exists in metastore (as part of add test)
 --         p1=5 exists in filesystem but not in metastore
 -- test drop partition: only removes p1=3 from metastore but doesn't update metadata for p1=5
-MSCK TABLE default.repairtable DROP PARTITIONS;
-MSCK REPAIR TABLE default.repairtable DROP PARTITIONS;
-show partitions default.repairtable;
+MSCK TABLE default.repairtable_n1 DROP PARTITIONS;
+MSCK REPAIR TABLE default.repairtable_n1 DROP PARTITIONS;
+show partitions default.repairtable_n1;
 -- test drop partition keyword: end
 
 -- test sync partition keyword: begin
@@ -174,7 +174,7 @@ dfs -rmr ${system:test.warehouse.dir}/repairtable/p1=4;
 -- Status: p1=4 dropped from filesystem, but exists in metastore
 --         p1=5 exists in filesystem but not in metastore (as part of drop test)
 -- test sync partition: removes p1=4 from metastore and updates metadata for p1=5
-MSCK TABLE default.repairtable SYNC PARTITIONS;
-MSCK REPAIR TABLE default.repairtable SYNC PARTITIONS;
-show partitions default.repairtable;
+MSCK TABLE default.repairtable_n1 SYNC PARTITIONS;
+MSCK REPAIR TABLE default.repairtable_n1 SYNC PARTITIONS;
+show partitions default.repairtable_n1;
 -- test sync partition keyword: end

http://git-wip-us.apache.org/repos/asf/hive/blob/38d3b8e1/ql/src/test/queries/clientpositive/multiMapJoin2.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/multiMapJoin2.q b/ql/src/test/queries/clientpositive/multiMapJoin2.q
index efaadcb..166cb09 100644
--- a/ql/src/test/queries/clientpositive/multiMapJoin2.q
+++ b/ql/src/test/queries/clientpositive/multiMapJoin2.q
@@ -179,18 +179,18 @@ GROUP BY tmp1.key
 ORDER BY key, cnt;
 
 -- Check if we can correctly handle partitioned table.
-CREATE TABLE part_table(key string, value string) PARTITIONED BY (partitionId int);
-INSERT OVERWRITE TABLE part_table PARTITION (partitionId=1)
+CREATE TABLE part_table_n0(key string, value string) PARTITIONED BY (partitionId int);
+INSERT OVERWRITE TABLE part_table_n0 PARTITION (partitionId=1)
   SELECT key, value FROM src ORDER BY key, value LIMIT 100;
-INSERT OVERWRITE TABLE part_table PARTITION (partitionId=2)
+INSERT OVERWRITE TABLE part_table_n0 PARTITION (partitionId=2)
   SELECT key, value FROM src1 ORDER BY key, value;
 
 EXPLAIN
 SELECT count(*)
-FROM part_table x JOIN src1 y ON (x.key = y.key);
+FROM part_table_n0 x JOIN src1 y ON (x.key = y.key);
 
 SELECT count(*)
-FROM part_table x JOIN src1 y ON (x.key = y.key);
+FROM part_table_n0 x JOIN src1 y ON (x.key = y.key);
 
 set hive.auto.convert.join.noconditionaltask.size=10000000;
 set hive.optimize.correlation=false;

http://git-wip-us.apache.org/repos/asf/hive/blob/38d3b8e1/ql/src/test/queries/clientpositive/multi_column_in.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/multi_column_in.q b/ql/src/test/queries/clientpositive/multi_column_in.q
index 288406c..44fb1d9 100644
--- a/ql/src/test/queries/clientpositive/multi_column_in.q
+++ b/ql/src/test/queries/clientpositive/multi_column_in.q
@@ -1,72 +1,72 @@
 set hive.mapred.mode=nonstrict;
-drop table emps;
+drop table emps_n1;
 
-create table emps (empno int, deptno int, empname string);
+create table emps_n1 (empno int, deptno int, empname string);
 
-insert into table emps values (1,2,"11"),(1,2,"11"),(3,4,"33"),(1,3,"11"),(2,5,"22"),(2,5,"22");
+insert into table emps_n1 values (1,2,"11"),(1,2,"11"),(3,4,"33"),(1,3,"11"),(2,5,"22"),(2,5,"22");
 
-select * from emps;
+select * from emps_n1;
 
-select * from emps where (int(empno+deptno/2), int(deptno/3)) in ((2,0),(3,2));
+select * from emps_n1 where (int(empno+deptno/2), int(deptno/3)) in ((2,0),(3,2));
 
-select * from emps where (int(empno+deptno/2), int(deptno/3)) not in ((2,0),(3,2));
+select * from emps_n1 where (int(empno+deptno/2), int(deptno/3)) not in ((2,0),(3,2));
 
-select * from emps where (empno,deptno) in ((1,2),(3,2));
+select * from emps_n1 where (empno,deptno) in ((1,2),(3,2));
 
-select * from emps where (empno,deptno) not in ((1,2),(3,2));
+select * from emps_n1 where (empno,deptno) not in ((1,2),(3,2));
 
-select * from emps where (empno,deptno) in ((1,2),(1,3));
+select * from emps_n1 where (empno,deptno) in ((1,2),(1,3));
 
-select * from emps where (empno,deptno) not in ((1,2),(1,3));
+select * from emps_n1 where (empno,deptno) not in ((1,2),(1,3));
 
 explain
-select * from emps where (empno+1,deptno) in ((1,2),(3,2));
+select * from emps_n1 where (empno+1,deptno) in ((1,2),(3,2));
 
 explain 
-select * from emps where (empno+1,deptno) not in ((1,2),(3,2));
+select * from emps_n1 where (empno+1,deptno) not in ((1,2),(3,2));
 
-select * from emps where empno in (1,2);
+select * from emps_n1 where empno in (1,2);
 
-select * from emps where empno in (1,2) and deptno > 2;
+select * from emps_n1 where empno in (1,2) and deptno > 2;
 
-select * from emps where (empno) in (1,2) and deptno > 2;
+select * from emps_n1 where (empno) in (1,2) and deptno > 2;
 
-select * from emps where ((empno) in (1,2) and deptno > 2);
+select * from emps_n1 where ((empno) in (1,2) and deptno > 2);
 
-explain select * from emps where ((empno*2)|1,deptno) in ((empno+1,2),(empno+2,2));
+explain select * from emps_n1 where ((empno*2)|1,deptno) in ((empno+1,2),(empno+2,2));
 
-select * from emps where ((empno*2)|1,deptno) in ((empno+1,2),(empno+2,2));
+select * from emps_n1 where ((empno*2)|1,deptno) in ((empno+1,2),(empno+2,2));
 
-select (empno*2)|1,substr(empname,1,1) from emps;
+select (empno*2)|1,substr(empname,1,1) from emps_n1;
 
-select * from emps where ((empno*2)|1,substr(empname,1,1)) in ((empno+1,'2'),(empno+2,'2'));
+select * from emps_n1 where ((empno*2)|1,substr(empname,1,1)) in ((empno+1,'2'),(empno+2,'2'));
 
-select * from emps where ((empno*2)|1,substr(empname,1,1)) not in ((empno+1,'2'),(empno+2,'2'));
+select * from emps_n1 where ((empno*2)|1,substr(empname,1,1)) not in ((empno+1,'2'),(empno+2,'2'));
 
-select * from emps where ((empno*2)|1,substr(empname,1,1)) in ((empno+1,'2'),(empno+3,'2'));
+select * from emps_n1 where ((empno*2)|1,substr(empname,1,1)) in ((empno+1,'2'),(empno+3,'2'));
 
-select * from emps where ((empno*2)|1,substr(empname,1,1)) not in ((empno+1,'2'),(empno+3,'2'));
+select * from emps_n1 where ((empno*2)|1,substr(empname,1,1)) not in ((empno+1,'2'),(empno+3,'2'));
 
 
-select sum(empno), empname from emps where ((empno*2)|1,substr(empname,1,1)) in ((empno+1,'2'),(empno+3,'2'))
+select sum(empno), empname from emps_n1 where ((empno*2)|1,substr(empname,1,1)) in ((empno+1,'2'),(empno+3,'2'))
 group by empname;
 
-select * from emps where ((empno*2)|1,substr(empname,1,1)) in ((empno+1,'2'),(empno+3,'2'))
+select * from emps_n1 where ((empno*2)|1,substr(empname,1,1)) in ((empno+1,'2'),(empno+3,'2'))
 union
-select * from emps where (empno,deptno) in ((1,2),(3,2));
+select * from emps_n1 where (empno,deptno) in ((1,2),(3,2));
 
-drop view v;
+drop view v_n2;
 
-create view v as 
+create view v_n2 as 
 select * from(
-select * from emps where ((empno*2)|1,substr(empname,1,1)) in ((empno+1,'2'),(empno+3,'2'))
+select * from emps_n1 where ((empno*2)|1,substr(empname,1,1)) in ((empno+1,'2'),(empno+3,'2'))
 union
-select * from emps where (empno,deptno) in ((1,2),(3,2)))subq order by empno desc;
+select * from emps_n1 where (empno,deptno) in ((1,2),(3,2)))subq order by empno desc;
 
-select * from v;
+select * from v_n2;
 
 select subq.e1 from 
-(select (empno*2)|1 as e1, substr(empname,1,1) as n1 from emps)subq
+(select (empno*2)|1 as e1, substr(empname,1,1) as n1 from emps_n1)subq
 join
-(select empno as e2 from emps where ((empno*2)|1,substr(empname,1,1)) in ((empno+1,'2'),(empno+3,'2')))subq2
+(select empno as e2 from emps_n1 where ((empno*2)|1,substr(empname,1,1)) in ((empno+1,'2'),(empno+3,'2')))subq2
 on e1=e2+1;

http://git-wip-us.apache.org/repos/asf/hive/blob/38d3b8e1/ql/src/test/queries/clientpositive/multi_column_in_single.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/multi_column_in_single.q b/ql/src/test/queries/clientpositive/multi_column_in_single.q
index 7151fc4..32a2167 100644
--- a/ql/src/test/queries/clientpositive/multi_column_in_single.q
+++ b/ql/src/test/queries/clientpositive/multi_column_in_single.q
@@ -3,66 +3,66 @@ set hive.mapred.mode=nonstrict;
 
 select * from src where (key, value) in (('238','val_238'));
 
-drop table emps;
+drop table emps_n7;
 
-create table emps (empno int, deptno int, empname string);
+create table emps_n7 (empno int, deptno int, empname string);
 
-insert into table emps values (1,2,"11"),(1,2,"11"),(3,4,"33"),(1,3,"11"),(2,5,"22"),(2,5,"22");
+insert into table emps_n7 values (1,2,"11"),(1,2,"11"),(3,4,"33"),(1,3,"11"),(2,5,"22"),(2,5,"22");
 
-select * from emps;
+select * from emps_n7;
 
-select * from emps where (int(empno+deptno/2), int(deptno/3)) in ((3,2));
+select * from emps_n7 where (int(empno+deptno/2), int(deptno/3)) in ((3,2));
 
-select * from emps where (int(empno+deptno/2), int(deptno/3)) not in ((3,2));
+select * from emps_n7 where (int(empno+deptno/2), int(deptno/3)) not in ((3,2));
 
-select * from emps where (empno,deptno) in ((3,2));
+select * from emps_n7 where (empno,deptno) in ((3,2));
 
-select * from emps where (empno,deptno) not in ((3,2));
+select * from emps_n7 where (empno,deptno) not in ((3,2));
 
-select * from emps where (empno,deptno) in ((1,3));
+select * from emps_n7 where (empno,deptno) in ((1,3));
 
-select * from emps where (empno,deptno) not in ((1,3));
+select * from emps_n7 where (empno,deptno) not in ((1,3));
 
 explain
-select * from emps where (empno+1,deptno) in ((3,2));
+select * from emps_n7 where (empno+1,deptno) in ((3,2));
 
 explain 
-select * from emps where (empno+1,deptno) not in ((3,2));
+select * from emps_n7 where (empno+1,deptno) not in ((3,2));
 
-explain select * from emps where ((empno*2)|1,deptno) in ((empno+2,2));
+explain select * from emps_n7 where ((empno*2)|1,deptno) in ((empno+2,2));
 
-select * from emps where ((empno*2)|1,deptno) in ((empno+2,2));
+select * from emps_n7 where ((empno*2)|1,deptno) in ((empno+2,2));
 
-select (empno*2)|1,substr(empname,1,1) from emps;
+select (empno*2)|1,substr(empname,1,1) from emps_n7;
 
-select * from emps where ((empno*2)|1,substr(empname,1,1)) in ((empno+2,'2'));
+select * from emps_n7 where ((empno*2)|1,substr(empname,1,1)) in ((empno+2,'2'));
 
-select * from emps where ((empno*2)|1,substr(empname,1,1)) not in ((empno+2,'2'));
+select * from emps_n7 where ((empno*2)|1,substr(empname,1,1)) not in ((empno+2,'2'));
 
-select * from emps where ((empno*2)|1,substr(empname,1,1)) in ((empno+3,'2'));
+select * from emps_n7 where ((empno*2)|1,substr(empname,1,1)) in ((empno+3,'2'));
 
-select * from emps where ((empno*2)|1,substr(empname,1,1)) not in ((empno+3,'2'));
+select * from emps_n7 where ((empno*2)|1,substr(empname,1,1)) not in ((empno+3,'2'));
 
 
-select sum(empno), empname from emps where ((empno*2)|1,substr(empname,1,1)) in ((empno+3,'2'))
+select sum(empno), empname from emps_n7 where ((empno*2)|1,substr(empname,1,1)) in ((empno+3,'2'))
 group by empname;
 
-select * from emps where ((empno*2)|1,substr(empname,1,1)) in ((empno+3,'2'))
+select * from emps_n7 where ((empno*2)|1,substr(empname,1,1)) in ((empno+3,'2'))
 union
-select * from emps where (empno,deptno) in ((3,2));
+select * from emps_n7 where (empno,deptno) in ((3,2));
 
-drop view v;
+drop view v_n11;
 
-create view v as 
+create view v_n11 as 
 select * from(
-select * from emps where ((empno*2)|1,substr(empname,1,1)) in ((empno+3,'2'))
+select * from emps_n7 where ((empno*2)|1,substr(empname,1,1)) in ((empno+3,'2'))
 union
-select * from emps where (empno,deptno) in ((3,2)))subq order by empno desc;
+select * from emps_n7 where (empno,deptno) in ((3,2)))subq order by empno desc;
 
-select * from v;
+select * from v_n11;
 
 select subq.e1 from 
-(select (empno*2)|1 as e1, substr(empname,1,1) as n1 from emps)subq
+(select (empno*2)|1 as e1, substr(empname,1,1) as n1 from emps_n7)subq
 join
-(select empno as e2 from emps where ((empno*2)|1,substr(empname,1,1)) in ((empno+3,'2')))subq2
+(select empno as e2 from emps_n7 where ((empno*2)|1,substr(empname,1,1)) in ((empno+3,'2')))subq2
 on e1=e2+1;

http://git-wip-us.apache.org/repos/asf/hive/blob/38d3b8e1/ql/src/test/queries/clientpositive/multi_count_distinct.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/multi_count_distinct.q b/ql/src/test/queries/clientpositive/multi_count_distinct.q
index 038be3e..6a9cbc9 100644
--- a/ql/src/test/queries/clientpositive/multi_count_distinct.q
+++ b/ql/src/test/queries/clientpositive/multi_count_distinct.q
@@ -1,36 +1,36 @@
 SET hive.vectorized.execution.enabled=false;
 set hive.mapred.mode=nonstrict;
 
-drop table employee;
+drop table employee_n1;
 
-create table employee (department_id int, gender varchar(10), education_level int);
+create table employee_n1 (department_id int, gender varchar(10), education_level int);
  
-insert into employee values (1, 'M', 1),(1, 'M', 1),(2, 'F', 1),(1, 'F', 3),(1, 'M', 2),(4, 'M', 1),(2, 'F', 1),(2, 'F', 3),(3, 'M', 2);
+insert into employee_n1 values (1, 'M', 1),(1, 'M', 1),(2, 'F', 1),(1, 'F', 3),(1, 'M', 2),(4, 'M', 1),(2, 'F', 1),(2, 'F', 3),(3, 'M', 2);
 
-explain select count(distinct department_id), count(distinct gender), count(distinct education_level) from employee;
+explain select count(distinct department_id), count(distinct gender), count(distinct education_level) from employee_n1;
 
-select count(distinct department_id), count(distinct gender), count(distinct education_level) from employee;
+select count(distinct department_id), count(distinct gender), count(distinct education_level) from employee_n1;
 
-select count(distinct department_id), count(distinct gender), count(distinct education_level), count(distinct education_level) from employee;
+select count(distinct department_id), count(distinct gender), count(distinct education_level), count(distinct education_level) from employee_n1;
 
 select count(distinct department_id), count(distinct gender), count(distinct education_level), 
-count(distinct education_level, department_id) from employee;
+count(distinct education_level, department_id) from employee_n1;
 
 select count(distinct gender), count(distinct department_id), count(distinct gender), count(distinct education_level),
-count(distinct education_level, department_id), count(distinct department_id, education_level) from employee;
+count(distinct education_level, department_id), count(distinct department_id, education_level) from employee_n1;
 
 explain select count(distinct gender), count(distinct department_id), count(distinct gender), count(distinct education_level),
-count(distinct education_level, department_id), count(distinct department_id, education_level), count(distinct department_id, education_level, gender) from employee;
+count(distinct education_level, department_id), count(distinct department_id, education_level), count(distinct department_id, education_level, gender) from employee_n1;
 
 select count(distinct gender), count(distinct department_id), count(distinct gender), count(distinct education_level),
-count(distinct education_level, department_id), count(distinct department_id, education_level), count(distinct department_id, education_level, gender) from employee;
+count(distinct education_level, department_id), count(distinct department_id, education_level), count(distinct department_id, education_level, gender) from employee_n1;
 
 select count(case i when 3 then 1 else null end) as c0, count(case i when 5 then 1 else null end) as c1, 
 count(case i when 6 then 1 else null end) as c2 from (select grouping__id as i, department_id, gender, 
-education_level from employee group by department_id, gender, education_level grouping sets 
+education_level from employee_n1 group by department_id, gender, education_level grouping sets 
 (department_id, gender, education_level))subq;
 
-select grouping__id as i, department_id, gender, education_level from employee 
+select grouping__id as i, department_id, gender, education_level from employee_n1 
 group by department_id, gender, education_level grouping sets 
 (department_id, gender, education_level, (education_level, department_id));
 

http://git-wip-us.apache.org/repos/asf/hive/blob/38d3b8e1/ql/src/test/queries/clientpositive/multi_insert.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/multi_insert.q b/ql/src/test/queries/clientpositive/multi_insert.q
index 9e33a84..634bdb4 100644
--- a/ql/src/test/queries/clientpositive/multi_insert.q
+++ b/ql/src/test/queries/clientpositive/multi_insert.q
@@ -1,23 +1,23 @@
 --! qt:dataset:src
 -- SORT_QUERY_RESULTS
 
-create table src_multi1 like src;
-create table src_multi2 like src;
+create table src_multi1_n5 like src;
+create table src_multi2_n6 like src;
 
 set hive.merge.mapfiles=false;
 set hive.merge.mapredfiles=false;
 set hive.stats.dbclass=fs;
 explain
 from src
-insert overwrite table src_multi1 select * where key < 10
-insert overwrite table src_multi2 select * where key > 10 and key < 20;
+insert overwrite table src_multi1_n5 select * where key < 10
+insert overwrite table src_multi2_n6 select * where key > 10 and key < 20;
 
 from src
-insert overwrite table src_multi1 select * where key < 10
-insert overwrite table src_multi2 select * where key > 10 and key < 20;
+insert overwrite table src_multi1_n5 select * where key < 10
+insert overwrite table src_multi2_n6 select * where key > 10 and key < 20;
 
-select * from src_multi1;
-select * from src_multi2;
+select * from src_multi1_n5;
+select * from src_multi2_n6;
 
 
 set hive.merge.mapfiles=true;
@@ -25,45 +25,45 @@ set hive.merge.mapredfiles=false;
 
 explain
 from src
-insert overwrite table src_multi1 select * where key < 10
-insert overwrite table src_multi2 select * where key > 10 and key < 20;
+insert overwrite table src_multi1_n5 select * where key < 10
+insert overwrite table src_multi2_n6 select * where key > 10 and key < 20;
 
 from src
-insert overwrite table src_multi1 select * where key < 10
-insert overwrite table src_multi2 select * where key > 10 and key < 20;
+insert overwrite table src_multi1_n5 select * where key < 10
+insert overwrite table src_multi2_n6 select * where key > 10 and key < 20;
 
-select * from src_multi1;
-select * from src_multi2;
+select * from src_multi1_n5;
+select * from src_multi2_n6;
 
 set hive.merge.mapfiles=false;
 set hive.merge.mapredfiles=true;
 
 explain
 from src
-insert overwrite table src_multi1 select * where key < 10
-insert overwrite table src_multi2 select * where key > 10 and key < 20;
+insert overwrite table src_multi1_n5 select * where key < 10
+insert overwrite table src_multi2_n6 select * where key > 10 and key < 20;
 
 from src
-insert overwrite table src_multi1 select * where key < 10
-insert overwrite table src_multi2 select * where key > 10 and key < 20;
+insert overwrite table src_multi1_n5 select * where key < 10
+insert overwrite table src_multi2_n6 select * where key > 10 and key < 20;
 
-select * from src_multi1;
-select * from src_multi2;
+select * from src_multi1_n5;
+select * from src_multi2_n6;
 
 set hive.merge.mapfiles=true;
 set hive.merge.mapredfiles=true;
 
 explain
 from src
-insert overwrite table src_multi1 select * where key < 10
-insert overwrite table src_multi2 select * where key > 10 and key < 20;
+insert overwrite table src_multi1_n5 select * where key < 10
+insert overwrite table src_multi2_n6 select * where key > 10 and key < 20;
 
 from src
-insert overwrite table src_multi1 select * where key < 10
-insert overwrite table src_multi2 select * where key > 10 and key < 20;
+insert overwrite table src_multi1_n5 select * where key < 10
+insert overwrite table src_multi2_n6 select * where key > 10 and key < 20;
 
-select * from src_multi1;
-select * from src_multi2;
+select * from src_multi1_n5;
+select * from src_multi2_n6;
 
 
 
@@ -72,15 +72,15 @@ set hive.merge.mapredfiles=false;
 
 explain
 from src
-insert overwrite table src_multi1 select * where key < 10 group by key, value
-insert overwrite table src_multi2 select * where key > 10 and key < 20 group by key, value;
+insert overwrite table src_multi1_n5 select * where key < 10 group by key, value
+insert overwrite table src_multi2_n6 select * where key > 10 and key < 20 group by key, value;
 
 from src
-insert overwrite table src_multi1 select * where key < 10 group by key, value
-insert overwrite table src_multi2 select * where key > 10 and key < 20 group by key, value;
+insert overwrite table src_multi1_n5 select * where key < 10 group by key, value
+insert overwrite table src_multi2_n6 select * where key > 10 and key < 20 group by key, value;
 
-select * from src_multi1;
-select * from src_multi2;
+select * from src_multi1_n5;
+select * from src_multi2_n6;
 
 
 set hive.merge.mapfiles=false;
@@ -88,30 +88,30 @@ set hive.merge.mapredfiles=true;
 
 explain
 from src
-insert overwrite table src_multi1 select * where key < 10 group by key, value
-insert overwrite table src_multi2 select * where key > 10 and key < 20 group by key, value;
+insert overwrite table src_multi1_n5 select * where key < 10 group by key, value
+insert overwrite table src_multi2_n6 select * where key > 10 and key < 20 group by key, value;
 
 from src
-insert overwrite table src_multi1 select * where key < 10 group by key, value
-insert overwrite table src_multi2 select * where key > 10 and key < 20 group by key, value;
+insert overwrite table src_multi1_n5 select * where key < 10 group by key, value
+insert overwrite table src_multi2_n6 select * where key > 10 and key < 20 group by key, value;
 
-select * from src_multi1;
-select * from src_multi2;
+select * from src_multi1_n5;
+select * from src_multi2_n6;
 
 set hive.merge.mapfiles=true;
 set hive.merge.mapredfiles=false;
 
 explain
 from src
-insert overwrite table src_multi1 select * where key < 10 group by key, value
-insert overwrite table src_multi2 select * where key > 10 and key < 20 group by key, value;
+insert overwrite table src_multi1_n5 select * where key < 10 group by key, value
+insert overwrite table src_multi2_n6 select * where key > 10 and key < 20 group by key, value;
 
 from src
-insert overwrite table src_multi1 select * where key < 10 group by key, value
-insert overwrite table src_multi2 select * where key > 10 and key < 20 group by key, value;
+insert overwrite table src_multi1_n5 select * where key < 10 group by key, value
+insert overwrite table src_multi2_n6 select * where key > 10 and key < 20 group by key, value;
 
-select * from src_multi1;
-select * from src_multi2;
+select * from src_multi1_n5;
+select * from src_multi2_n6;
 
 
 set hive.merge.mapfiles=true;
@@ -119,15 +119,15 @@ set hive.merge.mapredfiles=true;
 
 explain
 from src
-insert overwrite table src_multi1 select * where key < 10 group by key, value
-insert overwrite table src_multi2 select * where key > 10 and key < 20 group by key, value;
+insert overwrite table src_multi1_n5 select * where key < 10 group by key, value
+insert overwrite table src_multi2_n6 select * where key > 10 and key < 20 group by key, value;
 
 from src
-insert overwrite table src_multi1 select * where key < 10 group by key, value
-insert overwrite table src_multi2 select * where key > 10 and key < 20 group by key, value;
+insert overwrite table src_multi1_n5 select * where key < 10 group by key, value
+insert overwrite table src_multi2_n6 select * where key > 10 and key < 20 group by key, value;
 
-select * from src_multi1;
-select * from src_multi2;
+select * from src_multi1_n5;
+select * from src_multi2_n6;
 
 
 
@@ -137,60 +137,60 @@ set hive.merge.mapredfiles=false;
 
 explain
 from (select * from src  union all select * from src) s
-insert overwrite table src_multi1 select * where key < 10
-insert overwrite table src_multi2 select * where key > 10 and key < 20;
+insert overwrite table src_multi1_n5 select * where key < 10
+insert overwrite table src_multi2_n6 select * where key > 10 and key < 20;
 
 from (select * from src  union all select * from src) s
-insert overwrite table src_multi1 select * where key < 10
-insert overwrite table src_multi2 select * where key > 10 and key < 20;
+insert overwrite table src_multi1_n5 select * where key < 10
+insert overwrite table src_multi2_n6 select * where key > 10 and key < 20;
 
-select * from src_multi1;
-select * from src_multi2;
+select * from src_multi1_n5;
+select * from src_multi2_n6;
 
 set hive.merge.mapfiles=true;
 set hive.merge.mapredfiles=false;
 
 explain
 from (select * from src  union all select * from src) s
-insert overwrite table src_multi1 select * where key < 10
-insert overwrite table src_multi2 select * where key > 10 and key < 20;
+insert overwrite table src_multi1_n5 select * where key < 10
+insert overwrite table src_multi2_n6 select * where key > 10 and key < 20;
 
 from (select * from src  union all select * from src) s
-insert overwrite table src_multi1 select * where key < 10
-insert overwrite table src_multi2 select * where key > 10 and key < 20;
+insert overwrite table src_multi1_n5 select * where key < 10
+insert overwrite table src_multi2_n6 select * where key > 10 and key < 20;
 
-select * from src_multi1;
-select * from src_multi2;
+select * from src_multi1_n5;
+select * from src_multi2_n6;
 
 set hive.merge.mapfiles=false;
 set hive.merge.mapredfiles=true;
 
 explain
 from (select * from src  union all select * from src) s
-insert overwrite table src_multi1 select * where key < 10
-insert overwrite table src_multi2 select * where key > 10 and key < 20;
+insert overwrite table src_multi1_n5 select * where key < 10
+insert overwrite table src_multi2_n6 select * where key > 10 and key < 20;
 
 from (select * from src  union all select * from src) s
-insert overwrite table src_multi1 select * where key < 10
-insert overwrite table src_multi2 select * where key > 10 and key < 20;
+insert overwrite table src_multi1_n5 select * where key < 10
+insert overwrite table src_multi2_n6 select * where key > 10 and key < 20;
 
-select * from src_multi1;
-select * from src_multi2;
+select * from src_multi1_n5;
+select * from src_multi2_n6;
 
 set hive.merge.mapfiles=true;
 set hive.merge.mapredfiles=true;
 
 explain
 from (select * from src  union all select * from src) s
-insert overwrite table src_multi1 select * where key < 10
-insert overwrite table src_multi2 select * where key > 10 and key < 20;
+insert overwrite table src_multi1_n5 select * where key < 10
+insert overwrite table src_multi2_n6 select * where key > 10 and key < 20;
 
 from (select * from src  union all select * from src) s
-insert overwrite table src_multi1 select * where key < 10
-insert overwrite table src_multi2 select * where key > 10 and key < 20;
+insert overwrite table src_multi1_n5 select * where key < 10
+insert overwrite table src_multi2_n6 select * where key > 10 and key < 20;
 
-select * from src_multi1;
-select * from src_multi2;
+select * from src_multi1_n5;
+select * from src_multi2_n6;
 
 
 

http://git-wip-us.apache.org/repos/asf/hive/blob/38d3b8e1/ql/src/test/queries/clientpositive/multi_insert_gby.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/multi_insert_gby.q b/ql/src/test/queries/clientpositive/multi_insert_gby.q
index 7c9e8de..0ec2b35 100644
--- a/ql/src/test/queries/clientpositive/multi_insert_gby.q
+++ b/ql/src/test/queries/clientpositive/multi_insert_gby.q
@@ -2,35 +2,35 @@
 -- SORT_QUERY_RESULTS
 
 --HIVE-3699 Multiple insert overwrite into multiple tables query stores same results in all tables
-create table e1 (key string, count int);
-create table e2 (key string, count int);
+create table e1_n0 (key string, count int);
+create table e2_n1 (key string, count int);
 
 explain FROM src
-INSERT OVERWRITE TABLE e1
+INSERT OVERWRITE TABLE e1_n0
     SELECT key, COUNT(*) WHERE key>450 GROUP BY key
-INSERT OVERWRITE TABLE e2
+INSERT OVERWRITE TABLE e2_n1
     SELECT key, COUNT(*) WHERE key>500 GROUP BY key;
 
 FROM src
-INSERT OVERWRITE TABLE e1
+INSERT OVERWRITE TABLE e1_n0
     SELECT key, COUNT(*) WHERE key>450 GROUP BY key
-INSERT OVERWRITE TABLE e2
+INSERT OVERWRITE TABLE e2_n1
     SELECT key, COUNT(*) WHERE key>500 GROUP BY key;
 
-select * from e1;
-select * from e2;
+select * from e1_n0;
+select * from e2_n1;
 
 explain FROM src
-INSERT OVERWRITE TABLE e1
+INSERT OVERWRITE TABLE e1_n0
     SELECT key, COUNT(*) WHERE key>450 GROUP BY key
-INSERT OVERWRITE TABLE e2
+INSERT OVERWRITE TABLE e2_n1
     SELECT key, COUNT(*) GROUP BY key;
 
 FROM src
-INSERT OVERWRITE TABLE e1
+INSERT OVERWRITE TABLE e1_n0
     SELECT key, COUNT(*) WHERE key>450 GROUP BY key
-INSERT OVERWRITE TABLE e2
+INSERT OVERWRITE TABLE e2_n1
     SELECT key, COUNT(*) GROUP BY key;
 
-select * from e1;
-select * from e2;
+select * from e1_n0;
+select * from e2_n1;

http://git-wip-us.apache.org/repos/asf/hive/blob/38d3b8e1/ql/src/test/queries/clientpositive/multi_insert_gby2.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/multi_insert_gby2.q b/ql/src/test/queries/clientpositive/multi_insert_gby2.q
index a477f83..67077ba 100644
--- a/ql/src/test/queries/clientpositive/multi_insert_gby2.q
+++ b/ql/src/test/queries/clientpositive/multi_insert_gby2.q
@@ -2,20 +2,20 @@
 set hive.mapred.mode=nonstrict;
 --HIVE-3699 Multiple insert overwrite into multiple tables query stores same results in all tables
 create table e1 (count int);
-create table e2 (percentile double);
+create table e2_n0 (percentile double);
 set hive.stats.dbclass=fs;
 explain
 FROM (select key, cast(key as double) as value from src order by key) a
 INSERT OVERWRITE TABLE e1
     SELECT COUNT(*)
-INSERT OVERWRITE TABLE e2
+INSERT OVERWRITE TABLE e2_n0
     SELECT percentile_approx(value, 0.5);
 
 FROM (select key, cast(key as double) as value from src order by key) a
 INSERT OVERWRITE TABLE e1
     SELECT COUNT(*)
-INSERT OVERWRITE TABLE e2
+INSERT OVERWRITE TABLE e2_n0
     SELECT percentile_approx(value, 0.5);
 
 select * from e1;
-select * from e2;
+select * from e2_n0;

http://git-wip-us.apache.org/repos/asf/hive/blob/38d3b8e1/ql/src/test/queries/clientpositive/multi_insert_gby3.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/multi_insert_gby3.q b/ql/src/test/queries/clientpositive/multi_insert_gby3.q
index 6bed476..de60fa7 100644
--- a/ql/src/test/queries/clientpositive/multi_insert_gby3.q
+++ b/ql/src/test/queries/clientpositive/multi_insert_gby3.q
@@ -1,45 +1,45 @@
 --! qt:dataset:src
 set hive.mapred.mode=nonstrict;
 -- SORT_QUERY_RESULTS
-create table e1 (key string, keyD double);
-create table e2 (key string, keyD double, value string);
+create table e1_n2 (key string, keyD double);
+create table e2_n3 (key string, keyD double, value string);
 create table e3 (key string, keyD double);
 set hive.stats.dbclass=fs;
 explain
 FROM (select key, cast(key as double) as keyD, value from src order by key) a
-INSERT OVERWRITE TABLE e1
+INSERT OVERWRITE TABLE e1_n2
     SELECT key, COUNT(distinct value) group by key
-INSERT OVERWRITE TABLE e2
+INSERT OVERWRITE TABLE e2_n3
     SELECT key, sum(keyD), value group by key, value;
 
 explain
 FROM (select key, cast(key as double) as keyD, value from src order by key) a
-INSERT OVERWRITE TABLE e2
+INSERT OVERWRITE TABLE e2_n3
     SELECT key, sum(keyD), value group by key, value
-INSERT OVERWRITE TABLE e1
+INSERT OVERWRITE TABLE e1_n2
     SELECT key, COUNT(distinct value) group by key;
 
 FROM (select key, cast(key as double) as keyD, value from src order by key) a
-INSERT OVERWRITE TABLE e1
+INSERT OVERWRITE TABLE e1_n2
     SELECT key, COUNT(distinct value) group by key
-INSERT OVERWRITE TABLE e2
+INSERT OVERWRITE TABLE e2_n3
     SELECT key, sum(keyD), value group by key, value;
 
-select * from e1;
-select * from e2;
+select * from e1_n2;
+select * from e2_n3;
 
 FROM (select key, cast(key as double) as keyD, value from src order by key) a
-INSERT OVERWRITE TABLE e2
+INSERT OVERWRITE TABLE e2_n3
     SELECT key, sum(keyD), value group by key, value
-INSERT OVERWRITE TABLE e1
+INSERT OVERWRITE TABLE e1_n2
     SELECT key, COUNT(distinct value) group by key;
 
-select * from e1;
-select * from e2;
+select * from e1_n2;
+select * from e2_n3;
 
 explain
 from src
-insert overwrite table e1
+insert overwrite table e1_n2
 select key, count(distinct value) group by key
 insert overwrite table e3
 select value, count(distinct key) group by value;
@@ -47,9 +47,9 @@ select value, count(distinct key) group by value;
 
 explain
 FROM (select key, cast(key as double) as keyD, value from src order by key) a
-INSERT OVERWRITE TABLE e1
+INSERT OVERWRITE TABLE e1_n2
     SELECT key, COUNT(distinct value) group by key
-INSERT OVERWRITE TABLE e2
+INSERT OVERWRITE TABLE e2_n3
     SELECT key, sum(keyD), value group by key, value
 INSERT overwrite table e3
     SELECT key, COUNT(distinct keyD) group by key, keyD, value;

http://git-wip-us.apache.org/repos/asf/hive/blob/38d3b8e1/ql/src/test/queries/clientpositive/multi_insert_gby4.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/multi_insert_gby4.q b/ql/src/test/queries/clientpositive/multi_insert_gby4.q
index d485813..b89d674 100644
--- a/ql/src/test/queries/clientpositive/multi_insert_gby4.q
+++ b/ql/src/test/queries/clientpositive/multi_insert_gby4.q
@@ -1,27 +1,27 @@
 --! qt:dataset:src
 -- SORT_QUERY_RESULTS
 
-create table e1 (key string, count int);
-create table e2 (key string, count int);
-create table e3 (key string, count int);
+create table e1_n4 (key string, count int);
+create table e2_n5 (key string, count int);
+create table e3_n0 (key string, count int);
 
 explain
 FROM (SELECT key, value FROM src) a
-INSERT OVERWRITE TABLE e1
+INSERT OVERWRITE TABLE e1_n4
     SELECT key, COUNT(*) WHERE key>450 GROUP BY key
-INSERT OVERWRITE TABLE e2
+INSERT OVERWRITE TABLE e2_n5
     SELECT key, COUNT(*) WHERE key>500 GROUP BY key
-INSERT OVERWRITE TABLE e3
+INSERT OVERWRITE TABLE e3_n0
     SELECT key, COUNT(*) WHERE key>490 GROUP BY key;
 
 FROM (SELECT key, value FROM src) a
-INSERT OVERWRITE TABLE e1
+INSERT OVERWRITE TABLE e1_n4
     SELECT key, COUNT(*) WHERE key>450 GROUP BY key
-INSERT OVERWRITE TABLE e2
+INSERT OVERWRITE TABLE e2_n5
     SELECT key, COUNT(*) WHERE key>500 GROUP BY key
-INSERT OVERWRITE TABLE e3
+INSERT OVERWRITE TABLE e3_n0
     SELECT key, COUNT(*) WHERE key>490 GROUP BY key;
 
-select * from e1;
-select * from e2;
-select * from e3;
+select * from e1_n4;
+select * from e2_n5;
+select * from e3_n0;

http://git-wip-us.apache.org/repos/asf/hive/blob/38d3b8e1/ql/src/test/queries/clientpositive/multi_insert_lateral_view.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/multi_insert_lateral_view.q b/ql/src/test/queries/clientpositive/multi_insert_lateral_view.q
index 481bdf2..3442e82 100644
--- a/ql/src/test/queries/clientpositive/multi_insert_lateral_view.q
+++ b/ql/src/test/queries/clientpositive/multi_insert_lateral_view.q
@@ -2,7 +2,7 @@
 set hive.stats.dbclass=fs;
 -- SORT_QUERY_RESULTS
 
-create table src_10 as select * from src limit 10;
+create table src_10_n0 as select * from src limit 10;
 
 create table src_lv1 (key string, value string);
 create table src_lv2 (key string, value string);
@@ -14,11 +14,11 @@ create table src_lv3 (key string, value string);
 --      -LVF[6]-SEL[7]-LVJ[10]-SEL[13]-FS[14]
 --             -SEL[8]-UDTF[9]-LVJ[10]
 explain
-from src_10
+from src_10_n0
 insert overwrite table src_lv1 select key, C lateral view explode(array(key+1, key+2)) A as C
 insert overwrite table src_lv2 select key, C lateral view explode(array(key+3, key+4)) A as C;
 
-from src_10
+from src_10_n0
 insert overwrite table src_lv1 select key, C lateral view explode(array(key+1, key+2)) A as C
 insert overwrite table src_lv2 select key, C lateral view explode(array(key+3, key+4)) A as C;
 
@@ -31,11 +31,11 @@ select * from src_lv2;
 --      -LVF[6]-SEL[7]-LVJ[10]-SEL[17]-GBY[18]-RS[19]-GBY[20]-SEL[21]-FS[22]
 --             -SEL[8]-UDTF[9]-LVJ[10]
 explain
-from src_10
+from src_10_n0
 insert overwrite table src_lv1 select key, sum(C) lateral view explode(array(key+1, key+2)) A as C group by key
 insert overwrite table src_lv2 select key, sum(C) lateral view explode(array(key+3, key+4)) A as C group by key;
 
-from src_10
+from src_10_n0
 insert overwrite table src_lv1 select key, sum(C) lateral view explode(array(key+1, key+2)) A as C group by key
 insert overwrite table src_lv2 select key, sum(C) lateral view explode(array(key+3, key+4)) A as C group by key;
 
@@ -48,12 +48,12 @@ select * from src_lv2;
 --      -FIL[12]-SEL[13]-RS[14]-FOR[15]-FIL[16]-GBY[17]-SEL[18]-FS[19]
 --                                     -FIL[20]-GBY[21]-SEL[22]-FS[23]
 explain
-from src_10
+from src_10_n0
 insert overwrite table src_lv1 select key, sum(C) lateral view explode(array(key+1, key+2)) A as C group by key
 insert overwrite table src_lv2 select key, count(value) where key > 200 group by key
 insert overwrite table src_lv3 select key, count(value) where key < 200 group by key;
 
-from src_10
+from src_10_n0
 insert overwrite table src_lv1 select key, sum(C) lateral view explode(array(key+1, key+2)) A as C group by key
 insert overwrite table src_lv2 select key, count(value) where key > 200 group by key
 insert overwrite table src_lv3 select key, count(value) where key < 200 group by key;
@@ -70,12 +70,12 @@ select * from src_lv3;
 --             -SEL[8]-UDTF[9]-LVJ[10]
 --      -SEL[23]-GBY[24]-RS[25]-GBY[26]-SEL[27]-FS[28]
 explain
-from src_10
+from src_10_n0
 insert overwrite table src_lv1 select C, sum(distinct key) lateral view explode(array(key+1, key+2)) A as C group by C
 insert overwrite table src_lv2 select C, sum(distinct key) lateral view explode(array(key+3, key+4)) A as C group by C
 insert overwrite table src_lv3 select value, sum(distinct key) group by value;
 
-from src_10
+from src_10_n0
 insert overwrite table src_lv1 select C, sum(distinct key) lateral view explode(array(key+1, key+2)) A as C group by C
 insert overwrite table src_lv2 select C, sum(distinct key) lateral view explode(array(key+3, key+4)) A as C group by C
 insert overwrite table src_lv3 select value, sum(distinct key) group by value;
@@ -88,13 +88,13 @@ create table src_lv4 (key string, value string);
 
 -- Common distincts optimization works across non-lateral view queries, but not across lateral view multi inserts
 explain
-from src_10
+from src_10_n0
 insert overwrite table src_lv1 select key, sum(distinct C) lateral view explode(array(key+1, key+2)) A as C group by key
 insert overwrite table src_lv2 select key, sum(distinct C) lateral view explode(array(key+3, key+4)) A as C group by key
 insert overwrite table src_lv3 select value, sum(distinct key) where key > 200 group by value
 insert overwrite table src_lv4 select value, sum(distinct key) where key < 200 group by value;
 
-from src_10
+from src_10_n0
 insert overwrite table src_lv1 select key, sum(distinct C) lateral view explode(array(key+1, key+2)) A as C group by key
 insert overwrite table src_lv2 select key, sum(distinct C) lateral view explode(array(key+3, key+4)) A as C group by key
 insert overwrite table src_lv3 select value, sum(distinct key) where key > 200 group by value

http://git-wip-us.apache.org/repos/asf/hive/blob/38d3b8e1/ql/src/test/queries/clientpositive/multi_insert_mixed.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/multi_insert_mixed.q b/ql/src/test/queries/clientpositive/multi_insert_mixed.q
index a5c1e87..9c8fb2e 100644
--- a/ql/src/test/queries/clientpositive/multi_insert_mixed.q
+++ b/ql/src/test/queries/clientpositive/multi_insert_mixed.q
@@ -1,7 +1,7 @@
 --! qt:dataset:src
 set hive.mapred.mode=nonstrict;
-create table src_multi1 like src;
-create table src_multi2 like src;
+create table src_multi1_n2 like src;
+create table src_multi2_n3 like src;
 create table src_multi3 like src;
 set hive.stats.dbclass=fs;
 -- Testing the case where a map work contains both shuffling (ReduceSinkOperator)
@@ -9,15 +9,15 @@ set hive.stats.dbclass=fs;
 
 explain
 from src
-insert overwrite table src_multi1 select key, count(1) group by key order by key
-insert overwrite table src_multi2 select value, count(1) group by value order by value
+insert overwrite table src_multi1_n2 select key, count(1) group by key order by key
+insert overwrite table src_multi2_n3 select value, count(1) group by value order by value
 insert overwrite table src_multi3 select * where key < 10;
 
 from src
-insert overwrite table src_multi1 select key, count(1) group by key order by key
-insert overwrite table src_multi2 select value, count(1) group by value order by value
+insert overwrite table src_multi1_n2 select key, count(1) group by key order by key
+insert overwrite table src_multi2_n3 select value, count(1) group by value order by value
 insert overwrite table src_multi3 select * where key < 10;
 
-select * from src_multi1;
-select * from src_multi2;
+select * from src_multi1_n2;
+select * from src_multi2_n3;
 select * from src_multi3;