You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hive.apache.org by jc...@apache.org on 2018/05/25 00:57:53 UTC

[21/58] [abbrv] [partial] hive git commit: HIVE-19617: Rename test tables to avoid collisions during execution in batches (Jesus Camacho Rodriguez, reviewed by Gunther Hagleitner)

http://git-wip-us.apache.org/repos/asf/hive/blob/9bf28a3c/ql/src/test/queries/clientpositive/orc_ppd_date.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/orc_ppd_date.q b/ql/src/test/queries/clientpositive/orc_ppd_date.q
index 32767d8..2069fbf 100644
--- a/ql/src/test/queries/clientpositive/orc_ppd_date.q
+++ b/ql/src/test/queries/clientpositive/orc_ppd_date.q
@@ -3,99 +3,99 @@ SET hive.input.format=org.apache.hadoop.hive.ql.io.HiveInputFormat;
 SET mapred.min.split.size=1000;
 SET mapred.max.split.size=5000;
 
-create table newtypesorc(c char(10), v varchar(10), d decimal(5,3), da date) stored as orc tblproperties("orc.stripe.size"="16777216"); 
+create table newtypesorc_n3(c char(10), v varchar(10), d decimal(5,3), da date) stored as orc tblproperties("orc.stripe.size"="16777216"); 
 
-insert overwrite table newtypesorc select * from (select cast("apple" as char(10)), cast("bee" as varchar(10)), 0.22, cast("1970-02-20" as date) from src src1 union all select cast("hello" as char(10)), cast("world" as varchar(10)), 11.22, cast("1970-02-27" as date) from src src2) uniontbl;
+insert overwrite table newtypesorc_n3 select * from (select cast("apple" as char(10)), cast("bee" as varchar(10)), 0.22, cast("1970-02-20" as date) from src src1 union all select cast("hello" as char(10)), cast("world" as varchar(10)), 11.22, cast("1970-02-27" as date) from src src2) uniontbl;
 
 -- date data types (EQUAL, NOT_EQUAL, LESS_THAN, LESS_THAN_EQUALS, IN, BETWEEN tests)
-select sum(hash(*)) from newtypesorc where da='1970-02-20';
+select sum(hash(*)) from newtypesorc_n3 where da='1970-02-20';
 
 set hive.optimize.index.filter=true;
-select sum(hash(*)) from newtypesorc where da='1970-02-20';
+select sum(hash(*)) from newtypesorc_n3 where da='1970-02-20';
 
 set hive.optimize.index.filter=true;
-select sum(hash(*)) from newtypesorc where da= date '1970-02-20';
+select sum(hash(*)) from newtypesorc_n3 where da= date '1970-02-20';
 
 set hive.optimize.index.filter=false;
-select sum(hash(*)) from newtypesorc where da=cast('1970-02-20' as date);
+select sum(hash(*)) from newtypesorc_n3 where da=cast('1970-02-20' as date);
 
 set hive.optimize.index.filter=true;
-select sum(hash(*)) from newtypesorc where da=cast('1970-02-20' as date);
+select sum(hash(*)) from newtypesorc_n3 where da=cast('1970-02-20' as date);
 
 set hive.optimize.index.filter=false;
-select sum(hash(*)) from newtypesorc where da=cast('1970-02-20' as varchar(20));
+select sum(hash(*)) from newtypesorc_n3 where da=cast('1970-02-20' as varchar(20));
 
 set hive.optimize.index.filter=true;
-select sum(hash(*)) from newtypesorc where da=cast('1970-02-20' as varchar(20));
+select sum(hash(*)) from newtypesorc_n3 where da=cast('1970-02-20' as varchar(20));
 
 set hive.optimize.index.filter=false;
-select sum(hash(*)) from newtypesorc where da!='1970-02-20';
+select sum(hash(*)) from newtypesorc_n3 where da!='1970-02-20';
 
 set hive.optimize.index.filter=true;
-select sum(hash(*)) from newtypesorc where da!='1970-02-20';
+select sum(hash(*)) from newtypesorc_n3 where da!='1970-02-20';
 
 set hive.optimize.index.filter=false;
-select sum(hash(*)) from newtypesorc where da<'1970-02-27';
+select sum(hash(*)) from newtypesorc_n3 where da<'1970-02-27';
 
 set hive.optimize.index.filter=true;
-select sum(hash(*)) from newtypesorc where da<'1970-02-27';
+select sum(hash(*)) from newtypesorc_n3 where da<'1970-02-27';
 
 set hive.optimize.index.filter=false;
-select sum(hash(*)) from newtypesorc where da<'1970-02-29';
+select sum(hash(*)) from newtypesorc_n3 where da<'1970-02-29';
 
 set hive.optimize.index.filter=true;
-select sum(hash(*)) from newtypesorc where da<'1970-02-29';
+select sum(hash(*)) from newtypesorc_n3 where da<'1970-02-29';
 
 set hive.optimize.index.filter=false;
-select sum(hash(*)) from newtypesorc where da<'1970-02-15';
+select sum(hash(*)) from newtypesorc_n3 where da<'1970-02-15';
 
 set hive.optimize.index.filter=true;
-select sum(hash(*)) from newtypesorc where da<'1970-02-15';
+select sum(hash(*)) from newtypesorc_n3 where da<'1970-02-15';
 
 set hive.optimize.index.filter=false;
-select sum(hash(*)) from newtypesorc where da<='1970-02-20';
+select sum(hash(*)) from newtypesorc_n3 where da<='1970-02-20';
 
 set hive.optimize.index.filter=true;
-select sum(hash(*)) from newtypesorc where da<='1970-02-20';
+select sum(hash(*)) from newtypesorc_n3 where da<='1970-02-20';
 
 set hive.optimize.index.filter=false;
-select sum(hash(*)) from newtypesorc where da<='1970-02-27';
+select sum(hash(*)) from newtypesorc_n3 where da<='1970-02-27';
 
 set hive.optimize.index.filter=true;
-select sum(hash(*)) from newtypesorc where da<='1970-02-27';
+select sum(hash(*)) from newtypesorc_n3 where da<='1970-02-27';
 
 set hive.optimize.index.filter=false;
-select sum(hash(*)) from newtypesorc where da in (cast('1970-02-21' as date), cast('1970-02-27' as date));
+select sum(hash(*)) from newtypesorc_n3 where da in (cast('1970-02-21' as date), cast('1970-02-27' as date));
 
 set hive.optimize.index.filter=true;
-select sum(hash(*)) from newtypesorc where da in (cast('1970-02-21' as date), cast('1970-02-27' as date));
+select sum(hash(*)) from newtypesorc_n3 where da in (cast('1970-02-21' as date), cast('1970-02-27' as date));
 
 set hive.optimize.index.filter=false;
-select sum(hash(*)) from newtypesorc where da in (cast('1970-02-20' as date), cast('1970-02-27' as date));
+select sum(hash(*)) from newtypesorc_n3 where da in (cast('1970-02-20' as date), cast('1970-02-27' as date));
 
 set hive.optimize.index.filter=true;
-select sum(hash(*)) from newtypesorc where da in (cast('1970-02-20' as date), cast('1970-02-27' as date));
+select sum(hash(*)) from newtypesorc_n3 where da in (cast('1970-02-20' as date), cast('1970-02-27' as date));
 
 set hive.optimize.index.filter=false;
-select sum(hash(*)) from newtypesorc where da in (cast('1970-02-21' as date), cast('1970-02-22' as date));
+select sum(hash(*)) from newtypesorc_n3 where da in (cast('1970-02-21' as date), cast('1970-02-22' as date));
 
 set hive.optimize.index.filter=true;
-select sum(hash(*)) from newtypesorc where da in (cast('1970-02-21' as date), cast('1970-02-22' as date));
+select sum(hash(*)) from newtypesorc_n3 where da in (cast('1970-02-21' as date), cast('1970-02-22' as date));
 
 set hive.optimize.index.filter=false;
-select sum(hash(*)) from newtypesorc where da between '1970-02-19' and '1970-02-22';
+select sum(hash(*)) from newtypesorc_n3 where da between '1970-02-19' and '1970-02-22';
 
 set hive.optimize.index.filter=true;
-select sum(hash(*)) from newtypesorc where da between '1970-02-19' and '1970-02-22';
+select sum(hash(*)) from newtypesorc_n3 where da between '1970-02-19' and '1970-02-22';
 
 set hive.optimize.index.filter=false;
-select sum(hash(*)) from newtypesorc where da between '1970-02-19' and '1970-02-28';
+select sum(hash(*)) from newtypesorc_n3 where da between '1970-02-19' and '1970-02-28';
 
 set hive.optimize.index.filter=true;
-select sum(hash(*)) from newtypesorc where da between '1970-02-19' and '1970-02-28';
+select sum(hash(*)) from newtypesorc_n3 where da between '1970-02-19' and '1970-02-28';
 
 set hive.optimize.index.filter=false;
-select sum(hash(*)) from newtypesorc where da between '1970-02-18' and '1970-02-19';
+select sum(hash(*)) from newtypesorc_n3 where da between '1970-02-18' and '1970-02-19';
 
 set hive.optimize.index.filter=true;
-select sum(hash(*)) from newtypesorc where da between '1970-02-18' and '1970-02-19';
+select sum(hash(*)) from newtypesorc_n3 where da between '1970-02-18' and '1970-02-19';

http://git-wip-us.apache.org/repos/asf/hive/blob/9bf28a3c/ql/src/test/queries/clientpositive/orc_ppd_decimal.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/orc_ppd_decimal.q b/ql/src/test/queries/clientpositive/orc_ppd_decimal.q
index 268d5ae..3aa8e6c 100644
--- a/ql/src/test/queries/clientpositive/orc_ppd_decimal.q
+++ b/ql/src/test/queries/clientpositive/orc_ppd_decimal.q
@@ -3,162 +3,162 @@ SET hive.input.format=org.apache.hadoop.hive.ql.io.HiveInputFormat;
 SET mapred.min.split.size=1000;
 SET mapred.max.split.size=5000;
 
-create table newtypesorc(c char(10), v varchar(10), d decimal(5,3), da date) stored as orc tblproperties("orc.stripe.size"="16777216"); 
+create table newtypesorc_n5(c char(10), v varchar(10), d decimal(5,3), da date) stored as orc tblproperties("orc.stripe.size"="16777216"); 
 
-insert overwrite table newtypesorc select * from (select cast("apple" as char(10)), cast("bee" as varchar(10)), 0.22, cast("1970-02-20" as date) from src src1 union all select cast("hello" as char(10)), cast("world" as varchar(10)), 11.22, cast("1970-02-27" as date) from src src2) uniontbl;
+insert overwrite table newtypesorc_n5 select * from (select cast("apple" as char(10)), cast("bee" as varchar(10)), 0.22, cast("1970-02-20" as date) from src src1 union all select cast("hello" as char(10)), cast("world" as varchar(10)), 11.22, cast("1970-02-27" as date) from src src2) uniontbl;
 
 -- decimal data types (EQUAL, NOT_EQUAL, LESS_THAN, LESS_THAN_EQUALS, IN, BETWEEN tests)
-select sum(hash(*)) from newtypesorc where d=0.22;
+select sum(hash(*)) from newtypesorc_n5 where d=0.22;
 
 set hive.optimize.index.filter=true;
-select sum(hash(*)) from newtypesorc where d=0.22;
+select sum(hash(*)) from newtypesorc_n5 where d=0.22;
 
 set hive.optimize.index.filter=false;
-select sum(hash(*)) from newtypesorc where d='0.22';
+select sum(hash(*)) from newtypesorc_n5 where d='0.22';
 
 set hive.optimize.index.filter=true;
-select sum(hash(*)) from newtypesorc where d='0.22';
+select sum(hash(*)) from newtypesorc_n5 where d='0.22';
 
 set hive.optimize.index.filter=false;
-select sum(hash(*)) from newtypesorc where d=cast('0.22' as float);
+select sum(hash(*)) from newtypesorc_n5 where d=cast('0.22' as float);
 
 set hive.optimize.index.filter=true;
-select sum(hash(*)) from newtypesorc where d=cast('0.22' as float);
+select sum(hash(*)) from newtypesorc_n5 where d=cast('0.22' as float);
 
 set hive.optimize.index.filter=false;
-select sum(hash(*)) from newtypesorc where d!=0.22;
+select sum(hash(*)) from newtypesorc_n5 where d!=0.22;
 
 set hive.optimize.index.filter=true;
-select sum(hash(*)) from newtypesorc where d!=0.22;
+select sum(hash(*)) from newtypesorc_n5 where d!=0.22;
 
 set hive.optimize.index.filter=false;
-select sum(hash(*)) from newtypesorc where d!='0.22';
+select sum(hash(*)) from newtypesorc_n5 where d!='0.22';
 
 set hive.optimize.index.filter=true;
-select sum(hash(*)) from newtypesorc where d!='0.22';
+select sum(hash(*)) from newtypesorc_n5 where d!='0.22';
 
 set hive.optimize.index.filter=false;
-select sum(hash(*)) from newtypesorc where d!=cast('0.22' as float);
+select sum(hash(*)) from newtypesorc_n5 where d!=cast('0.22' as float);
 
 set hive.optimize.index.filter=true;
-select sum(hash(*)) from newtypesorc where d!=cast('0.22' as float);
+select sum(hash(*)) from newtypesorc_n5 where d!=cast('0.22' as float);
 
 set hive.optimize.index.filter=false;
-select sum(hash(*)) from newtypesorc where d<11.22;
+select sum(hash(*)) from newtypesorc_n5 where d<11.22;
 
 set hive.optimize.index.filter=true;
-select sum(hash(*)) from newtypesorc where d<11.22;
+select sum(hash(*)) from newtypesorc_n5 where d<11.22;
 
 set hive.optimize.index.filter=false;
-select sum(hash(*)) from newtypesorc where d<'11.22';
+select sum(hash(*)) from newtypesorc_n5 where d<'11.22';
 
 set hive.optimize.index.filter=true;
-select sum(hash(*)) from newtypesorc where d<'11.22';
+select sum(hash(*)) from newtypesorc_n5 where d<'11.22';
 
 set hive.optimize.index.filter=false;
-select sum(hash(*)) from newtypesorc where d<cast('11.22' as float);
+select sum(hash(*)) from newtypesorc_n5 where d<cast('11.22' as float);
 
 set hive.optimize.index.filter=true;
-select sum(hash(*)) from newtypesorc where d<cast('11.22' as float);
+select sum(hash(*)) from newtypesorc_n5 where d<cast('11.22' as float);
 
 set hive.optimize.index.filter=false;
-select sum(hash(*)) from newtypesorc where d<1;
+select sum(hash(*)) from newtypesorc_n5 where d<1;
 
 set hive.optimize.index.filter=true;
-select sum(hash(*)) from newtypesorc where d<1;
+select sum(hash(*)) from newtypesorc_n5 where d<1;
 
 set hive.optimize.index.filter=false;
-select sum(hash(*)) from newtypesorc where d<=11.22;
+select sum(hash(*)) from newtypesorc_n5 where d<=11.22;
 
 set hive.optimize.index.filter=true;
-select sum(hash(*)) from newtypesorc where d<=11.22;
+select sum(hash(*)) from newtypesorc_n5 where d<=11.22;
 
 set hive.optimize.index.filter=false;
-select sum(hash(*)) from newtypesorc where d<='11.22';
+select sum(hash(*)) from newtypesorc_n5 where d<='11.22';
 
 set hive.optimize.index.filter=true;
-select sum(hash(*)) from newtypesorc where d<='11.22';
+select sum(hash(*)) from newtypesorc_n5 where d<='11.22';
 
 set hive.optimize.index.filter=false;
-select sum(hash(*)) from newtypesorc where d<=cast('11.22' as float);
+select sum(hash(*)) from newtypesorc_n5 where d<=cast('11.22' as float);
 
 set hive.optimize.index.filter=true;
-select sum(hash(*)) from newtypesorc where d<=cast('11.22' as float);
+select sum(hash(*)) from newtypesorc_n5 where d<=cast('11.22' as float);
 
 set hive.optimize.index.filter=false;
-select sum(hash(*)) from newtypesorc where d<=cast('11.22' as decimal);
+select sum(hash(*)) from newtypesorc_n5 where d<=cast('11.22' as decimal);
 
 set hive.optimize.index.filter=true;
-select sum(hash(*)) from newtypesorc where d<=cast('11.22' as decimal);
+select sum(hash(*)) from newtypesorc_n5 where d<=cast('11.22' as decimal);
 
 set hive.optimize.index.filter=false;
-select sum(hash(*)) from newtypesorc where d<=11.22BD;
+select sum(hash(*)) from newtypesorc_n5 where d<=11.22BD;
 
 set hive.optimize.index.filter=true;
-select sum(hash(*)) from newtypesorc where d<=11.22BD;
+select sum(hash(*)) from newtypesorc_n5 where d<=11.22BD;
 
 set hive.optimize.index.filter=false;
-select sum(hash(*)) from newtypesorc where d<=12;
+select sum(hash(*)) from newtypesorc_n5 where d<=12;
 
 set hive.optimize.index.filter=true;
-select sum(hash(*)) from newtypesorc where d<=12;
+select sum(hash(*)) from newtypesorc_n5 where d<=12;
 
 set hive.optimize.index.filter=false;
-select sum(hash(*)) from newtypesorc where d in ('0.22', '1.0');
+select sum(hash(*)) from newtypesorc_n5 where d in ('0.22', '1.0');
 
 set hive.optimize.index.filter=true;
-select sum(hash(*)) from newtypesorc where d in ('0.22', '1.0');
+select sum(hash(*)) from newtypesorc_n5 where d in ('0.22', '1.0');
 
 set hive.optimize.index.filter=false;
-select sum(hash(*)) from newtypesorc where d in ('0.22', '11.22');
+select sum(hash(*)) from newtypesorc_n5 where d in ('0.22', '11.22');
 
 set hive.optimize.index.filter=true;
-select sum(hash(*)) from newtypesorc where d in ('0.22', '11.22');
+select sum(hash(*)) from newtypesorc_n5 where d in ('0.22', '11.22');
 
 set hive.optimize.index.filter=false;
-select sum(hash(*)) from newtypesorc where d in ('0.9', '1.0');
+select sum(hash(*)) from newtypesorc_n5 where d in ('0.9', '1.0');
 
 set hive.optimize.index.filter=true;
-select sum(hash(*)) from newtypesorc where d in ('0.9', '1.0');
+select sum(hash(*)) from newtypesorc_n5 where d in ('0.9', '1.0');
 
 set hive.optimize.index.filter=false;
-select sum(hash(*)) from newtypesorc where d in ('0.9', 0.22);
+select sum(hash(*)) from newtypesorc_n5 where d in ('0.9', 0.22);
 
 set hive.optimize.index.filter=true;
-select sum(hash(*)) from newtypesorc where d in ('0.9', 0.22);
+select sum(hash(*)) from newtypesorc_n5 where d in ('0.9', 0.22);
 
 set hive.optimize.index.filter=false;
-select sum(hash(*)) from newtypesorc where d in ('0.9', 0.22, cast('11.22' as float));
+select sum(hash(*)) from newtypesorc_n5 where d in ('0.9', 0.22, cast('11.22' as float));
 
 set hive.optimize.index.filter=true;
-select sum(hash(*)) from newtypesorc where d in ('0.9', 0.22, cast('11.22' as float));
+select sum(hash(*)) from newtypesorc_n5 where d in ('0.9', 0.22, cast('11.22' as float));
 
 set hive.optimize.index.filter=false;
-select sum(hash(*)) from newtypesorc where d between 0 and 1;
+select sum(hash(*)) from newtypesorc_n5 where d between 0 and 1;
 
 set hive.optimize.index.filter=true;
-select sum(hash(*)) from newtypesorc where d between 0 and 1;
+select sum(hash(*)) from newtypesorc_n5 where d between 0 and 1;
 
 set hive.optimize.index.filter=false;
-select sum(hash(*)) from newtypesorc where d between 0 and 1000;
+select sum(hash(*)) from newtypesorc_n5 where d between 0 and 1000;
 
 set hive.optimize.index.filter=true;
-select sum(hash(*)) from newtypesorc where d between 0 and 1000;
+select sum(hash(*)) from newtypesorc_n5 where d between 0 and 1000;
 
 set hive.optimize.index.filter=false;
-select sum(hash(*)) from newtypesorc where d between 0 and '2.0';
+select sum(hash(*)) from newtypesorc_n5 where d between 0 and '2.0';
 
 set hive.optimize.index.filter=true;
-select sum(hash(*)) from newtypesorc where d between 0 and '2.0';
+select sum(hash(*)) from newtypesorc_n5 where d between 0 and '2.0';
 
 set hive.optimize.index.filter=false;
-select sum(hash(*)) from newtypesorc where d between 0 and cast(3 as float);
+select sum(hash(*)) from newtypesorc_n5 where d between 0 and cast(3 as float);
 
 set hive.optimize.index.filter=true;
-select sum(hash(*)) from newtypesorc where d between 0 and cast(3 as float);
+select sum(hash(*)) from newtypesorc_n5 where d between 0 and cast(3 as float);
 
 set hive.optimize.index.filter=false;
-select sum(hash(*)) from newtypesorc where d between 1 and cast(30 as char(10));
+select sum(hash(*)) from newtypesorc_n5 where d between 1 and cast(30 as char(10));
 
 set hive.optimize.index.filter=true;
-select sum(hash(*)) from newtypesorc where d between 1 and cast(30 as char(10));
+select sum(hash(*)) from newtypesorc_n5 where d between 1 and cast(30 as char(10));

http://git-wip-us.apache.org/repos/asf/hive/blob/9bf28a3c/ql/src/test/queries/clientpositive/orc_ppd_same_table_multiple_aliases.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/orc_ppd_same_table_multiple_aliases.q b/ql/src/test/queries/clientpositive/orc_ppd_same_table_multiple_aliases.q
index cc7e3e0..27cf7d6 100644
--- a/ql/src/test/queries/clientpositive/orc_ppd_same_table_multiple_aliases.q
+++ b/ql/src/test/queries/clientpositive/orc_ppd_same_table_multiple_aliases.q
@@ -2,18 +2,18 @@
 
 set hive.vectorized.execution.enabled=false;
 set hive.optimize.index.filter=true;
-create table test_table(number int) stored as ORC;
+create table test_table_n6(number int) stored as ORC;
 
 -- Two insertions will create two files, with one stripe each
-insert into table test_table VALUES (1);
-insert into table test_table VALUES (2);
+insert into table test_table_n6 VALUES (1);
+insert into table test_table_n6 VALUES (2);
 
 -- This should return 2 records
-select * from test_table;
+select * from test_table_n6;
 
 -- These should each return 1 record
-select * from test_table where number = 1;
-select * from test_table where number = 2;
+select * from test_table_n6 where number = 1;
+select * from test_table_n6 where number = 2;
 
 -- This should return 2 records
-select * from test_table where number = 1 union all select * from test_table where number = 2;
+select * from test_table_n6 where number = 1 union all select * from test_table_n6 where number = 2;

http://git-wip-us.apache.org/repos/asf/hive/blob/9bf28a3c/ql/src/test/queries/clientpositive/orc_ppd_schema_evol_1a.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/orc_ppd_schema_evol_1a.q b/ql/src/test/queries/clientpositive/orc_ppd_schema_evol_1a.q
index c31b0c1..595fb66 100644
--- a/ql/src/test/queries/clientpositive/orc_ppd_schema_evol_1a.q
+++ b/ql/src/test/queries/clientpositive/orc_ppd_schema_evol_1a.q
@@ -18,28 +18,28 @@ stored as textfile;
 
 load data local inpath '../../data/files/unique_1.txt' into table unique_1;
 
-create table test1 stored as orc as select * from unique_1;
+create table test1_n1 stored as orc as select * from unique_1;
 
 SET hive.exec.post.hooks=org.apache.hadoop.hive.ql.hooks.PostExecTezSummaryPrinter;
 
-alter table test1 change column i i string;
+alter table test1_n1 change column i i string;
 
 set hive.optimize.ppd=false;
 set hive.optimize.index.filter=false;
 set hive.input.format=org.apache.hadoop.hive.ql.io.CombineHiveInputFormat;
 
-select s from test1 where i = '-1591211872';
+select s from test1_n1 where i = '-1591211872';
 
 set hive.input.format=org.apache.hadoop.hive.ql.io.HiveInputFormat;
 
-select s from test1 where i = -1591211872;
+select s from test1_n1 where i = -1591211872;
 
 set hive.optimize.ppd=true;
 set hive.optimize.index.filter=true;
 set hive.input.format=org.apache.hadoop.hive.ql.io.CombineHiveInputFormat;
 
-select s from test1 where i = '-1591211872';
+select s from test1_n1 where i = '-1591211872';
 
 set hive.input.format=org.apache.hadoop.hive.ql.io.HiveInputFormat;
 
-select s from test1 where i = -1591211872;
+select s from test1_n1 where i = -1591211872;

http://git-wip-us.apache.org/repos/asf/hive/blob/9bf28a3c/ql/src/test/queries/clientpositive/orc_ppd_schema_evol_1b.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/orc_ppd_schema_evol_1b.q b/ql/src/test/queries/clientpositive/orc_ppd_schema_evol_1b.q
index 7cbfbcf..4932222 100644
--- a/ql/src/test/queries/clientpositive/orc_ppd_schema_evol_1b.q
+++ b/ql/src/test/queries/clientpositive/orc_ppd_schema_evol_1b.q
@@ -7,7 +7,7 @@ set hive.input.format=org.apache.hadoop.hive.ql.io.CombineHiveInputFormat;
 SET hive.map.aggr=false;
 -- disabling map side aggregation as that can lead to different intermediate record counts
 
-create table unique_1( 
+create table unique_1_n0( 
 i int, 
 d double, 
 s string) 
@@ -15,7 +15,7 @@ row format delimited
 fields terminated by '|' 
 stored as textfile;
 
-load data local inpath '../../data/files/unique_1.txt' into table unique_1;
+load data local inpath '../../data/files/unique_1.txt' into table unique_1_n0;
 
 create table unique_2( 
 i int, 
@@ -33,7 +33,7 @@ d double,
 s string)
 stored as orc;
 
-insert into table test_two_files select * from unique_1 where i <= 0;
+insert into table test_two_files select * from unique_1_n0 where i <= 0;
 insert into table test_two_files select * from unique_2 where i > 0;
 
 SET hive.exec.post.hooks=org.apache.hadoop.hive.ql.hooks.PostExecTezSummaryPrinter;

http://git-wip-us.apache.org/repos/asf/hive/blob/9bf28a3c/ql/src/test/queries/clientpositive/orc_ppd_schema_evol_2a.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/orc_ppd_schema_evol_2a.q b/ql/src/test/queries/clientpositive/orc_ppd_schema_evol_2a.q
index 0f48730..cacbff3 100644
--- a/ql/src/test/queries/clientpositive/orc_ppd_schema_evol_2a.q
+++ b/ql/src/test/queries/clientpositive/orc_ppd_schema_evol_2a.q
@@ -9,7 +9,7 @@ SET hive.map.aggr=false;
 
 
 
-create table unique_1( 
+create table unique_1_n2( 
 i int, 
 d string, 
 s string) 
@@ -17,30 +17,30 @@ row format delimited
 fields terminated by '|' 
 stored as textfile;
 
-load data local inpath '../../data/files/unique_1.txt' into table unique_1;
+load data local inpath '../../data/files/unique_1.txt' into table unique_1_n2;
 
-create table test1 stored as orc as select * from unique_1 order by d;
+create table test1_n13 stored as orc as select * from unique_1_n2 order by d;
 
 SET hive.exec.post.hooks=org.apache.hadoop.hive.ql.hooks.PostExecTezSummaryPrinter;
 
-alter table test1 change column d d double;
+alter table test1_n13 change column d d double;
 
 set hive.optimize.ppd=false;
 set hive.optimize.index.filter=false;
 set hive.input.format=org.apache.hadoop.hive.ql.io.CombineHiveInputFormat;
 
-select s from test1 where d = -4996703.42;
+select s from test1_n13 where d = -4996703.42;
 
 set hive.input.format=org.apache.hadoop.hive.ql.io.HiveInputFormat;
 
-select s from test1 where d = -4996703.42;
+select s from test1_n13 where d = -4996703.42;
 
 set hive.optimize.ppd=true;
 set hive.optimize.index.filter=true;
 set hive.input.format=org.apache.hadoop.hive.ql.io.CombineHiveInputFormat;
 
-select s from test1 where d = -4996703.42;
+select s from test1_n13 where d = -4996703.42;
 
 set hive.input.format=org.apache.hadoop.hive.ql.io.HiveInputFormat;
 
-select s from test1 where d = -4996703.42;
+select s from test1_n13 where d = -4996703.42;

http://git-wip-us.apache.org/repos/asf/hive/blob/9bf28a3c/ql/src/test/queries/clientpositive/orc_ppd_schema_evol_2b.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/orc_ppd_schema_evol_2b.q b/ql/src/test/queries/clientpositive/orc_ppd_schema_evol_2b.q
index b64471e..7fd8dee 100644
--- a/ql/src/test/queries/clientpositive/orc_ppd_schema_evol_2b.q
+++ b/ql/src/test/queries/clientpositive/orc_ppd_schema_evol_2b.q
@@ -8,7 +8,7 @@ SET hive.map.aggr=false;
 -- disabling map side aggregation as that can lead to different intermediate record counts
 
 
-create table unique_1( 
+create table unique_1_n1( 
 i int, 
 d string, 
 s string) 
@@ -16,9 +16,9 @@ row format delimited
 fields terminated by '|' 
 stored as textfile;
 
-load data local inpath '../../data/files/unique_1.txt' into table unique_1;
+load data local inpath '../../data/files/unique_1.txt' into table unique_1_n1;
 
-create table unique_2( 
+create table unique_2_n0( 
 i int, 
 d string, 
 s string)
@@ -26,40 +26,40 @@ row format delimited
 fields terminated by '|' 
 stored as textfile;
 
-load data local inpath '../../data/files/unique_2.txt' into table unique_2;
+load data local inpath '../../data/files/unique_2.txt' into table unique_2_n0;
 
-create table test_two_files( 
+create table test_two_files_n0( 
 i int, 
 d string, 
 s string)
 stored as orc;
 
-insert into table test_two_files select * from unique_1 where cast(d as double) <= 0 order by cast(d as double);
-insert into table test_two_files select * from unique_2 where cast(d as double) > 0 order by cast(d as double);
+insert into table test_two_files_n0 select * from unique_1_n1 where cast(d as double) <= 0 order by cast(d as double);
+insert into table test_two_files_n0 select * from unique_2_n0 where cast(d as double) > 0 order by cast(d as double);
 
 SET hive.exec.post.hooks=org.apache.hadoop.hive.ql.hooks.PostExecTezSummaryPrinter;
 
-alter table test_two_files change column d d double;
+alter table test_two_files_n0 change column d d double;
 
 set hive.optimize.ppd=false;
 set hive.optimize.index.filter=false;
 set hive.input.format=org.apache.hadoop.hive.ql.io.CombineHiveInputFormat;
 
-select s from test_two_files where d = -4996703.42;
+select s from test_two_files_n0 where d = -4996703.42;
 
 set hive.input.format=org.apache.hadoop.hive.ql.io.HiveInputFormat;
 
-select s from test_two_files where d = -4996703.42;
+select s from test_two_files_n0 where d = -4996703.42;
 
 
 set hive.optimize.ppd=true;
 set hive.optimize.index.filter=true;
 set hive.input.format=org.apache.hadoop.hive.ql.io.CombineHiveInputFormat;
 
-select s from test_two_files where d = -4996703.42;
+select s from test_two_files_n0 where d = -4996703.42;
 
 set hive.input.format=org.apache.hadoop.hive.ql.io.HiveInputFormat;
 
-select s from test_two_files where d = -4996703.42;
+select s from test_two_files_n0 where d = -4996703.42;
 
 

http://git-wip-us.apache.org/repos/asf/hive/blob/9bf28a3c/ql/src/test/queries/clientpositive/orc_ppd_schema_evol_3a.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/orc_ppd_schema_evol_3a.q b/ql/src/test/queries/clientpositive/orc_ppd_schema_evol_3a.q
index e4a9268..4235c2c 100644
--- a/ql/src/test/queries/clientpositive/orc_ppd_schema_evol_3a.q
+++ b/ql/src/test/queries/clientpositive/orc_ppd_schema_evol_3a.q
@@ -5,7 +5,7 @@ SET hive.cbo.enable=false;
 SET hive.map.aggr=false;
 -- disabling map side aggregation as that can lead to different intermediate record counts
 
-CREATE TABLE staging(t tinyint,
+CREATE TABLE staging_n8(t tinyint,
            si smallint,
            i int,
            b bigint,
@@ -19,10 +19,10 @@ CREATE TABLE staging(t tinyint,
 ROW FORMAT DELIMITED FIELDS TERMINATED BY '|'
 STORED AS TEXTFILE;
 
-LOAD DATA LOCAL INPATH '../../data/files/over1k' OVERWRITE INTO TABLE staging;
-LOAD DATA LOCAL INPATH '../../data/files/over1k' INTO TABLE staging;
+LOAD DATA LOCAL INPATH '../../data/files/over1k' OVERWRITE INTO TABLE staging_n8;
+LOAD DATA LOCAL INPATH '../../data/files/over1k' INTO TABLE staging_n8;
 
-CREATE TABLE orc_ppd_staging(t tinyint,
+CREATE TABLE orc_ppd_staging_n2(t tinyint,
            si smallint,
            i int,
            b bigint,
@@ -37,14 +37,14 @@ CREATE TABLE orc_ppd_staging(t tinyint,
            bin binary)
 STORED AS ORC tblproperties("orc.row.index.stride" = "1000", "orc.bloom.filter.columns"="*");
 
-insert overwrite table orc_ppd_staging select t, si, i, b, f, d, bo, s, cast(s as char(50)), cast(s as varchar(50)), cast(ts as date), `dec`, bin from staging order by t, s;
+insert overwrite table orc_ppd_staging_n2 select t, si, i, b, f, d, bo, s, cast(s as char(50)), cast(s as varchar(50)), cast(ts as date), `dec`, bin from staging_n8 order by t, s;
 
 -- just to introduce a gap in min/max range for bloom filters. The dataset has contiguous values
 -- which makes it hard to test bloom filters
-insert into orc_ppd_staging select -10,-321,-65680,-4294967430,-97.94,-13.07,true,"aaa","aaa","aaa","1990-03-11",-71.54,"aaa" from staging limit 1;
-insert into orc_ppd_staging select 127,331,65690,4294967440,107.94,23.07,true,"zzz","zzz","zzz","2023-03-11",71.54,"zzz" from staging limit 1;
+insert into orc_ppd_staging_n2 select -10,-321,-65680,-4294967430,-97.94,-13.07,true,"aaa","aaa","aaa","1990-03-11",-71.54,"aaa" from staging_n8 limit 1;
+insert into orc_ppd_staging_n2 select 127,331,65690,4294967440,107.94,23.07,true,"zzz","zzz","zzz","2023-03-11",71.54,"zzz" from staging_n8 limit 1;
 
-CREATE TABLE orc_ppd(t tinyint,
+CREATE TABLE orc_ppd_n3(t tinyint,
            si smallint,
            i int,
            b bigint,
@@ -59,7 +59,7 @@ CREATE TABLE orc_ppd(t tinyint,
            bin binary)
 STORED AS ORC tblproperties("orc.row.index.stride" = "1000", "orc.bloom.filter.columns"="*");
 
-insert overwrite table orc_ppd select t, si, i, b, f, d, bo, s, cast(s as char(50)), cast(s as varchar(50)), da, `dec`, bin from orc_ppd_staging order by t, s;
+insert overwrite table orc_ppd_n3 select t, si, i, b, f, d, bo, s, cast(s as char(50)), cast(s as varchar(50)), da, `dec`, bin from orc_ppd_staging_n2 order by t, s;
 
 SET hive.exec.post.hooks=org.apache.hadoop.hive.ql.hooks.PostExecTezSummaryPrinter;
 SET hive.optimize.index.filter=false;
@@ -70,177 +70,177 @@ SET hive.optimize.index.filter=false;
 -- Entry 2: count: 100 hasNull: false min: 118 max: 127 sum: 12151 positions: 0,4,119,0,0,244,19
 
 -- INPUT_RECORDS: 0 (no row groups)
-select count(*) from orc_ppd where t > 127;
+select count(*) from orc_ppd_n3 where t > 127;
 SET hive.optimize.index.filter=true;
 -- INPUT_RECORDS: 0 (no row groups)
-select count(*) from orc_ppd where t > 127;
+select count(*) from orc_ppd_n3 where t > 127;
 
 SET hive.optimize.index.filter=false;
 -- INPUT_RECORDS: 1000 (1 row group)
-select count(*) from orc_ppd where t = 55;
+select count(*) from orc_ppd_n3 where t = 55;
 SET hive.optimize.index.filter=true;
 -- INPUT_RECORDS: 1000 (1 row group)
-select count(*) from orc_ppd where t = 55;
+select count(*) from orc_ppd_n3 where t = 55;
 
 SET hive.optimize.index.filter=false;
 -- INPUT_RECORDS: 2000 (2 row groups)
-select count(*) from orc_ppd where t = 54;
+select count(*) from orc_ppd_n3 where t = 54;
 SET hive.optimize.index.filter=true;
 -- INPUT_RECORDS: 2000 (2 row groups)
-select count(*) from orc_ppd where t = 54;
+select count(*) from orc_ppd_n3 where t = 54;
 
-alter table orc_ppd change column t t smallint; 
+alter table orc_ppd_n3 change column t t smallint; 
 
 SET hive.optimize.index.filter=false;
 -- INPUT_RECORDS: 0 (no row groups)
-select count(*) from orc_ppd where t > 127;
+select count(*) from orc_ppd_n3 where t > 127;
 SET hive.optimize.index.filter=true;
 -- INPUT_RECORDS: 0 (no row groups)
-select count(*) from orc_ppd where t > 127;
+select count(*) from orc_ppd_n3 where t > 127;
 
 SET hive.optimize.index.filter=false;
 -- INPUT_RECORDS: 1000 (1 row group)
-select count(*) from orc_ppd where t = 55;
+select count(*) from orc_ppd_n3 where t = 55;
 SET hive.optimize.index.filter=true;
 -- INPUT_RECORDS: 1000 (1 row group)
-select count(*) from orc_ppd where t = 55;
+select count(*) from orc_ppd_n3 where t = 55;
 
 SET hive.optimize.index.filter=false;
 -- INPUT_RECORDS: 2000 (2 row groups)
-select count(*) from orc_ppd where t = 54;
+select count(*) from orc_ppd_n3 where t = 54;
 SET hive.optimize.index.filter=true;
 -- INPUT_RECORDS: 2000 (2 row groups)
-select count(*) from orc_ppd where t = 54;
+select count(*) from orc_ppd_n3 where t = 54;
 
-alter table orc_ppd change column t t int;
+alter table orc_ppd_n3 change column t t int;
 
 SET hive.optimize.index.filter=false;
 -- INPUT_RECORDS: 0 (no row groups)
-select count(*) from orc_ppd where t > 127;
+select count(*) from orc_ppd_n3 where t > 127;
 SET hive.optimize.index.filter=true;
 -- INPUT_RECORDS: 0 (no row groups)
-select count(*) from orc_ppd where t > 127;
+select count(*) from orc_ppd_n3 where t > 127;
 
 SET hive.optimize.index.filter=false;
 -- INPUT_RECORDS: 1000 (1 row group)
-select count(*) from orc_ppd where t = 55;
+select count(*) from orc_ppd_n3 where t = 55;
 SET hive.optimize.index.filter=true;
 -- INPUT_RECORDS: 1000 (1 row group)
-select count(*) from orc_ppd where t = 55;
+select count(*) from orc_ppd_n3 where t = 55;
 
 SET hive.optimize.index.filter=false;
 -- INPUT_RECORDS: 2000 (2 row groups)
-select count(*) from orc_ppd where t = 54;
+select count(*) from orc_ppd_n3 where t = 54;
 SET hive.optimize.index.filter=true;
 -- INPUT_RECORDS: 2000 (2 row groups)
-select count(*) from orc_ppd where t = 54;
+select count(*) from orc_ppd_n3 where t = 54;
 
-alter table orc_ppd change column t t bigint;
+alter table orc_ppd_n3 change column t t bigint;
 
 SET hive.optimize.index.filter=false;
 -- INPUT_RECORDS: 0 (no row groups)
-select count(*) from orc_ppd where t > 127;
+select count(*) from orc_ppd_n3 where t > 127;
 SET hive.optimize.index.filter=true;
 -- INPUT_RECORDS: 0 (no row groups)
-select count(*) from orc_ppd where t > 127;
+select count(*) from orc_ppd_n3 where t > 127;
 
 SET hive.optimize.index.filter=false;
 -- INPUT_RECORDS: 1000 (1 row group)
-select count(*) from orc_ppd where t = 55;
+select count(*) from orc_ppd_n3 where t = 55;
 SET hive.optimize.index.filter=true;
 -- INPUT_RECORDS: 1000 (1 row group)
-select count(*) from orc_ppd where t = 55;
+select count(*) from orc_ppd_n3 where t = 55;
 
 SET hive.optimize.index.filter=false;
 -- INPUT_RECORDS: 2000 (2 row groups)
-select count(*) from orc_ppd where t = 54;
+select count(*) from orc_ppd_n3 where t = 54;
 SET hive.optimize.index.filter=true;
 -- INPUT_RECORDS: 2000 (2 row groups)
-select count(*) from orc_ppd where t = 54;
+select count(*) from orc_ppd_n3 where t = 54;
 
-alter table orc_ppd change column t t string;
+alter table orc_ppd_n3 change column t t string;
 
 SET hive.optimize.index.filter=false;
 -- INPUT_RECORDS: 0 (no row groups)
-select count(*) from orc_ppd where t > '127';
+select count(*) from orc_ppd_n3 where t > '127';
 SET hive.optimize.index.filter=true;
 -- INPUT_RECORDS: 0 (no row groups)
-select count(*) from orc_ppd where t > '127';
+select count(*) from orc_ppd_n3 where t > '127';
 
 SET hive.optimize.index.filter=false;
 -- INPUT_RECORDS: 1000 (1 row group)
-select count(*) from orc_ppd where t = '55';
+select count(*) from orc_ppd_n3 where t = '55';
 SET hive.optimize.index.filter=true;
 -- INPUT_RECORDS: 1000 (1 row group)
-select count(*) from orc_ppd where t = '55';
+select count(*) from orc_ppd_n3 where t = '55';
 
 SET hive.optimize.index.filter=false;
 -- INPUT_RECORDS: 2000 (2 row groups)
-select count(*) from orc_ppd where t = '54';
+select count(*) from orc_ppd_n3 where t = '54';
 SET hive.optimize.index.filter=true;
 -- INPUT_RECORDS: 2000 (2 row groups)
-select count(*) from orc_ppd where t = '54';
+select count(*) from orc_ppd_n3 where t = '54';
 
 SET hive.optimize.index.filter=false;
 -- float tests
-select count(*) from orc_ppd where f = 74.72;
+select count(*) from orc_ppd_n3 where f = 74.72;
 SET hive.optimize.index.filter=true;
-select count(*) from orc_ppd where f = 74.72;
+select count(*) from orc_ppd_n3 where f = 74.72;
 
-alter table orc_ppd change column f f double;
+alter table orc_ppd_n3 change column f f double;
 
 SET hive.optimize.index.filter=false;
-select count(*) from orc_ppd where f = 74.72;
+select count(*) from orc_ppd_n3 where f = 74.72;
 SET hive.optimize.index.filter=true;
-select count(*) from orc_ppd where f = 74.72;
+select count(*) from orc_ppd_n3 where f = 74.72;
 
-alter table orc_ppd change column f f string;
+alter table orc_ppd_n3 change column f f string;
 
 SET hive.optimize.index.filter=false;
-select count(*) from orc_ppd where f = '74.72';
+select count(*) from orc_ppd_n3 where f = '74.72';
 SET hive.optimize.index.filter=true;
-select count(*) from orc_ppd where f = '74.72';
+select count(*) from orc_ppd_n3 where f = '74.72';
 
 SET hive.optimize.index.filter=false;
 -- string tests
-select count(*) from orc_ppd where s = 'bob davidson';
+select count(*) from orc_ppd_n3 where s = 'bob davidson';
 SET hive.optimize.index.filter=true;
-select count(*) from orc_ppd where s = 'bob davidson';
+select count(*) from orc_ppd_n3 where s = 'bob davidson';
 
-alter table orc_ppd change column s s char(50);
+alter table orc_ppd_n3 change column s s char(50);
 
 SET hive.optimize.index.filter=false;
-select count(*) from orc_ppd where s = 'bob davidson';
+select count(*) from orc_ppd_n3 where s = 'bob davidson';
 SET hive.optimize.index.filter=true;
-select count(*) from orc_ppd where s = 'bob davidson';
+select count(*) from orc_ppd_n3 where s = 'bob davidson';
 
-alter table orc_ppd change column s s varchar(50);
+alter table orc_ppd_n3 change column s s varchar(50);
 
 SET hive.optimize.index.filter=false;
-select count(*) from orc_ppd where s = 'bob davidson';
+select count(*) from orc_ppd_n3 where s = 'bob davidson';
 SET hive.optimize.index.filter=true;
-select count(*) from orc_ppd where s = 'bob davidson';
+select count(*) from orc_ppd_n3 where s = 'bob davidson';
 
-alter table orc_ppd change column s s char(50);
+alter table orc_ppd_n3 change column s s char(50);
 
 SET hive.optimize.index.filter=false;
-select count(*) from orc_ppd where s = 'bob davidson';
+select count(*) from orc_ppd_n3 where s = 'bob davidson';
 SET hive.optimize.index.filter=true;
-select count(*) from orc_ppd where s = 'bob davidson';
+select count(*) from orc_ppd_n3 where s = 'bob davidson';
 
-alter table orc_ppd change column s s string;
+alter table orc_ppd_n3 change column s s string;
 
 SET hive.optimize.index.filter=false;
-select count(*) from orc_ppd where s = 'bob davidson';
+select count(*) from orc_ppd_n3 where s = 'bob davidson';
 SET hive.optimize.index.filter=true;
-select count(*) from orc_ppd where s = 'bob davidson';
+select count(*) from orc_ppd_n3 where s = 'bob davidson';
 
-alter table orc_ppd add columns (boo boolean);
+alter table orc_ppd_n3 add columns (boo boolean);
 
 SET hive.optimize.index.filter=false;
 -- ppd on newly added column
-select count(*) from orc_ppd where si = 442;
-select count(*) from orc_ppd where si = 442 or boo is not null or boo = false;
+select count(*) from orc_ppd_n3 where si = 442;
+select count(*) from orc_ppd_n3 where si = 442 or boo is not null or boo = false;
 SET hive.optimize.index.filter=true;
-select count(*) from orc_ppd where si = 442;
-select count(*) from orc_ppd where si = 442 or boo is not null or boo = false;
+select count(*) from orc_ppd_n3 where si = 442;
+select count(*) from orc_ppd_n3 where si = 442 or boo is not null or boo = false;

http://git-wip-us.apache.org/repos/asf/hive/blob/9bf28a3c/ql/src/test/queries/clientpositive/orc_ppd_str_conversion.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/orc_ppd_str_conversion.q b/ql/src/test/queries/clientpositive/orc_ppd_str_conversion.q
index bba6c57..61f4cfa 100644
--- a/ql/src/test/queries/clientpositive/orc_ppd_str_conversion.q
+++ b/ql/src/test/queries/clientpositive/orc_ppd_str_conversion.q
@@ -1,18 +1,18 @@
 set hive.vectorized.execution.enabled=false;
 set hive.cbo.enable=false;
 
-create table orc_test( col1 varchar(15), col2 char(10)) stored as orc;
+create table orc_test_n0( col1 varchar(15), col2 char(10)) stored as orc;
 create table text_test( col1 varchar(15), col2 char(10));
 
-insert into orc_test values ('val1', '1');
-insert overwrite table text_test select * from orc_test;
+insert into orc_test_n0 values ('val1', '1');
+insert overwrite table text_test select * from orc_test_n0;
 
 explain select * from text_test where col2='1';
 select * from text_test where col2='1';
 
 set hive.optimize.index.filter=false;
-select * from orc_test where col2='1';
+select * from orc_test_n0 where col2='1';
 
 set hive.optimize.index.filter=true;
-select * from orc_test where col2='1';
+select * from orc_test_n0 where col2='1';
 

http://git-wip-us.apache.org/repos/asf/hive/blob/9bf28a3c/ql/src/test/queries/clientpositive/orc_ppd_timestamp.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/orc_ppd_timestamp.q b/ql/src/test/queries/clientpositive/orc_ppd_timestamp.q
index 07a77ae..cb1615f 100644
--- a/ql/src/test/queries/clientpositive/orc_ppd_timestamp.q
+++ b/ql/src/test/queries/clientpositive/orc_ppd_timestamp.q
@@ -3,96 +3,96 @@ SET hive.input.format=org.apache.hadoop.hive.ql.io.HiveInputFormat;
 SET mapred.min.split.size=1000;
 SET mapred.max.split.size=5000;
 
-create table newtypesorc(c char(10), v varchar(10), d decimal(5,3), ts timestamp) stored as orc tblproperties("orc.stripe.size"="16777216"); 
+create table newtypesorc_n2(c char(10), v varchar(10), d decimal(5,3), ts timestamp) stored as orc tblproperties("orc.stripe.size"="16777216"); 
 
-insert overwrite table newtypesorc select * from (select cast("apple" as char(10)), cast("bee" as varchar(10)), 0.22, cast("2011-01-01 01:01:01" as timestamp) from src src1 union all select cast("hello" as char(10)), cast("world" as varchar(10)), 11.22, cast("2011-01-20 01:01:01" as timestamp) from src src2) uniontbl;
+insert overwrite table newtypesorc_n2 select * from (select cast("apple" as char(10)), cast("bee" as varchar(10)), 0.22, cast("2011-01-01 01:01:01" as timestamp) from src src1 union all select cast("hello" as char(10)), cast("world" as varchar(10)), 11.22, cast("2011-01-20 01:01:01" as timestamp) from src src2) uniontbl;
 
 -- timestamp data types (EQUAL, NOT_EQUAL, LESS_THAN, LESS_THAN_EQUALS, IN, BETWEEN tests)
-select sum(hash(*)) from newtypesorc where cast(ts as string)='2011-01-01 01:01:01';
+select sum(hash(*)) from newtypesorc_n2 where cast(ts as string)='2011-01-01 01:01:01';
 
 set hive.optimize.index.filter=true;
-select sum(hash(*)) from newtypesorc where cast(ts as string)='2011-01-01 01:01:01';
+select sum(hash(*)) from newtypesorc_n2 where cast(ts as string)='2011-01-01 01:01:01';
 
 set hive.optimize.index.filter=false;
-select sum(hash(*)) from newtypesorc where ts=cast('2011-01-01 01:01:01' as timestamp);
+select sum(hash(*)) from newtypesorc_n2 where ts=cast('2011-01-01 01:01:01' as timestamp);
 
 set hive.optimize.index.filter=true;
-select sum(hash(*)) from newtypesorc where ts=cast('2011-01-01 01:01:01' as timestamp);
+select sum(hash(*)) from newtypesorc_n2 where ts=cast('2011-01-01 01:01:01' as timestamp);
 
 set hive.optimize.index.filter=false;
-select sum(hash(*)) from newtypesorc where ts=cast('2011-01-01 01:01:01' as varchar(20));
+select sum(hash(*)) from newtypesorc_n2 where ts=cast('2011-01-01 01:01:01' as varchar(20));
 
 set hive.optimize.index.filter=true;
-select sum(hash(*)) from newtypesorc where ts=cast('2011-01-01 01:01:01' as varchar(20));
+select sum(hash(*)) from newtypesorc_n2 where ts=cast('2011-01-01 01:01:01' as varchar(20));
 
 set hive.optimize.index.filter=false;
-select sum(hash(*)) from newtypesorc where ts!=cast('2011-01-01 01:01:01' as timestamp);
+select sum(hash(*)) from newtypesorc_n2 where ts!=cast('2011-01-01 01:01:01' as timestamp);
 
 set hive.optimize.index.filter=true;
-select sum(hash(*)) from newtypesorc where ts!=cast('2011-01-01 01:01:01' as timestamp);
+select sum(hash(*)) from newtypesorc_n2 where ts!=cast('2011-01-01 01:01:01' as timestamp);
 
 set hive.optimize.index.filter=false;
-select sum(hash(*)) from newtypesorc where ts<cast('2011-01-20 01:01:01' as timestamp);
+select sum(hash(*)) from newtypesorc_n2 where ts<cast('2011-01-20 01:01:01' as timestamp);
 
 set hive.optimize.index.filter=true;
-select sum(hash(*)) from newtypesorc where ts<cast('2011-01-20 01:01:01' as timestamp);
+select sum(hash(*)) from newtypesorc_n2 where ts<cast('2011-01-20 01:01:01' as timestamp);
 
 set hive.optimize.index.filter=false;
-select sum(hash(*)) from newtypesorc where ts<cast('2011-01-22 01:01:01' as timestamp);
+select sum(hash(*)) from newtypesorc_n2 where ts<cast('2011-01-22 01:01:01' as timestamp);
 
 set hive.optimize.index.filter=true;
-select sum(hash(*)) from newtypesorc where ts<cast('2011-01-22 01:01:01' as timestamp);
+select sum(hash(*)) from newtypesorc_n2 where ts<cast('2011-01-22 01:01:01' as timestamp);
 
 set hive.optimize.index.filter=false;
-select sum(hash(*)) from newtypesorc where ts<cast('2010-10-01 01:01:01' as timestamp);
+select sum(hash(*)) from newtypesorc_n2 where ts<cast('2010-10-01 01:01:01' as timestamp);
 
 set hive.optimize.index.filter=true;
-select sum(hash(*)) from newtypesorc where ts<cast('2010-10-01 01:01:01' as timestamp);
+select sum(hash(*)) from newtypesorc_n2 where ts<cast('2010-10-01 01:01:01' as timestamp);
 
 set hive.optimize.index.filter=false;
-select sum(hash(*)) from newtypesorc where ts<=cast('2011-01-01 01:01:01' as timestamp);
+select sum(hash(*)) from newtypesorc_n2 where ts<=cast('2011-01-01 01:01:01' as timestamp);
 
 set hive.optimize.index.filter=true;
-select sum(hash(*)) from newtypesorc where ts<=cast('2011-01-01 01:01:01' as timestamp);
+select sum(hash(*)) from newtypesorc_n2 where ts<=cast('2011-01-01 01:01:01' as timestamp);
 
 set hive.optimize.index.filter=false;
-select sum(hash(*)) from newtypesorc where ts<=cast('2011-01-20 01:01:01' as timestamp);
+select sum(hash(*)) from newtypesorc_n2 where ts<=cast('2011-01-20 01:01:01' as timestamp);
 
 set hive.optimize.index.filter=true;
-select sum(hash(*)) from newtypesorc where ts<=cast('2011-01-20 01:01:01' as timestamp);
+select sum(hash(*)) from newtypesorc_n2 where ts<=cast('2011-01-20 01:01:01' as timestamp);
 
 set hive.optimize.index.filter=false;
-select sum(hash(*)) from newtypesorc where ts in (cast('2011-01-02 01:01:01' as timestamp), cast('2011-01-20 01:01:01' as timestamp));
+select sum(hash(*)) from newtypesorc_n2 where ts in (cast('2011-01-02 01:01:01' as timestamp), cast('2011-01-20 01:01:01' as timestamp));
 
 set hive.optimize.index.filter=true;
-select sum(hash(*)) from newtypesorc where ts in (cast('2011-01-02 01:01:01' as timestamp), cast('2011-01-20 01:01:01' as timestamp));
+select sum(hash(*)) from newtypesorc_n2 where ts in (cast('2011-01-02 01:01:01' as timestamp), cast('2011-01-20 01:01:01' as timestamp));
 
 set hive.optimize.index.filter=false;
-select sum(hash(*)) from newtypesorc where ts in (cast('2011-01-01 01:01:01' as timestamp), cast('2011-01-20 01:01:01' as timestamp));
+select sum(hash(*)) from newtypesorc_n2 where ts in (cast('2011-01-01 01:01:01' as timestamp), cast('2011-01-20 01:01:01' as timestamp));
 
 set hive.optimize.index.filter=true;
-select sum(hash(*)) from newtypesorc where ts in (cast('2011-01-01 01:01:01' as timestamp), cast('2011-01-20 01:01:01' as timestamp));
+select sum(hash(*)) from newtypesorc_n2 where ts in (cast('2011-01-01 01:01:01' as timestamp), cast('2011-01-20 01:01:01' as timestamp));
 
 set hive.optimize.index.filter=false;
-select sum(hash(*)) from newtypesorc where ts in (cast('2011-01-02 01:01:01' as timestamp), cast('2011-01-08 01:01:01' as timestamp));
+select sum(hash(*)) from newtypesorc_n2 where ts in (cast('2011-01-02 01:01:01' as timestamp), cast('2011-01-08 01:01:01' as timestamp));
 
 set hive.optimize.index.filter=true;
-select sum(hash(*)) from newtypesorc where ts in (cast('2011-01-02 01:01:01' as timestamp), cast('2011-01-08 01:01:01' as timestamp));
+select sum(hash(*)) from newtypesorc_n2 where ts in (cast('2011-01-02 01:01:01' as timestamp), cast('2011-01-08 01:01:01' as timestamp));
 
 set hive.optimize.index.filter=false;
-select sum(hash(*)) from newtypesorc where ts between cast('2010-10-01 01:01:01' as timestamp) and cast('2011-01-08 01:01:01' as timestamp);
+select sum(hash(*)) from newtypesorc_n2 where ts between cast('2010-10-01 01:01:01' as timestamp) and cast('2011-01-08 01:01:01' as timestamp);
 
 set hive.optimize.index.filter=true;
-select sum(hash(*)) from newtypesorc where ts between cast('2010-10-01 01:01:01' as timestamp) and cast('2011-01-08 01:01:01' as timestamp);
+select sum(hash(*)) from newtypesorc_n2 where ts between cast('2010-10-01 01:01:01' as timestamp) and cast('2011-01-08 01:01:01' as timestamp);
 
 set hive.optimize.index.filter=false;
-select sum(hash(*)) from newtypesorc where ts between cast('2010-10-01 01:01:01' as timestamp) and cast('2011-01-25 01:01:01' as timestamp);
+select sum(hash(*)) from newtypesorc_n2 where ts between cast('2010-10-01 01:01:01' as timestamp) and cast('2011-01-25 01:01:01' as timestamp);
 
 set hive.optimize.index.filter=true;
-select sum(hash(*)) from newtypesorc where ts between cast('2010-10-01 01:01:01' as timestamp) and cast('2011-01-25 01:01:01' as timestamp);
+select sum(hash(*)) from newtypesorc_n2 where ts between cast('2010-10-01 01:01:01' as timestamp) and cast('2011-01-25 01:01:01' as timestamp);
 
 set hive.optimize.index.filter=false;
-select sum(hash(*)) from newtypesorc where ts between cast('2010-10-01 01:01:01' as timestamp) and cast('2010-11-01 01:01:01' as timestamp);
+select sum(hash(*)) from newtypesorc_n2 where ts between cast('2010-10-01 01:01:01' as timestamp) and cast('2010-11-01 01:01:01' as timestamp);
 
 set hive.optimize.index.filter=true;
-select sum(hash(*)) from newtypesorc where ts between cast('2010-10-01 01:01:01' as timestamp) and cast('2010-11-01 01:01:01' as timestamp);
+select sum(hash(*)) from newtypesorc_n2 where ts between cast('2010-10-01 01:01:01' as timestamp) and cast('2010-11-01 01:01:01' as timestamp);

http://git-wip-us.apache.org/repos/asf/hive/blob/9bf28a3c/ql/src/test/queries/clientpositive/orc_ppd_varchar.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/orc_ppd_varchar.q b/ql/src/test/queries/clientpositive/orc_ppd_varchar.q
index 07f5fa9..f331854 100644
--- a/ql/src/test/queries/clientpositive/orc_ppd_varchar.q
+++ b/ql/src/test/queries/clientpositive/orc_ppd_varchar.q
@@ -3,75 +3,75 @@ SET hive.input.format=org.apache.hadoop.hive.ql.io.HiveInputFormat;
 SET mapred.min.split.size=1000;
 SET mapred.max.split.size=5000;
 
-create table newtypesorc(c char(10), v varchar(10), d decimal(5,3), da date) stored as orc tblproperties("orc.stripe.size"="16777216"); 
+create table newtypesorc_n1(c char(10), v varchar(10), d decimal(5,3), da date) stored as orc tblproperties("orc.stripe.size"="16777216"); 
 
-insert overwrite table newtypesorc select * from (select cast("apple" as char(10)), cast("bee" as varchar(10)), 0.22, cast("1970-02-20" as date) from src src1 union all select cast("hello" as char(10)), cast("world" as varchar(10)), 11.22, cast("1970-02-27" as date) from src src2) uniontbl;
+insert overwrite table newtypesorc_n1 select * from (select cast("apple" as char(10)), cast("bee" as varchar(10)), 0.22, cast("1970-02-20" as date) from src src1 union all select cast("hello" as char(10)), cast("world" as varchar(10)), 11.22, cast("1970-02-27" as date) from src src2) uniontbl;
 
 set hive.optimize.index.filter=false;
 
 -- varchar data types (EQUAL, NOT_EQUAL, LESS_THAN, LESS_THAN_EQUALS, IN, BETWEEN tests)
-select sum(hash(*)) from newtypesorc where v="bee";
+select sum(hash(*)) from newtypesorc_n1 where v="bee";
 
 set hive.optimize.index.filter=true;
-select sum(hash(*)) from newtypesorc where v="bee";
+select sum(hash(*)) from newtypesorc_n1 where v="bee";
 
 set hive.optimize.index.filter=false;
-select sum(hash(*)) from newtypesorc where v!="bee";
+select sum(hash(*)) from newtypesorc_n1 where v!="bee";
 
 set hive.optimize.index.filter=true;
-select sum(hash(*)) from newtypesorc where v!="bee";
+select sum(hash(*)) from newtypesorc_n1 where v!="bee";
 
 set hive.optimize.index.filter=false;
-select sum(hash(*)) from newtypesorc where v<"world";
+select sum(hash(*)) from newtypesorc_n1 where v<"world";
 
 set hive.optimize.index.filter=true;
-select sum(hash(*)) from newtypesorc where v<"world";
+select sum(hash(*)) from newtypesorc_n1 where v<"world";
 
 set hive.optimize.index.filter=false;
-select sum(hash(*)) from newtypesorc where v<="world";
+select sum(hash(*)) from newtypesorc_n1 where v<="world";
 
 set hive.optimize.index.filter=true;
-select sum(hash(*)) from newtypesorc where v<="world";
+select sum(hash(*)) from newtypesorc_n1 where v<="world";
 
 set hive.optimize.index.filter=false;
-select sum(hash(*)) from newtypesorc where v="bee   ";
+select sum(hash(*)) from newtypesorc_n1 where v="bee   ";
 
 set hive.optimize.index.filter=true;
-select sum(hash(*)) from newtypesorc where v="bee   ";
+select sum(hash(*)) from newtypesorc_n1 where v="bee   ";
 
 set hive.optimize.index.filter=false;
-select sum(hash(*)) from newtypesorc where v in ("bee", "orange");
+select sum(hash(*)) from newtypesorc_n1 where v in ("bee", "orange");
 
 set hive.optimize.index.filter=true;
-select sum(hash(*)) from newtypesorc where v in ("bee", "orange");
+select sum(hash(*)) from newtypesorc_n1 where v in ("bee", "orange");
 
 set hive.optimize.index.filter=false;
-select sum(hash(*)) from newtypesorc where v in ("bee", "world");
+select sum(hash(*)) from newtypesorc_n1 where v in ("bee", "world");
 
 set hive.optimize.index.filter=true;
-select sum(hash(*)) from newtypesorc where v in ("bee", "world");
+select sum(hash(*)) from newtypesorc_n1 where v in ("bee", "world");
 
 set hive.optimize.index.filter=false;
-select sum(hash(*)) from newtypesorc where v in ("orange");
+select sum(hash(*)) from newtypesorc_n1 where v in ("orange");
 
 set hive.optimize.index.filter=true;
-select sum(hash(*)) from newtypesorc where v in ("orange");
+select sum(hash(*)) from newtypesorc_n1 where v in ("orange");
 
 set hive.optimize.index.filter=false;
-select sum(hash(*)) from newtypesorc where v between "bee" and "orange";
+select sum(hash(*)) from newtypesorc_n1 where v between "bee" and "orange";
 
 set hive.optimize.index.filter=true;
-select sum(hash(*)) from newtypesorc where v between "bee" and "orange";
+select sum(hash(*)) from newtypesorc_n1 where v between "bee" and "orange";
 
 set hive.optimize.index.filter=false;
-select sum(hash(*)) from newtypesorc where v between "bee" and "zombie";
+select sum(hash(*)) from newtypesorc_n1 where v between "bee" and "zombie";
 
 set hive.optimize.index.filter=true;
-select sum(hash(*)) from newtypesorc where v between "bee" and "zombie";
+select sum(hash(*)) from newtypesorc_n1 where v between "bee" and "zombie";
 
 set hive.optimize.index.filter=false;
-select sum(hash(*)) from newtypesorc where v between "orange" and "pine";
+select sum(hash(*)) from newtypesorc_n1 where v between "orange" and "pine";
 
 set hive.optimize.index.filter=true;
-select sum(hash(*)) from newtypesorc where v between "orange" and "pine";
+select sum(hash(*)) from newtypesorc_n1 where v between "orange" and "pine";
 

http://git-wip-us.apache.org/repos/asf/hive/blob/9bf28a3c/ql/src/test/queries/clientpositive/orc_predicate_pushdown.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/orc_predicate_pushdown.q b/ql/src/test/queries/clientpositive/orc_predicate_pushdown.q
index bc75ecf..6795f3e 100644
--- a/ql/src/test/queries/clientpositive/orc_predicate_pushdown.q
+++ b/ql/src/test/queries/clientpositive/orc_predicate_pushdown.q
@@ -18,7 +18,7 @@ STORED AS ORC;
 
 ALTER TABLE orc_pred SET SERDEPROPERTIES ('orc.row.index.stride' = '1000');
 
-CREATE TABLE staging(t tinyint,
+CREATE TABLE staging_n2(t tinyint,
            si smallint,
            i int,
            b bigint,
@@ -32,9 +32,9 @@ CREATE TABLE staging(t tinyint,
 ROW FORMAT DELIMITED FIELDS TERMINATED BY '|'
 STORED AS TEXTFILE;
 
-LOAD DATA LOCAL INPATH '../../data/files/over1k' OVERWRITE INTO TABLE staging;
+LOAD DATA LOCAL INPATH '../../data/files/over1k' OVERWRITE INTO TABLE staging_n2;
 
-INSERT INTO TABLE orc_pred select * from staging;
+INSERT INTO TABLE orc_pred select * from staging_n2;
 
 -- no predicate case. the explain plan should not have filter expression in table scan operator
 

http://git-wip-us.apache.org/repos/asf/hive/blob/9bf28a3c/ql/src/test/queries/clientpositive/orc_schema_evolution.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/orc_schema_evolution.q b/ql/src/test/queries/clientpositive/orc_schema_evolution.q
index 8241bb7..50a5c5e 100644
--- a/ql/src/test/queries/clientpositive/orc_schema_evolution.q
+++ b/ql/src/test/queries/clientpositive/orc_schema_evolution.q
@@ -2,21 +2,21 @@ set hive.vectorized.execution.enabled=false;
 set hive.fetch.task.conversion=none;
 SET hive.exec.schema.evolution=true;
 
-create table src_orc (key smallint, val string) stored as orc;
+create table src_orc_n3 (key smallint, val string) stored as orc;
 create table src_orc2 (key smallint, val string) stored as orc;
 
 -- integer type widening
-insert overwrite table src_orc select * from src;
-select sum(hash(*)) from src_orc;
+insert overwrite table src_orc_n3 select * from src;
+select sum(hash(*)) from src_orc_n3;
 
-alter table src_orc change key key smallint;
-select sum(hash(*)) from src_orc;
+alter table src_orc_n3 change key key smallint;
+select sum(hash(*)) from src_orc_n3;
 
-alter table src_orc change key key int;
-select sum(hash(*)) from src_orc;
+alter table src_orc_n3 change key key int;
+select sum(hash(*)) from src_orc_n3;
 
-alter table src_orc change key key bigint;
-select sum(hash(*)) from src_orc;
+alter table src_orc_n3 change key key bigint;
+select sum(hash(*)) from src_orc_n3;
 
 -- replace columns for adding columns and type widening
 insert overwrite table src_orc2 select * from src;

http://git-wip-us.apache.org/repos/asf/hive/blob/9bf28a3c/ql/src/test/queries/clientpositive/orc_wide_table.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/orc_wide_table.q b/ql/src/test/queries/clientpositive/orc_wide_table.q
index 1c418ee..08e2e76 100644
--- a/ql/src/test/queries/clientpositive/orc_wide_table.q
+++ b/ql/src/test/queries/clientpositive/orc_wide_table.q
@@ -3,7 +3,7 @@ set hive.mapred.mode=nonstrict;
 set hive.stats.column.autogather=false;
 
 drop table if exists test_txt;
-drop table if exists test_orc;
+drop table if exists test_orc_n1;
 create table test_txt(
 	c1 varchar(64),
 	c2 int,
@@ -2007,10 +2007,10 @@ create table test_txt(
 	c2000 boolean
 ) row format delimited fields terminated by ',';
 load data local inpath '../../data/files/2000_cols_data.csv' overwrite into table test_txt;
-create table test_orc like test_txt;
-alter table test_orc set fileformat orc;
-insert into table test_orc select * from test_txt;
+create table test_orc_n1 like test_txt;
+alter table test_orc_n1 set fileformat orc;
+insert into table test_orc_n1 select * from test_txt;
 
 select c1, c2, c1999 from test_txt order by c1;
-select c1, c2, c1999 from test_orc order by c1;
+select c1, c2, c1999 from test_orc_n1 order by c1;
 

http://git-wip-us.apache.org/repos/asf/hive/blob/9bf28a3c/ql/src/test/queries/clientpositive/order3.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/order3.q b/ql/src/test/queries/clientpositive/order3.q
index 3fdf917..c29afe2 100644
--- a/ql/src/test/queries/clientpositive/order3.q
+++ b/ql/src/test/queries/clientpositive/order3.q
@@ -1,26 +1,26 @@
 -- Test HIVE-16019
-drop table if exists test;
-create table test(key int, value1 int, value2 string);
-insert into table test values (1, 1, 'val111'), (1, 2, 'val121'), (1, 3, 'val131'), (2, 1, 'val211'), (2, 2, 'val221'), (2, 2, 'val222'), (2, 3, 'val231'), (2, 4, 'val241'),
+drop table if exists test_n0;
+create table test_n0(key int, value1 int, value2 string);
+insert into table test_n0 values (1, 1, 'val111'), (1, 2, 'val121'), (1, 3, 'val131'), (2, 1, 'val211'), (2, 2, 'val221'), (2, 2, 'val222'), (2, 3, 'val231'), (2, 4, 'val241'),
 (3, 1, 'val311'), (3, 2, 'val321'), (3, 2, 'val322'), (3, 3, 'val331'), (3, 3, 'val332'), (3, 3, 'val333'), (4, 1, 'val411');
 
 set hive.cbo.enable=true;
-EXPLAIN SELECT T1.KEY AS MYKEY FROM TEST T1 GROUP BY T1.KEY ORDER BY T1.KEY LIMIT 3;
-SELECT T1.KEY AS MYKEY FROM TEST T1 GROUP BY T1.KEY ORDER BY T1.KEY LIMIT 3;
+EXPLAIN SELECT T1.KEY AS MYKEY FROM TEST_n0 T1 GROUP BY T1.KEY ORDER BY T1.KEY LIMIT 3;
+SELECT T1.KEY AS MYKEY FROM TEST_n0 T1 GROUP BY T1.KEY ORDER BY T1.KEY LIMIT 3;
 
-EXPLAIN SELECT T1.KEY AS MYKEY, MAX(T1.VALUE1) AS MYVALUE1 FROM TEST T1 GROUP BY T1.KEY ORDER BY T1.KEY LIMIT 3;
-SELECT T1.KEY AS MYKEY, MAX(T1.VALUE1) AS MYVALUE1 FROM TEST T1 GROUP BY T1.KEY ORDER BY T1.KEY LIMIT 3;
+EXPLAIN SELECT T1.KEY AS MYKEY, MAX(T1.VALUE1) AS MYVALUE1 FROM TEST_n0 T1 GROUP BY T1.KEY ORDER BY T1.KEY LIMIT 3;
+SELECT T1.KEY AS MYKEY, MAX(T1.VALUE1) AS MYVALUE1 FROM TEST_n0 T1 GROUP BY T1.KEY ORDER BY T1.KEY LIMIT 3;
 
-EXPLAIN SELECT T1.KEY AS MYKEY, COUNT(T1.VALUE1) AS MYVALUE1, 'AAA' AS C FROM TEST T1 GROUP BY T1.KEY, 'AAA' ORDER BY T1.KEY, 'AAA' LIMIT 3;
-SELECT T1.KEY AS MYKEY, COUNT(T1.VALUE1) AS MYVALUE1, 'AAA' AS C FROM TEST T1 GROUP BY T1.KEY, 'AAA' ORDER BY T1.KEY, 'AAA' LIMIT 3;
+EXPLAIN SELECT T1.KEY AS MYKEY, COUNT(T1.VALUE1) AS MYVALUE1, 'AAA' AS C FROM TEST_n0 T1 GROUP BY T1.KEY, 'AAA' ORDER BY T1.KEY, 'AAA' LIMIT 3;
+SELECT T1.KEY AS MYKEY, COUNT(T1.VALUE1) AS MYVALUE1, 'AAA' AS C FROM TEST_n0 T1 GROUP BY T1.KEY, 'AAA' ORDER BY T1.KEY, 'AAA' LIMIT 3;
 
 set hive.cbo.enable=false;
-EXPLAIN SELECT T1.KEY AS MYKEY FROM TEST T1 GROUP BY T1.KEY ORDER BY T1.KEY LIMIT 3;
-SELECT T1.KEY AS MYKEY FROM TEST T1 GROUP BY T1.KEY ORDER BY T1.KEY LIMIT 3;
+EXPLAIN SELECT T1.KEY AS MYKEY FROM TEST_n0 T1 GROUP BY T1.KEY ORDER BY T1.KEY LIMIT 3;
+SELECT T1.KEY AS MYKEY FROM TEST_n0 T1 GROUP BY T1.KEY ORDER BY T1.KEY LIMIT 3;
 
-EXPLAIN SELECT T1.KEY AS MYKEY, MAX(T1.VALUE1) AS MYVALUE1 FROM TEST T1 GROUP BY T1.KEY ORDER BY T1.KEY LIMIT 3;
-SELECT T1.KEY AS MYKEY, MAX(T1.VALUE1) AS MYVALUE1 FROM TEST T1 GROUP BY T1.KEY ORDER BY T1.KEY LIMIT 3;
+EXPLAIN SELECT T1.KEY AS MYKEY, MAX(T1.VALUE1) AS MYVALUE1 FROM TEST_n0 T1 GROUP BY T1.KEY ORDER BY T1.KEY LIMIT 3;
+SELECT T1.KEY AS MYKEY, MAX(T1.VALUE1) AS MYVALUE1 FROM TEST_n0 T1 GROUP BY T1.KEY ORDER BY T1.KEY LIMIT 3;
 
-EXPLAIN SELECT T1.KEY AS MYKEY, COUNT(T1.VALUE1) AS MYVALUE1, 'AAA' AS C FROM TEST T1 GROUP BY T1.KEY, 'AAA' ORDER BY T1.KEY, 'AAA' LIMIT 3;
-SELECT T1.KEY AS MYKEY, COUNT(T1.VALUE1) AS MYVALUE1, 'AAA' AS C FROM TEST T1 GROUP BY T1.KEY, 'AAA' ORDER BY T1.KEY, 'AAA' LIMIT 3;
+EXPLAIN SELECT T1.KEY AS MYKEY, COUNT(T1.VALUE1) AS MYVALUE1, 'AAA' AS C FROM TEST_n0 T1 GROUP BY T1.KEY, 'AAA' ORDER BY T1.KEY, 'AAA' LIMIT 3;
+SELECT T1.KEY AS MYKEY, COUNT(T1.VALUE1) AS MYVALUE1, 'AAA' AS C FROM TEST_n0 T1 GROUP BY T1.KEY, 'AAA' ORDER BY T1.KEY, 'AAA' LIMIT 3;
 

http://git-wip-us.apache.org/repos/asf/hive/blob/9bf28a3c/ql/src/test/queries/clientpositive/order_by_expr_1.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/order_by_expr_1.q b/ql/src/test/queries/clientpositive/order_by_expr_1.q
index 1d99e6a..86911a5 100644
--- a/ql/src/test/queries/clientpositive/order_by_expr_1.q
+++ b/ql/src/test/queries/clientpositive/order_by_expr_1.q
@@ -1,44 +1,44 @@
 set hive.fetch.task.conversion=none;
 
-create table t(a int, b int);
+create table t_n5(a int, b int);
 
-insert into t values (1,2),(1,2),(1,3),(2,4),(20,-100),(-1000,100),(4,5),(3,7),(8,9);
+insert into t_n5 values (1,2),(1,2),(1,3),(2,4),(20,-100),(-1000,100),(4,5),(3,7),(8,9);
 
-select a, count(a) from t group by a order by count(a), a;
+select a, count(a) from t_n5 group by a order by count(a), a;
 
 explain
 select
   interval '2-2' year to month + interval '3-3' year to month,
   interval '2-2' year to month - interval '3-3' year to month
-from t 
+from t_n5 
 order by interval '2-2' year to month + interval '3-3' year to month
 limit 2;
 
-select a,b, count(*) from t group by a, b order by a+b;
-select a,b, count(*) from t group by a, b order by count(*), b desc; 
-select a,b,count(*),a+b from t group by a, b order by a+b;
-select a,b from t order by a+b;
-select a,b,a+b from t order by a+b;
-select a,b,a+b from t order by a+b desc;
-select cast(0.99999999999999999999 as decimal(20,19)) as c from t limit 1;
-select cast(0.99999999999999999999 as decimal(20,19)) as c from t order by c limit 1;
-select a from t order by b;
-select a from t order by 0-b;
-select b from t order by 0-b;
-select b from t order by a, 0-b;
-select b from t order by a+1, 0-b;
-select b from t order by 0-b, a+1;
-explain select b from t order by 0-b, a+1;
-select a,b from t order by 0-b;
-select a,b from t order by a, a+1, 0-b;
-select a,b from t order by 0-b, a+1;
-select a+1,b from t order by a, a+1, 0-b;
-select a+1 as c, b from t order by a, a+1, 0-b;
-select a, a+1 as c, b from t order by a, a+1, 0-b;
-select a, a+1 as c, b, 2*b from t order by a, a+1, 0-b;
-explain select a, a+1 as c, b, 2*b from t order by a, a+1, 0-b;
-select a, a+1 as c, b, 2*b from t order by a+1, 0-b;
-select a,b, count(*) as c from t group by a, b order by c, a+b desc; 
+select a,b, count(*) from t_n5 group by a, b order by a+b;
+select a,b, count(*) from t_n5 group by a, b order by count(*), b desc; 
+select a,b,count(*),a+b from t_n5 group by a, b order by a+b;
+select a,b from t_n5 order by a+b;
+select a,b,a+b from t_n5 order by a+b;
+select a,b,a+b from t_n5 order by a+b desc;
+select cast(0.99999999999999999999 as decimal(20,19)) as c from t_n5 limit 1;
+select cast(0.99999999999999999999 as decimal(20,19)) as c from t_n5 order by c limit 1;
+select a from t_n5 order by b;
+select a from t_n5 order by 0-b;
+select b from t_n5 order by 0-b;
+select b from t_n5 order by a, 0-b;
+select b from t_n5 order by a+1, 0-b;
+select b from t_n5 order by 0-b, a+1;
+explain select b from t_n5 order by 0-b, a+1;
+select a,b from t_n5 order by 0-b;
+select a,b from t_n5 order by a, a+1, 0-b;
+select a,b from t_n5 order by 0-b, a+1;
+select a+1,b from t_n5 order by a, a+1, 0-b;
+select a+1 as c, b from t_n5 order by a, a+1, 0-b;
+select a, a+1 as c, b from t_n5 order by a, a+1, 0-b;
+select a, a+1 as c, b, 2*b from t_n5 order by a, a+1, 0-b;
+explain select a, a+1 as c, b, 2*b from t_n5 order by a, a+1, 0-b;
+select a, a+1 as c, b, 2*b from t_n5 order by a+1, 0-b;
+select a,b, count(*) as c from t_n5 group by a, b order by c, a+b desc; 
 
-select a, max(b) from t group by a order by count(b), a desc; 
-select a, max(b) from t group by a order by count(b), a; 
+select a, max(b) from t_n5 group by a order by count(b), a desc; 
+select a, max(b) from t_n5 group by a order by count(b), a; 

http://git-wip-us.apache.org/repos/asf/hive/blob/9bf28a3c/ql/src/test/queries/clientpositive/order_by_expr_2.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/order_by_expr_2.q b/ql/src/test/queries/clientpositive/order_by_expr_2.q
index 043f8ed..759186d 100644
--- a/ql/src/test/queries/clientpositive/order_by_expr_2.q
+++ b/ql/src/test/queries/clientpositive/order_by_expr_2.q
@@ -1,11 +1,11 @@
 set hive.fetch.task.conversion=none;
 
-create table t(a int, b int);
+create table t_n14(a int, b int);
 
-insert into t values (1,2),(1,2),(1,3),(2,4),(20,-100),(-1000,100),(4,5),(3,7),(8,9);
+insert into t_n14 values (1,2),(1,2),(1,3),(2,4),(20,-100),(-1000,100),(4,5),(3,7),(8,9);
 
-select a as b, b as a from t order by a;
-select a as b, b as a from t order by t.a;
-select a as b from t order by b;
-select a as b from t order by 0-a;
-select a,b,count(*),a+b from t group by a, b order by a+b;
+select a as b, b as a from t_n14 order by a;
+select a as b, b as a from t_n14 order by t_n14.a;
+select a as b from t_n14 order by b;
+select a as b from t_n14 order by 0-a;
+select a,b,count(*),a+b from t_n14 group by a, b order by a+b;

http://git-wip-us.apache.org/repos/asf/hive/blob/9bf28a3c/ql/src/test/queries/clientpositive/order_by_pos.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/order_by_pos.q b/ql/src/test/queries/clientpositive/order_by_pos.q
index 744abe5..ab4f226 100644
--- a/ql/src/test/queries/clientpositive/order_by_pos.q
+++ b/ql/src/test/queries/clientpositive/order_by_pos.q
@@ -1,20 +1,20 @@
 set hive.fetch.task.conversion=none;
 
-create table t(a int, b int);
+create table t_n3(a int, b int);
 
-insert into t values (1,2),(1,2),(1,3),(2,4),(20,-100),(-1000,100),(4,5),(3,7),(8,9);
+insert into t_n3 values (1,2),(1,2),(1,3),(2,4),(20,-100),(-1000,100),(4,5),(3,7),(8,9);
 
-select * from t order by 2;
+select * from t_n3 order by 2;
 
-select * from t order by 1;
+select * from t_n3 order by 1;
 
-select * from t union select * from t order by 1, 2;
+select * from t_n3 union select * from t_n3 order by 1, 2;
 
-select * from t union select * from t order by 2;
+select * from t_n3 union select * from t_n3 order by 2;
 
-select * from t union select * from t order by 1;
+select * from t_n3 union select * from t_n3 order by 1;
 
-select * from (select a, count(a) from t group by a)subq order by 2, 1;
+select * from (select a, count(a) from t_n3 group by a)subq order by 2, 1;
 
-select * from (select a,b, count(*) from t group by a, b)subq order by 3, 2 desc;
+select * from (select a,b, count(*) from t_n3 group by a, b)subq order by 3, 2 desc;
  

http://git-wip-us.apache.org/repos/asf/hive/blob/9bf28a3c/ql/src/test/queries/clientpositive/order_null.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/order_null.q b/ql/src/test/queries/clientpositive/order_null.q
index ea833e0..fbbe7da 100644
--- a/ql/src/test/queries/clientpositive/order_null.q
+++ b/ql/src/test/queries/clientpositive/order_null.q
@@ -1,29 +1,29 @@
-create table src_null (a int, b string);
-insert into src_null values (1, 'A');
-insert into src_null values (null, null);
-insert into src_null values (3, null);
-insert into src_null values (2, null);
-insert into src_null values (2, 'A');
-insert into src_null values (2, 'B');
+create table src_null_n1 (a int, b string);
+insert into src_null_n1 values (1, 'A');
+insert into src_null_n1 values (null, null);
+insert into src_null_n1 values (3, null);
+insert into src_null_n1 values (2, null);
+insert into src_null_n1 values (2, 'A');
+insert into src_null_n1 values (2, 'B');
 
-SELECT x.* FROM src_null x ORDER BY a asc;
+SELECT x.* FROM src_null_n1 x ORDER BY a asc;
 
-SELECT x.* FROM src_null x ORDER BY a desc;
+SELECT x.* FROM src_null_n1 x ORDER BY a desc;
 
-SELECT x.* FROM src_null x ORDER BY b asc, a asc nulls last;
+SELECT x.* FROM src_null_n1 x ORDER BY b asc, a asc nulls last;
 
-SELECT x.* FROM src_null x ORDER BY b desc, a asc;
+SELECT x.* FROM src_null_n1 x ORDER BY b desc, a asc;
 
-SELECT x.* FROM src_null x ORDER BY a asc nulls first;
+SELECT x.* FROM src_null_n1 x ORDER BY a asc nulls first;
 
-SELECT x.* FROM src_null x ORDER BY a desc nulls first;
+SELECT x.* FROM src_null_n1 x ORDER BY a desc nulls first;
 
-SELECT x.* FROM src_null x ORDER BY b asc nulls last, a;
+SELECT x.* FROM src_null_n1 x ORDER BY b asc nulls last, a;
 
-SELECT x.* FROM src_null x ORDER BY b desc nulls last, a;
+SELECT x.* FROM src_null_n1 x ORDER BY b desc nulls last, a;
 
-SELECT x.* FROM src_null x ORDER BY a asc nulls last, b desc;
+SELECT x.* FROM src_null_n1 x ORDER BY a asc nulls last, b desc;
 
-SELECT x.* FROM src_null x ORDER BY b desc nulls last, a desc nulls last;
+SELECT x.* FROM src_null_n1 x ORDER BY b desc nulls last, a desc nulls last;
 
-SELECT x.* FROM src_null x ORDER BY b asc nulls first, a asc nulls last;
+SELECT x.* FROM src_null_n1 x ORDER BY b asc nulls first, a asc nulls last;

http://git-wip-us.apache.org/repos/asf/hive/blob/9bf28a3c/ql/src/test/queries/clientpositive/parallel.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/parallel.q b/ql/src/test/queries/clientpositive/parallel.q
index f2f0d35..fdcc1c0 100644
--- a/ql/src/test/queries/clientpositive/parallel.q
+++ b/ql/src/test/queries/clientpositive/parallel.q
@@ -6,26 +6,26 @@ set hive.input.format=org.apache.hadoop.hive.ql.io.HiveInputFormat;
 -- SORT_QUERY_RESULTS
 
 create table if not exists src_a like src;
-create table if not exists src_b like src;
+create table if not exists src_b_n0 like src;
 
 explain
 from (select key, value from src group by key, value) s
 insert overwrite table src_a select s.key, s.value group by s.key, s.value
-insert overwrite table src_b select s.key, s.value group by s.key, s.value;
+insert overwrite table src_b_n0 select s.key, s.value group by s.key, s.value;
 
 from (select key, value from src group by key, value) s
 insert overwrite table src_a select s.key, s.value group by s.key, s.value
-insert overwrite table src_b select s.key, s.value group by s.key, s.value;
+insert overwrite table src_b_n0 select s.key, s.value group by s.key, s.value;
 
 select * from src_a;
-select * from src_b;
+select * from src_b_n0;
 
 
 set hive.input.format=org.apache.hadoop.hive.ql.io.CombineHiveInputFormat;
 
 from (select key, value from src group by key, value) s
 insert overwrite table src_a select s.key, s.value group by s.key, s.value
-insert overwrite table src_b select s.key, s.value group by s.key, s.value;
+insert overwrite table src_b_n0 select s.key, s.value group by s.key, s.value;
 
 select * from src_a;
-select * from src_b;
+select * from src_b_n0;

http://git-wip-us.apache.org/repos/asf/hive/blob/9bf28a3c/ql/src/test/queries/clientpositive/parallel_colstats.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/parallel_colstats.q b/ql/src/test/queries/clientpositive/parallel_colstats.q
index 86603c1..5673f5b 100644
--- a/ql/src/test/queries/clientpositive/parallel_colstats.q
+++ b/ql/src/test/queries/clientpositive/parallel_colstats.q
@@ -6,27 +6,27 @@ set hive.stats.column.autogather=true;
 
 -- SORT_QUERY_RESULTS
 
-create table if not exists src_a like src;
-create table if not exists src_b like src;
+create table if not exists src_a_n0 like src;
+create table if not exists src_b_n1 like src;
 
 explain
 from (select key, value from src group by key, value) s
-insert overwrite table src_a select s.key, s.value group by s.key, s.value
-insert overwrite table src_b select s.key, s.value group by s.key, s.value;
+insert overwrite table src_a_n0 select s.key, s.value group by s.key, s.value
+insert overwrite table src_b_n1 select s.key, s.value group by s.key, s.value;
 
 from (select key, value from src group by key, value) s
-insert overwrite table src_a select s.key, s.value group by s.key, s.value
-insert overwrite table src_b select s.key, s.value group by s.key, s.value;
+insert overwrite table src_a_n0 select s.key, s.value group by s.key, s.value
+insert overwrite table src_b_n1 select s.key, s.value group by s.key, s.value;
 
-select * from src_a;
-select * from src_b;
+select * from src_a_n0;
+select * from src_b_n1;
 
 
 set hive.input.format=org.apache.hadoop.hive.ql.io.CombineHiveInputFormat;
 
 from (select key, value from src group by key, value) s
-insert overwrite table src_a select s.key, s.value group by s.key, s.value
-insert overwrite table src_b select s.key, s.value group by s.key, s.value;
+insert overwrite table src_a_n0 select s.key, s.value group by s.key, s.value
+insert overwrite table src_b_n1 select s.key, s.value group by s.key, s.value;
 
-select * from src_a;
-select * from src_b;
+select * from src_a_n0;
+select * from src_b_n1;

http://git-wip-us.apache.org/repos/asf/hive/blob/9bf28a3c/ql/src/test/queries/clientpositive/parallel_join1.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/parallel_join1.q b/ql/src/test/queries/clientpositive/parallel_join1.q
index ac92efd..7ac96b3 100644
--- a/ql/src/test/queries/clientpositive/parallel_join1.q
+++ b/ql/src/test/queries/clientpositive/parallel_join1.q
@@ -4,13 +4,13 @@ set hive.input.format=org.apache.hadoop.hive.ql.io.HiveInputFormat;
 
 -- SORT_QUERY_RESULTS
 
-CREATE TABLE dest_j1(key INT, value STRING) STORED AS TEXTFILE;
+CREATE TABLE dest_j1_n19(key INT, value STRING) STORED AS TEXTFILE;
 
 EXPLAIN
 FROM src src1 JOIN src src2 ON (src1.key = src2.key)
-INSERT OVERWRITE TABLE dest_j1 SELECT src1.key, src2.value;
+INSERT OVERWRITE TABLE dest_j1_n19 SELECT src1.key, src2.value;
 
 FROM src src1 JOIN src src2 ON (src1.key = src2.key)
-INSERT OVERWRITE TABLE dest_j1 SELECT src1.key, src2.value;
+INSERT OVERWRITE TABLE dest_j1_n19 SELECT src1.key, src2.value;
 
-SELECT dest_j1.* FROM dest_j1;
+SELECT dest_j1_n19.* FROM dest_j1_n19;

http://git-wip-us.apache.org/repos/asf/hive/blob/9bf28a3c/ql/src/test/queries/clientpositive/parallel_orderby.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/parallel_orderby.q b/ql/src/test/queries/clientpositive/parallel_orderby.q
index 9e76ac4..41fe9e8 100644
--- a/ql/src/test/queries/clientpositive/parallel_orderby.q
+++ b/ql/src/test/queries/clientpositive/parallel_orderby.q
@@ -1,15 +1,15 @@
 set hive.mapred.mode=nonstrict;
-create table src5 (key string, value string);
-load data local inpath '../../data/files/kv5.txt' into table src5;
-load data local inpath '../../data/files/kv5.txt' into table src5;
+create table src5_n2 (key string, value string);
+load data local inpath '../../data/files/kv5.txt' into table src5_n2;
+load data local inpath '../../data/files/kv5.txt' into table src5_n2;
 
 set mapred.reduce.tasks = 4;
 set hive.optimize.sampling.orderby=true;
 set hive.optimize.sampling.orderby.percent=0.66f;
 
 explain
-create table total_ordered as select * from src5 order by key, value;
-create table total_ordered as select * from src5 order by key, value;
+create table total_ordered as select * from src5_n2 order by key, value;
+create table total_ordered as select * from src5_n2 order by key, value;
 
 desc formatted total_ordered;
 select * from total_ordered;
@@ -18,7 +18,7 @@ set hive.optimize.sampling.orderby.percent=0.0001f;
 -- rolling back to single task in case that the number of sample is not enough
 
 drop table total_ordered;
-create table total_ordered as select * from src5 order by key, value;
+create table total_ordered as select * from src5_n2 order by key, value;
 
 desc formatted total_ordered;
 select * from total_ordered;

http://git-wip-us.apache.org/repos/asf/hive/blob/9bf28a3c/ql/src/test/queries/clientpositive/parquet_ctas.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/parquet_ctas.q b/ql/src/test/queries/clientpositive/parquet_ctas.q
index 431bfc1..e61b069 100644
--- a/ql/src/test/queries/clientpositive/parquet_ctas.q
+++ b/ql/src/test/queries/clientpositive/parquet_ctas.q
@@ -1,26 +1,27 @@
 set hive.vectorized.execution.enabled=false;
 set hive.mapred.mode=nonstrict;
-drop table staging;
+
+drop table staging_n3;
 drop table parquet_ctas;
 drop table parquet_ctas_advanced;
 drop table parquet_ctas_alias;
 drop table parquet_ctas_mixed;
 
-create table staging (key int, value string) stored as textfile;
-insert into table staging select * from src order by key limit 10;
+create table staging_n3 (key int, value string) stored as textfile;
+insert into table staging_n3 select * from src order by key limit 10;
 
-create table parquet_ctas stored as parquet as select * from staging;
+create table parquet_ctas stored as parquet as select * from staging_n3;
 describe parquet_ctas;
 select * from parquet_ctas;
 
-create table parquet_ctas_advanced stored as parquet as select key+1,concat(value,"value") from staging;
+create table parquet_ctas_advanced stored as parquet as select key+1,concat(value,"value") from staging_n3;
 describe parquet_ctas_advanced;
 select * from parquet_ctas_advanced;
 
-create table parquet_ctas_alias stored as parquet as select key+1 as mykey,concat(value,"value") as myvalue from staging;
+create table parquet_ctas_alias stored as parquet as select key+1 as mykey,concat(value,"value") as myvalue from staging_n3;
 describe parquet_ctas_alias;
 select * from parquet_ctas_alias;
 
-create table parquet_ctas_mixed stored as parquet as select key,key+1,concat(value,"value") as myvalue from staging;
+create table parquet_ctas_mixed stored as parquet as select key,key+1,concat(value,"value") as myvalue from staging_n3;
 describe parquet_ctas_mixed;
 select * from parquet_ctas_mixed;
\ No newline at end of file

http://git-wip-us.apache.org/repos/asf/hive/blob/9bf28a3c/ql/src/test/queries/clientpositive/parquet_decimal.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/parquet_decimal.q b/ql/src/test/queries/clientpositive/parquet_decimal.q
index 1d3559a..dfed699 100644
--- a/ql/src/test/queries/clientpositive/parquet_decimal.q
+++ b/ql/src/test/queries/clientpositive/parquet_decimal.q
@@ -1,29 +1,29 @@
 set hive.vectorized.execution.enabled=false;
 set hive.mapred.mode=nonstrict;
 
-DROP TABLE IF EXISTS `dec`;
+DROP TABLE IF EXISTS `dec_n1`;
 
-CREATE TABLE `dec`(name string, value decimal(8,4));
+CREATE TABLE `dec_n1`(name string, value decimal(8,4));
 
-LOAD DATA LOCAL INPATH '../../data/files/dec.txt' INTO TABLE `dec`;
+LOAD DATA LOCAL INPATH '../../data/files/dec.txt' INTO TABLE `dec_n1`;
 
-DROP TABLE IF EXISTS parq_dec;
+DROP TABLE IF EXISTS parq_dec_n1;
 
-CREATE TABLE parq_dec(name string, value decimal(5,2)) STORED AS PARQUET;
+CREATE TABLE parq_dec_n1(name string, value decimal(5,2)) STORED AS PARQUET;
 
-DESC parq_dec;
+DESC parq_dec_n1;
 
-INSERT OVERWRITE TABLE parq_dec SELECT name, value FROM `dec`;
+INSERT OVERWRITE TABLE parq_dec_n1 SELECT name, value FROM `dec_n1`;
 
-SELECT * FROM parq_dec;
+SELECT * FROM parq_dec_n1;
 
-SELECT value, count(*) FROM parq_dec GROUP BY value ORDER BY value;
+SELECT value, count(*) FROM parq_dec_n1 GROUP BY value ORDER BY value;
 
-TRUNCATE TABLE parq_dec;
+TRUNCATE TABLE parq_dec_n1;
 
-INSERT OVERWRITE TABLE parq_dec SELECT name, NULL FROM `dec`;
+INSERT OVERWRITE TABLE parq_dec_n1 SELECT name, NULL FROM `dec_n1`;
 
-SELECT * FROM parq_dec;
+SELECT * FROM parq_dec_n1;
 
 DROP TABLE IF EXISTS parq_dec1;
 
@@ -35,6 +35,6 @@ LOAD DATA LOCAL INPATH '../../data/files/dec.parq' INTO TABLE parq_dec1;
 
 SELECT VALUE FROM parq_dec1;
 
-DROP TABLE `dec`;
-DROP TABLE parq_dec;
+DROP TABLE `dec_n1`;
+DROP TABLE parq_dec_n1;
 DROP TABLE parq_dec1;

http://git-wip-us.apache.org/repos/asf/hive/blob/9bf28a3c/ql/src/test/queries/clientpositive/parquet_join.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/parquet_join.q b/ql/src/test/queries/clientpositive/parquet_join.q
index 0bb5482..caf93ce 100644
--- a/ql/src/test/queries/clientpositive/parquet_join.q
+++ b/ql/src/test/queries/clientpositive/parquet_join.q
@@ -1,18 +1,18 @@
 set hive.vectorized.execution.enabled=false;
 set hive.mapred.mode=nonstrict;
 
-drop table if exists staging;
+drop table if exists staging_n5;
 drop table if exists parquet_jointable1;
 drop table if exists parquet_jointable2;
 drop table if exists parquet_jointable1_bucketed_sorted;
 drop table if exists parquet_jointable2_bucketed_sorted;
 
-create table staging (key int, value string) stored as textfile;
-insert into table staging select distinct key, value from src order by key limit 2;
+create table staging_n5 (key int, value string) stored as textfile;
+insert into table staging_n5 select distinct key, value from src order by key limit 2;
 
-create table parquet_jointable1 stored as parquet as select * from staging;
+create table parquet_jointable1 stored as parquet as select * from staging_n5;
 
-create table parquet_jointable2 stored as parquet as select key,key+1,concat(value,"value") as myvalue from staging;
+create table parquet_jointable2 stored as parquet as select key,key+1,concat(value,"value") as myvalue from staging_n5;
 
 -- SORT_QUERY_RESULTS
 
@@ -40,8 +40,8 @@ set hive.input.format=org.apache.hadoop.hive.ql.io.BucketizedHiveInputFormat;
 -- SMB join
 
 create table parquet_jointable1_bucketed_sorted (key int,value string) clustered by (key) sorted by (key ASC) INTO 1 BUCKETS stored as parquet;
-insert overwrite table parquet_jointable1_bucketed_sorted select key,concat(value,"value1") as value from staging cluster by key;
+insert overwrite table parquet_jointable1_bucketed_sorted select key,concat(value,"value1") as value from staging_n5 cluster by key;
 create table parquet_jointable2_bucketed_sorted (key int,value1 string, value2 string) clustered by (key) sorted by (key ASC) INTO 1 BUCKETS stored as parquet;
-insert overwrite table parquet_jointable2_bucketed_sorted select key,concat(value,"value2-1") as value1,concat(value,"value2-2") as value2 from staging cluster by key;
+insert overwrite table parquet_jointable2_bucketed_sorted select key,concat(value,"value2-1") as value1,concat(value,"value2-2") as value2 from staging_n5 cluster by key;
 explain select p1.value,p2.value2 from parquet_jointable1_bucketed_sorted p1 join parquet_jointable2_bucketed_sorted p2 on p1.key=p2.key;
 select p1.value,p2.value2 from parquet_jointable1_bucketed_sorted p1 join parquet_jointable2_bucketed_sorted p2 on p1.key=p2.key;

http://git-wip-us.apache.org/repos/asf/hive/blob/9bf28a3c/ql/src/test/queries/clientpositive/parquet_join2.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/parquet_join2.q b/ql/src/test/queries/clientpositive/parquet_join2.q
index bd0ce4b..6981689 100644
--- a/ql/src/test/queries/clientpositive/parquet_join2.q
+++ b/ql/src/test/queries/clientpositive/parquet_join2.q
@@ -2,14 +2,14 @@ set hive.vectorized.execution.enabled=false;
 set hive.optimize.index.filter = true;
 set hive.auto.convert.join=false;
 
-CREATE TABLE tbl1(id INT) STORED AS PARQUET;
-INSERT INTO tbl1 VALUES(1), (2);
+CREATE TABLE tbl1_n6(id INT) STORED AS PARQUET;
+INSERT INTO tbl1_n6 VALUES(1), (2);
 
-CREATE TABLE tbl2(id INT, value STRING) STORED AS PARQUET;
-INSERT INTO tbl2 VALUES(1, 'value1');
-INSERT INTO tbl2 VALUES(1, 'value2');
+CREATE TABLE tbl2_n5(id INT, value STRING) STORED AS PARQUET;
+INSERT INTO tbl2_n5 VALUES(1, 'value1');
+INSERT INTO tbl2_n5 VALUES(1, 'value2');
 
-select tbl1.id, t1.value, t2.value
-FROM tbl1
-JOIN (SELECT * FROM tbl2 WHERE value='value1') t1 ON tbl1.id=t1.id
-JOIN (SELECT * FROM tbl2 WHERE value='value2') t2 ON tbl1.id=t2.id;
+select tbl1_n6.id, t1.value, t2.value
+FROM tbl1_n6
+JOIN (SELECT * FROM tbl2_n5 WHERE value='value1') t1 ON tbl1_n6.id=t1.id
+JOIN (SELECT * FROM tbl2_n5 WHERE value='value2') t2 ON tbl1_n6.id=t2.id;