You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hive.apache.org by jc...@apache.org on 2018/05/25 00:58:01 UTC

[29/58] [abbrv] [partial] hive git commit: HIVE-19617: Rename test tables to avoid collisions during execution in batches (Jesus Camacho Rodriguez, reviewed by Gunther Hagleitner)

http://git-wip-us.apache.org/repos/asf/hive/blob/9bf28a3c/ql/src/test/queries/clientpositive/join_is_not_distinct_from.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/join_is_not_distinct_from.q b/ql/src/test/queries/clientpositive/join_is_not_distinct_from.q
index ebe832d..aec7dea 100644
--- a/ql/src/test/queries/clientpositive/join_is_not_distinct_from.q
+++ b/ql/src/test/queries/clientpositive/join_is_not_distinct_from.q
@@ -1,71 +1,71 @@
 set hive.explain.user=false;
 -- SORT_QUERY_RESULTS
 
-CREATE TABLE myinput1(key int, value int);
-LOAD DATA LOCAL INPATH '../../data/files/in8.txt' INTO TABLE myinput1;
+CREATE TABLE myinput1_n10(key int, value int);
+LOAD DATA LOCAL INPATH '../../data/files/in8.txt' INTO TABLE myinput1_n10;
 
 -- merging
-explain select * from myinput1 a join myinput1 b on a.key is not distinct from b.value;
+explain select * from myinput1_n10 a join myinput1_n10 b on a.key is not distinct from b.value;
 -- SORT_QUERY_RESULTS
-select * from myinput1 a join myinput1 b on a.key is not distinct from b.value;
+select * from myinput1_n10 a join myinput1_n10 b on a.key is not distinct from b.value;
 
-explain select * from myinput1 a join myinput1 b on a.key is not distinct from b.value join myinput1 c on a.key=c.key;
-select * from myinput1 a join myinput1 b on a.key is not distinct from b.value join myinput1 c on a.key=c.key;
+explain select * from myinput1_n10 a join myinput1_n10 b on a.key is not distinct from b.value join myinput1_n10 c on a.key=c.key;
+select * from myinput1_n10 a join myinput1_n10 b on a.key is not distinct from b.value join myinput1_n10 c on a.key=c.key;
 
-explain select * from myinput1 a join myinput1 b on a.key is not distinct from b.value join myinput1 c on a.key is not distinct from c.key;
-select * from myinput1 a join myinput1 b on a.key is not distinct from b.value join myinput1 c on a.key is not distinct from c.key;
+explain select * from myinput1_n10 a join myinput1_n10 b on a.key is not distinct from b.value join myinput1_n10 c on a.key is not distinct from c.key;
+select * from myinput1_n10 a join myinput1_n10 b on a.key is not distinct from b.value join myinput1_n10 c on a.key is not distinct from c.key;
 
-explain select * from myinput1 a join myinput1 b on a.key is not distinct from b.value AND a.value=b.key join myinput1 c on a.key is not distinct from c.key AND a.value=c.value;
+explain select * from myinput1_n10 a join myinput1_n10 b on a.key is not distinct from b.value AND a.value=b.key join myinput1_n10 c on a.key is not distinct from c.key AND a.value=c.value;
 
-select * from myinput1 a join myinput1 b on a.key is not distinct from b.value AND a.value=b.key join myinput1 c on a.key is not distinct from c.key AND a.value=c.value;
+select * from myinput1_n10 a join myinput1_n10 b on a.key is not distinct from b.value AND a.value=b.key join myinput1_n10 c on a.key is not distinct from c.key AND a.value=c.value;
 
-explain select * from myinput1 a join myinput1 b on a.key is not distinct from b.value AND a.value is not distinct from b.key join myinput1 c on a.key is not distinct from c.key AND a.value is not distinct from c.value;
-select * from myinput1 a join myinput1 b on a.key is not distinct from b.value AND a.value is not distinct from b.key join myinput1 c on a.key is not distinct from c.key AND a.value is not distinct from c.value;
+explain select * from myinput1_n10 a join myinput1_n10 b on a.key is not distinct from b.value AND a.value is not distinct from b.key join myinput1_n10 c on a.key is not distinct from c.key AND a.value is not distinct from c.value;
+select * from myinput1_n10 a join myinput1_n10 b on a.key is not distinct from b.value AND a.value is not distinct from b.key join myinput1_n10 c on a.key is not distinct from c.key AND a.value is not distinct from c.value;
 
 -- outer joins
-SELECT * FROM myinput1 a LEFT OUTER JOIN myinput1 b ON a.key is not distinct from b.value;
-SELECT * FROM myinput1 a RIGHT OUTER JOIN myinput1 b ON a.key is not distinct from b.value;
-SELECT * FROM myinput1 a FULL OUTER JOIN myinput1 b ON a.key is not distinct from b.value;
+SELECT * FROM myinput1_n10 a LEFT OUTER JOIN myinput1_n10 b ON a.key is not distinct from b.value;
+SELECT * FROM myinput1_n10 a RIGHT OUTER JOIN myinput1_n10 b ON a.key is not distinct from b.value;
+SELECT * FROM myinput1_n10 a FULL OUTER JOIN myinput1_n10 b ON a.key is not distinct from b.value;
 
 -- map joins
-SELECT /*+ MAPJOIN(a) */ * FROM myinput1 a JOIN myinput1 b ON a.key is not distinct from b.value;
-SELECT /*+ MAPJOIN(b) */ * FROM myinput1 a JOIN myinput1 b ON a.key is not distinct from b.value;
+SELECT /*+ MAPJOIN(a) */ * FROM myinput1_n10 a JOIN myinput1_n10 b ON a.key is not distinct from b.value;
+SELECT /*+ MAPJOIN(b) */ * FROM myinput1_n10 a JOIN myinput1_n10 b ON a.key is not distinct from b.value;
 
-CREATE TABLE smb_input(key int, value int);
-LOAD DATA LOCAL INPATH '../../data/files/in4.txt' into table smb_input;
-LOAD DATA LOCAL INPATH '../../data/files/in5.txt' into table smb_input;
+CREATE TABLE smb_input_n2(key int, value int);
+LOAD DATA LOCAL INPATH '../../data/files/in4.txt' into table smb_input_n2;
+LOAD DATA LOCAL INPATH '../../data/files/in5.txt' into table smb_input_n2;
 
 
 ;
 
 -- smbs
-CREATE TABLE smb_input1(key int, value int) CLUSTERED BY (key) SORTED BY (key) INTO 2 BUCKETS;
-CREATE TABLE smb_input2(key int, value int) CLUSTERED BY (value) SORTED BY (value) INTO 2 BUCKETS;
+CREATE TABLE smb_input1_n5(key int, value int) CLUSTERED BY (key) SORTED BY (key) INTO 2 BUCKETS;
+CREATE TABLE smb_input2_n5(key int, value int) CLUSTERED BY (value) SORTED BY (value) INTO 2 BUCKETS;
 
-from smb_input
-insert overwrite table smb_input1 select *
-insert overwrite table smb_input2 select *;
+from smb_input_n2
+insert overwrite table smb_input1_n5 select *
+insert overwrite table smb_input2_n5 select *;
 
 SET hive.optimize.bucketmapjoin = true;
 SET hive.optimize.bucketmapjoin.sortedmerge = true;
 SET hive.input.format = org.apache.hadoop.hive.ql.io.BucketizedHiveInputFormat;
 
-SELECT /*+ MAPJOIN(a) */ * FROM smb_input1 a JOIN smb_input1 b ON a.key  is not distinct from  b.key;
-SELECT /*+ MAPJOIN(a) */ * FROM smb_input1 a JOIN smb_input1 b ON a.key  is not distinct from  b.key AND a.value  is not distinct from  b.value;
-SELECT /*+ MAPJOIN(a) */ * FROM smb_input1 a RIGHT OUTER JOIN smb_input1 b ON a.key  is not distinct from  b.key;
-SELECT /*+ MAPJOIN(b) */ * FROM smb_input1 a JOIN smb_input1 b ON a.key  is not distinct from  b.key;
-SELECT /*+ MAPJOIN(b) */ * FROM smb_input1 a LEFT OUTER JOIN smb_input1 b ON a.key  is not distinct from  b.key;
+SELECT /*+ MAPJOIN(a) */ * FROM smb_input1_n5 a JOIN smb_input1_n5 b ON a.key  is not distinct from  b.key;
+SELECT /*+ MAPJOIN(a) */ * FROM smb_input1_n5 a JOIN smb_input1_n5 b ON a.key  is not distinct from  b.key AND a.value  is not distinct from  b.value;
+SELECT /*+ MAPJOIN(a) */ * FROM smb_input1_n5 a RIGHT OUTER JOIN smb_input1_n5 b ON a.key  is not distinct from  b.key;
+SELECT /*+ MAPJOIN(b) */ * FROM smb_input1_n5 a JOIN smb_input1_n5 b ON a.key  is not distinct from  b.key;
+SELECT /*+ MAPJOIN(b) */ * FROM smb_input1_n5 a LEFT OUTER JOIN smb_input1_n5 b ON a.key  is not distinct from  b.key;
 
-SELECT /*+ MAPJOIN(a) */ * FROM smb_input1 a JOIN smb_input2 b ON a.key  is not distinct from  b.value;
-SELECT /*+ MAPJOIN(b) */ * FROM smb_input1 a JOIN smb_input2 b ON a.key  is not distinct from  b.value;
-SELECT /*+ MAPJOIN(b) */ * FROM smb_input1 a LEFT OUTER JOIN smb_input2 b ON a.key  is not distinct from  b.value;
-SELECT /*+ MAPJOIN(a) */ * FROM smb_input1 a RIGHT OUTER JOIN smb_input2 b ON a.key  is not distinct from  b.value;
+SELECT /*+ MAPJOIN(a) */ * FROM smb_input1_n5 a JOIN smb_input2_n5 b ON a.key  is not distinct from  b.value;
+SELECT /*+ MAPJOIN(b) */ * FROM smb_input1_n5 a JOIN smb_input2_n5 b ON a.key  is not distinct from  b.value;
+SELECT /*+ MAPJOIN(b) */ * FROM smb_input1_n5 a LEFT OUTER JOIN smb_input2_n5 b ON a.key  is not distinct from  b.value;
+SELECT /*+ MAPJOIN(a) */ * FROM smb_input1_n5 a RIGHT OUTER JOIN smb_input2_n5 b ON a.key  is not distinct from  b.value;
 
-SELECT /*+ MAPJOIN(a) */ * FROM smb_input2 a JOIN smb_input2 b ON a.value  is not distinct from  b.value;
-SELECT /*+ MAPJOIN(a) */ * FROM smb_input2 a RIGHT OUTER JOIN smb_input2 b ON a.value  is not distinct from  b.value;
-SELECT /*+ MAPJOIN(b) */ * FROM smb_input2 a JOIN smb_input2 b ON a.value  is not distinct from  b.value;
-SELECT /*+ MAPJOIN(b) */ * FROM smb_input2 a LEFT OUTER JOIN smb_input2 b ON a.value  is not distinct from  b.value;
+SELECT /*+ MAPJOIN(a) */ * FROM smb_input2_n5 a JOIN smb_input2_n5 b ON a.value  is not distinct from  b.value;
+SELECT /*+ MAPJOIN(a) */ * FROM smb_input2_n5 a RIGHT OUTER JOIN smb_input2_n5 b ON a.value  is not distinct from  b.value;
+SELECT /*+ MAPJOIN(b) */ * FROM smb_input2_n5 a JOIN smb_input2_n5 b ON a.value  is not distinct from  b.value;
+SELECT /*+ MAPJOIN(b) */ * FROM smb_input2_n5 a LEFT OUTER JOIN smb_input2_n5 b ON a.value  is not distinct from  b.value;
 
 --HIVE-3315 join predicate transitive
-explain select * from myinput1 a join myinput1 b on a.key is not distinct from b.value AND a.key is NULL;
-select * from myinput1 a join myinput1 b on a.key is not distinct from b.value AND a.key is NULL;
+explain select * from myinput1_n10 a join myinput1_n10 b on a.key is not distinct from b.value AND a.key is NULL;
+select * from myinput1_n10 a join myinput1_n10 b on a.key is not distinct from b.value AND a.key is NULL;

http://git-wip-us.apache.org/repos/asf/hive/blob/9bf28a3c/ql/src/test/queries/clientpositive/join_map_ppr.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/join_map_ppr.q b/ql/src/test/queries/clientpositive/join_map_ppr.q
index 840d303..a3f8f24 100644
--- a/ql/src/test/queries/clientpositive/join_map_ppr.q
+++ b/ql/src/test/queries/clientpositive/join_map_ppr.q
@@ -1,22 +1,22 @@
 -- SORT_QUERY_RESULTS
 
-CREATE TABLE dest_j1(key STRING, value STRING, val2 STRING) STORED AS TEXTFILE;
+CREATE TABLE dest_j1_n4(key STRING, value STRING, val2 STRING) STORED AS TEXTFILE;
 set hive.cbo.enable=false;
 
 EXPLAIN EXTENDED
-INSERT OVERWRITE TABLE dest_j1
+INSERT OVERWRITE TABLE dest_j1_n4
 SELECT /*+ MAPJOIN(x,y) */ x.key, z.value, y.value
 FROM src1 x JOIN src y ON (x.key = y.key) 
 JOIN srcpart z ON (x.key = z.key)
 WHERE z.ds='2008-04-08' and z.hr=11;
 
-INSERT OVERWRITE TABLE dest_j1
+INSERT OVERWRITE TABLE dest_j1_n4
 SELECT /*+ MAPJOIN(x,y) */ x.key, z.value, y.value
 FROM src1 x JOIN src y ON (x.key = y.key) 
 JOIN srcpart z ON (x.key = z.key)
 WHERE z.ds='2008-04-08' and z.hr=11;
 
-select * from dest_j1;
+select * from dest_j1_n4;
 
 CREATE TABLE src_copy(key int, value string);
 CREATE TABLE src1_copy(key string, value string);
@@ -24,19 +24,19 @@ INSERT OVERWRITE TABLE src_copy select key, value from src;
 INSERT OVERWRITE TABLE src1_copy select key, value from src1;
 
 EXPLAIN EXTENDED
-INSERT OVERWRITE TABLE dest_j1
+INSERT OVERWRITE TABLE dest_j1_n4
 SELECT /*+ MAPJOIN(x,y) */ x.key, z.value, y.value
 FROM src1_copy x JOIN src_copy y ON (x.key = y.key) 
 JOIN srcpart z ON (x.key = z.key)
 WHERE z.ds='2008-04-08' and z.hr=11;
 
-INSERT OVERWRITE TABLE dest_j1
+INSERT OVERWRITE TABLE dest_j1_n4
 SELECT /*+ MAPJOIN(x,y) */ x.key, z.value, y.value
 FROM src1_copy x JOIN src_copy y ON (x.key = y.key) 
 JOIN srcpart z ON (x.key = z.key)
 WHERE z.ds='2008-04-08' and z.hr=11;
 
-select * from dest_j1;
+select * from dest_j1_n4;
 
 
 

http://git-wip-us.apache.org/repos/asf/hive/blob/9bf28a3c/ql/src/test/queries/clientpositive/join_nulls.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/join_nulls.q b/ql/src/test/queries/clientpositive/join_nulls.q
index 3b7ecd9..4c85512 100644
--- a/ql/src/test/queries/clientpositive/join_nulls.q
+++ b/ql/src/test/queries/clientpositive/join_nulls.q
@@ -1,47 +1,47 @@
 set hive.mapred.mode=nonstrict;
 -- SORT_QUERY_RESULTS
 
-CREATE TABLE myinput1(key int, value int);
-LOAD DATA LOCAL INPATH '../../data/files/in1.txt' INTO TABLE myinput1;
+CREATE TABLE myinput1_n3(key int, value int);
+LOAD DATA LOCAL INPATH '../../data/files/in1.txt' INTO TABLE myinput1_n3;
 
-SELECT * FROM myinput1 a JOIN myinput1 b;
-SELECT * FROM myinput1 a LEFT OUTER JOIN myinput1 b;
-SELECT * FROM myinput1 a RIGHT OUTER JOIN myinput1 b;
-SELECT * FROM myinput1 a JOIN myinput1 b ON a.key = b.value;
-SELECT * FROM myinput1 a JOIN myinput1 b ON a.key = b.key;
-SELECT * FROM myinput1 a JOIN myinput1 b ON a.value = b.value;
-SELECT * FROM myinput1 a JOIN myinput1 b ON a.value = b.value and a.key=b.key;
-SELECT * FROM myinput1 a LEFT OUTER JOIN myinput1 b ON a.key = b.value;
-SELECT * FROM myinput1 a LEFT OUTER JOIN myinput1 b ON a.value = b.value;
-SELECT * FROM myinput1 a LEFT OUTER JOIN myinput1 b ON a.key = b.key;
-SELECT * FROM myinput1 a LEFT OUTER JOIN myinput1 b ON a.key = b.key and a.value=b.value;
-SELECT * FROM myinput1 a RIGHT OUTER JOIN myinput1 b ON a.key = b.value;
-SELECT * FROM myinput1 a RIGHT OUTER JOIN myinput1 b ON a.key = b.key;
-SELECT * FROM myinput1 a RIGHT OUTER JOIN myinput1 b ON a.value = b.value;
-SELECT * FROM myinput1 a RIGHT OUTER JOIN myinput1 b ON a.key=b.key and a.value = b.value;
-SELECT * FROM myinput1 a FULL OUTER JOIN myinput1 b ON a.key = b.value;
-SELECT * FROM myinput1 a FULL OUTER JOIN myinput1 b ON a.key = b.key;
-SELECT * FROM myinput1 a FULL OUTER JOIN myinput1 b ON a.value = b.value;
-SELECT * FROM myinput1 a FULL OUTER JOIN myinput1 b ON a.value = b.value and a.key=b.key;
+SELECT * FROM myinput1_n3 a JOIN myinput1_n3 b;
+SELECT * FROM myinput1_n3 a LEFT OUTER JOIN myinput1_n3 b;
+SELECT * FROM myinput1_n3 a RIGHT OUTER JOIN myinput1_n3 b;
+SELECT * FROM myinput1_n3 a JOIN myinput1_n3 b ON a.key = b.value;
+SELECT * FROM myinput1_n3 a JOIN myinput1_n3 b ON a.key = b.key;
+SELECT * FROM myinput1_n3 a JOIN myinput1_n3 b ON a.value = b.value;
+SELECT * FROM myinput1_n3 a JOIN myinput1_n3 b ON a.value = b.value and a.key=b.key;
+SELECT * FROM myinput1_n3 a LEFT OUTER JOIN myinput1_n3 b ON a.key = b.value;
+SELECT * FROM myinput1_n3 a LEFT OUTER JOIN myinput1_n3 b ON a.value = b.value;
+SELECT * FROM myinput1_n3 a LEFT OUTER JOIN myinput1_n3 b ON a.key = b.key;
+SELECT * FROM myinput1_n3 a LEFT OUTER JOIN myinput1_n3 b ON a.key = b.key and a.value=b.value;
+SELECT * FROM myinput1_n3 a RIGHT OUTER JOIN myinput1_n3 b ON a.key = b.value;
+SELECT * FROM myinput1_n3 a RIGHT OUTER JOIN myinput1_n3 b ON a.key = b.key;
+SELECT * FROM myinput1_n3 a RIGHT OUTER JOIN myinput1_n3 b ON a.value = b.value;
+SELECT * FROM myinput1_n3 a RIGHT OUTER JOIN myinput1_n3 b ON a.key=b.key and a.value = b.value;
+SELECT * FROM myinput1_n3 a FULL OUTER JOIN myinput1_n3 b ON a.key = b.value;
+SELECT * FROM myinput1_n3 a FULL OUTER JOIN myinput1_n3 b ON a.key = b.key;
+SELECT * FROM myinput1_n3 a FULL OUTER JOIN myinput1_n3 b ON a.value = b.value;
+SELECT * FROM myinput1_n3 a FULL OUTER JOIN myinput1_n3 b ON a.value = b.value and a.key=b.key;
 
-SELECT * from myinput1 a LEFT OUTER JOIN myinput1 b ON (a.value=b.value) RIGHT OUTER JOIN myinput1 c ON (b.value=c.value);
-SELECT * from myinput1 a RIGHT OUTER JOIN myinput1 b ON (a.value=b.value) LEFT OUTER JOIN myinput1 c ON (b.value=c.value);
-SELECT * FROM myinput1 a LEFT OUTER JOIN myinput1 b RIGHT OUTER JOIN myinput1 c ON a.value = b.value and b.value = c.value;
+SELECT * from myinput1_n3 a LEFT OUTER JOIN myinput1_n3 b ON (a.value=b.value) RIGHT OUTER JOIN myinput1_n3 c ON (b.value=c.value);
+SELECT * from myinput1_n3 a RIGHT OUTER JOIN myinput1_n3 b ON (a.value=b.value) LEFT OUTER JOIN myinput1_n3 c ON (b.value=c.value);
+SELECT * FROM myinput1_n3 a LEFT OUTER JOIN myinput1_n3 b RIGHT OUTER JOIN myinput1_n3 c ON a.value = b.value and b.value = c.value;
 
-SELECT /*+ MAPJOIN(a) */ * FROM myinput1 a JOIN myinput1 b;
-SELECT /*+ MAPJOIN(a) */ * FROM myinput1 a JOIN myinput1 b ON a.key = b.value;
-SELECT /*+ MAPJOIN(a) */ * FROM myinput1 a JOIN myinput1 b ON a.key = b.key;
-SELECT /*+ MAPJOIN(a) */ * FROM myinput1 a JOIN myinput1 b ON a.value = b.value;
-SELECT /*+ MAPJOIN(b) */ * FROM myinput1 a JOIN myinput1 b ON a.key = b.value;
-SELECT /*+ MAPJOIN(b) */ * FROM myinput1 a JOIN myinput1 b ON a.key = b.key;
-SELECT /*+ MAPJOIN(b) */ * FROM myinput1 a JOIN myinput1 b ON a.value = b.value;
-SELECT /*+ MAPJOIN(b) */ * FROM myinput1 a JOIN myinput1 b ON a.value = b.value and a.key = b.key;
-SELECT /*+ MAPJOIN(b) */ * FROM myinput1 a LEFT OUTER JOIN myinput1 b ON a.key = b.value;
-SELECT /*+ MAPJOIN(b) */ * FROM myinput1 a LEFT OUTER JOIN myinput1 b ON a.key = b.key;
-SELECT /*+ MAPJOIN(b) */ * FROM myinput1 a LEFT OUTER JOIN myinput1 b ON a.value = b.value;
-SELECT /*+ MAPJOIN(a) */ * FROM myinput1 a RIGHT OUTER JOIN myinput1 b ON a.key = b.value;
-SELECT /*+ MAPJOIN(a) */ * FROM myinput1 a RIGHT OUTER JOIN myinput1 b ON a.key = b.key;
-SELECT /*+ MAPJOIN(a) */ * FROM myinput1 a RIGHT OUTER JOIN myinput1 b ON a.value = b.value;
+SELECT /*+ MAPJOIN(a) */ * FROM myinput1_n3 a JOIN myinput1_n3 b;
+SELECT /*+ MAPJOIN(a) */ * FROM myinput1_n3 a JOIN myinput1_n3 b ON a.key = b.value;
+SELECT /*+ MAPJOIN(a) */ * FROM myinput1_n3 a JOIN myinput1_n3 b ON a.key = b.key;
+SELECT /*+ MAPJOIN(a) */ * FROM myinput1_n3 a JOIN myinput1_n3 b ON a.value = b.value;
+SELECT /*+ MAPJOIN(b) */ * FROM myinput1_n3 a JOIN myinput1_n3 b ON a.key = b.value;
+SELECT /*+ MAPJOIN(b) */ * FROM myinput1_n3 a JOIN myinput1_n3 b ON a.key = b.key;
+SELECT /*+ MAPJOIN(b) */ * FROM myinput1_n3 a JOIN myinput1_n3 b ON a.value = b.value;
+SELECT /*+ MAPJOIN(b) */ * FROM myinput1_n3 a JOIN myinput1_n3 b ON a.value = b.value and a.key = b.key;
+SELECT /*+ MAPJOIN(b) */ * FROM myinput1_n3 a LEFT OUTER JOIN myinput1_n3 b ON a.key = b.value;
+SELECT /*+ MAPJOIN(b) */ * FROM myinput1_n3 a LEFT OUTER JOIN myinput1_n3 b ON a.key = b.key;
+SELECT /*+ MAPJOIN(b) */ * FROM myinput1_n3 a LEFT OUTER JOIN myinput1_n3 b ON a.value = b.value;
+SELECT /*+ MAPJOIN(a) */ * FROM myinput1_n3 a RIGHT OUTER JOIN myinput1_n3 b ON a.key = b.value;
+SELECT /*+ MAPJOIN(a) */ * FROM myinput1_n3 a RIGHT OUTER JOIN myinput1_n3 b ON a.key = b.key;
+SELECT /*+ MAPJOIN(a) */ * FROM myinput1_n3 a RIGHT OUTER JOIN myinput1_n3 b ON a.value = b.value;
 
 CREATE TABLE smb_input1(key int, value int) CLUSTERED BY (key) SORTED BY (key) INTO 2 BUCKETS; 
 CREATE TABLE smb_input2(key int, value int) CLUSTERED BY (value) SORTED BY (value) INTO 2 BUCKETS;

http://git-wip-us.apache.org/repos/asf/hive/blob/9bf28a3c/ql/src/test/queries/clientpositive/join_nullsafe.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/join_nullsafe.q b/ql/src/test/queries/clientpositive/join_nullsafe.q
index e96cc71..0cf0056 100644
--- a/ql/src/test/queries/clientpositive/join_nullsafe.q
+++ b/ql/src/test/queries/clientpositive/join_nullsafe.q
@@ -1,71 +1,71 @@
 set hive.explain.user=false;
 -- SORT_QUERY_RESULTS
 
-CREATE TABLE myinput1(key int, value int);
-LOAD DATA LOCAL INPATH '../../data/files/in8.txt' INTO TABLE myinput1;
+CREATE TABLE myinput1_n9(key int, value int);
+LOAD DATA LOCAL INPATH '../../data/files/in8.txt' INTO TABLE myinput1_n9;
 
 -- merging
-explain select * from myinput1 a join myinput1 b on a.key<=>b.value;
+explain select * from myinput1_n9 a join myinput1_n9 b on a.key<=>b.value;
 -- SORT_QUERY_RESULTS
-select * from myinput1 a join myinput1 b on a.key<=>b.value;
+select * from myinput1_n9 a join myinput1_n9 b on a.key<=>b.value;
 
-explain select * from myinput1 a join myinput1 b on a.key<=>b.value join myinput1 c on a.key=c.key;
-select * from myinput1 a join myinput1 b on a.key<=>b.value join myinput1 c on a.key=c.key;
+explain select * from myinput1_n9 a join myinput1_n9 b on a.key<=>b.value join myinput1_n9 c on a.key=c.key;
+select * from myinput1_n9 a join myinput1_n9 b on a.key<=>b.value join myinput1_n9 c on a.key=c.key;
 
-explain select * from myinput1 a join myinput1 b on a.key<=>b.value join myinput1 c on a.key<=>c.key;
-select * from myinput1 a join myinput1 b on a.key<=>b.value join myinput1 c on a.key<=>c.key;
+explain select * from myinput1_n9 a join myinput1_n9 b on a.key<=>b.value join myinput1_n9 c on a.key<=>c.key;
+select * from myinput1_n9 a join myinput1_n9 b on a.key<=>b.value join myinput1_n9 c on a.key<=>c.key;
 
-explain select * from myinput1 a join myinput1 b on a.key<=>b.value AND a.value=b.key join myinput1 c on a.key<=>c.key AND a.value=c.value;
+explain select * from myinput1_n9 a join myinput1_n9 b on a.key<=>b.value AND a.value=b.key join myinput1_n9 c on a.key<=>c.key AND a.value=c.value;
 
-select * from myinput1 a join myinput1 b on a.key<=>b.value AND a.value=b.key join myinput1 c on a.key<=>c.key AND a.value=c.value;
+select * from myinput1_n9 a join myinput1_n9 b on a.key<=>b.value AND a.value=b.key join myinput1_n9 c on a.key<=>c.key AND a.value=c.value;
 
-explain select * from myinput1 a join myinput1 b on a.key<=>b.value AND a.value<=>b.key join myinput1 c on a.key<=>c.key AND a.value<=>c.value;
-select * from myinput1 a join myinput1 b on a.key<=>b.value AND a.value<=>b.key join myinput1 c on a.key<=>c.key AND a.value<=>c.value;
+explain select * from myinput1_n9 a join myinput1_n9 b on a.key<=>b.value AND a.value<=>b.key join myinput1_n9 c on a.key<=>c.key AND a.value<=>c.value;
+select * from myinput1_n9 a join myinput1_n9 b on a.key<=>b.value AND a.value<=>b.key join myinput1_n9 c on a.key<=>c.key AND a.value<=>c.value;
 
 -- outer joins
-SELECT * FROM myinput1 a LEFT OUTER JOIN myinput1 b ON a.key<=>b.value;
-SELECT * FROM myinput1 a RIGHT OUTER JOIN myinput1 b ON a.key<=>b.value;
-SELECT * FROM myinput1 a FULL OUTER JOIN myinput1 b ON a.key<=>b.value;
+SELECT * FROM myinput1_n9 a LEFT OUTER JOIN myinput1_n9 b ON a.key<=>b.value;
+SELECT * FROM myinput1_n9 a RIGHT OUTER JOIN myinput1_n9 b ON a.key<=>b.value;
+SELECT * FROM myinput1_n9 a FULL OUTER JOIN myinput1_n9 b ON a.key<=>b.value;
 
 -- map joins
-SELECT /*+ MAPJOIN(a) */ * FROM myinput1 a JOIN myinput1 b ON a.key<=>b.value;
-SELECT /*+ MAPJOIN(b) */ * FROM myinput1 a JOIN myinput1 b ON a.key<=>b.value;
+SELECT /*+ MAPJOIN(a) */ * FROM myinput1_n9 a JOIN myinput1_n9 b ON a.key<=>b.value;
+SELECT /*+ MAPJOIN(b) */ * FROM myinput1_n9 a JOIN myinput1_n9 b ON a.key<=>b.value;
 
-CREATE TABLE smb_input(key int, value int);
-LOAD DATA LOCAL INPATH '../../data/files/in4.txt' into table smb_input;
-LOAD DATA LOCAL INPATH '../../data/files/in5.txt' into table smb_input;
+CREATE TABLE smb_input_n1(key int, value int);
+LOAD DATA LOCAL INPATH '../../data/files/in4.txt' into table smb_input_n1;
+LOAD DATA LOCAL INPATH '../../data/files/in5.txt' into table smb_input_n1;
 
 
 ;
 
 -- smbs
-CREATE TABLE smb_input1(key int, value int) CLUSTERED BY (key) SORTED BY (key) INTO 2 BUCKETS;
-CREATE TABLE smb_input2(key int, value int) CLUSTERED BY (value) SORTED BY (value) INTO 2 BUCKETS;
+CREATE TABLE smb_input1_n4(key int, value int) CLUSTERED BY (key) SORTED BY (key) INTO 2 BUCKETS;
+CREATE TABLE smb_input2_n4(key int, value int) CLUSTERED BY (value) SORTED BY (value) INTO 2 BUCKETS;
 
-from smb_input
-insert overwrite table smb_input1 select *
-insert overwrite table smb_input2 select *;
+from smb_input_n1
+insert overwrite table smb_input1_n4 select *
+insert overwrite table smb_input2_n4 select *;
 
 SET hive.optimize.bucketmapjoin = true;
 SET hive.optimize.bucketmapjoin.sortedmerge = true;
 SET hive.input.format = org.apache.hadoop.hive.ql.io.BucketizedHiveInputFormat;
 
-SELECT /*+ MAPJOIN(a) */ * FROM smb_input1 a JOIN smb_input1 b ON a.key <=> b.key;
-SELECT /*+ MAPJOIN(a) */ * FROM smb_input1 a JOIN smb_input1 b ON a.key <=> b.key AND a.value <=> b.value;
-SELECT /*+ MAPJOIN(a) */ * FROM smb_input1 a RIGHT OUTER JOIN smb_input1 b ON a.key <=> b.key;
-SELECT /*+ MAPJOIN(b) */ * FROM smb_input1 a JOIN smb_input1 b ON a.key <=> b.key;
-SELECT /*+ MAPJOIN(b) */ * FROM smb_input1 a LEFT OUTER JOIN smb_input1 b ON a.key <=> b.key;
+SELECT /*+ MAPJOIN(a) */ * FROM smb_input1_n4 a JOIN smb_input1_n4 b ON a.key <=> b.key;
+SELECT /*+ MAPJOIN(a) */ * FROM smb_input1_n4 a JOIN smb_input1_n4 b ON a.key <=> b.key AND a.value <=> b.value;
+SELECT /*+ MAPJOIN(a) */ * FROM smb_input1_n4 a RIGHT OUTER JOIN smb_input1_n4 b ON a.key <=> b.key;
+SELECT /*+ MAPJOIN(b) */ * FROM smb_input1_n4 a JOIN smb_input1_n4 b ON a.key <=> b.key;
+SELECT /*+ MAPJOIN(b) */ * FROM smb_input1_n4 a LEFT OUTER JOIN smb_input1_n4 b ON a.key <=> b.key;
 
-SELECT /*+ MAPJOIN(a) */ * FROM smb_input1 a JOIN smb_input2 b ON a.key <=> b.value;
-SELECT /*+ MAPJOIN(b) */ * FROM smb_input1 a JOIN smb_input2 b ON a.key <=> b.value;
-SELECT /*+ MAPJOIN(b) */ * FROM smb_input1 a LEFT OUTER JOIN smb_input2 b ON a.key <=> b.value;
-SELECT /*+ MAPJOIN(a) */ * FROM smb_input1 a RIGHT OUTER JOIN smb_input2 b ON a.key <=> b.value;
+SELECT /*+ MAPJOIN(a) */ * FROM smb_input1_n4 a JOIN smb_input2_n4 b ON a.key <=> b.value;
+SELECT /*+ MAPJOIN(b) */ * FROM smb_input1_n4 a JOIN smb_input2_n4 b ON a.key <=> b.value;
+SELECT /*+ MAPJOIN(b) */ * FROM smb_input1_n4 a LEFT OUTER JOIN smb_input2_n4 b ON a.key <=> b.value;
+SELECT /*+ MAPJOIN(a) */ * FROM smb_input1_n4 a RIGHT OUTER JOIN smb_input2_n4 b ON a.key <=> b.value;
 
-SELECT /*+ MAPJOIN(a) */ * FROM smb_input2 a JOIN smb_input2 b ON a.value <=> b.value;
-SELECT /*+ MAPJOIN(a) */ * FROM smb_input2 a RIGHT OUTER JOIN smb_input2 b ON a.value <=> b.value;
-SELECT /*+ MAPJOIN(b) */ * FROM smb_input2 a JOIN smb_input2 b ON a.value <=> b.value;
-SELECT /*+ MAPJOIN(b) */ * FROM smb_input2 a LEFT OUTER JOIN smb_input2 b ON a.value <=> b.value;
+SELECT /*+ MAPJOIN(a) */ * FROM smb_input2_n4 a JOIN smb_input2_n4 b ON a.value <=> b.value;
+SELECT /*+ MAPJOIN(a) */ * FROM smb_input2_n4 a RIGHT OUTER JOIN smb_input2_n4 b ON a.value <=> b.value;
+SELECT /*+ MAPJOIN(b) */ * FROM smb_input2_n4 a JOIN smb_input2_n4 b ON a.value <=> b.value;
+SELECT /*+ MAPJOIN(b) */ * FROM smb_input2_n4 a LEFT OUTER JOIN smb_input2_n4 b ON a.value <=> b.value;
 
 --HIVE-3315 join predicate transitive
-explain select * from myinput1 a join myinput1 b on a.key<=>b.value AND a.key is NULL;
-select * from myinput1 a join myinput1 b on a.key<=>b.value AND a.key is NULL;
+explain select * from myinput1_n9 a join myinput1_n9 b on a.key<=>b.value AND a.key is NULL;
+select * from myinput1_n9 a join myinput1_n9 b on a.key<=>b.value AND a.key is NULL;

http://git-wip-us.apache.org/repos/asf/hive/blob/9bf28a3c/ql/src/test/queries/clientpositive/join_on_varchar.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/join_on_varchar.q b/ql/src/test/queries/clientpositive/join_on_varchar.q
index 5016369..faabdb9 100644
--- a/ql/src/test/queries/clientpositive/join_on_varchar.q
+++ b/ql/src/test/queries/clientpositive/join_on_varchar.q
@@ -1,14 +1,14 @@
 set hive.mapred.mode=nonstrict;
 -- SORT_QUERY_RESULTS
 
-create table tbl1(c1 varchar(10), intcol int);
-create table tbl2(c2 varchar(30));
-insert into table tbl1 select repeat('t', 10), 11 from src limit 1;
-insert into table tbl1 select repeat('s', 10), 22 from src limit 1;
-insert into table tbl2 select concat(repeat('t', 10), 'ppp') from src limit 1;
-insert into table tbl2 select repeat('s', 10) from src limit 1;
+create table tbl1_n3(c1 varchar(10), intcol int);
+create table tbl2_n2(c2 varchar(30));
+insert into table tbl1_n3 select repeat('t', 10), 11 from src limit 1;
+insert into table tbl1_n3 select repeat('s', 10), 22 from src limit 1;
+insert into table tbl2_n2 select concat(repeat('t', 10), 'ppp') from src limit 1;
+insert into table tbl2_n2 select repeat('s', 10) from src limit 1;
 set hive.auto.convert.join=true;
 
 explain
-select /*+ MAPJOIN(tbl2) */ c1,c2 from tbl1 join tbl2 on (c1 = c2) order by c1,c2;
-select /*+ MAPJOIN(tbl2) */ c1,c2 from tbl1 join tbl2 on (c1 = c2) order by c1,c2;
+select /*+ MAPJOIN(tbl2_n2) */ c1,c2 from tbl1_n3 join tbl2_n2 on (c1 = c2) order by c1,c2;
+select /*+ MAPJOIN(tbl2_n2) */ c1,c2 from tbl1_n3 join tbl2_n2 on (c1 = c2) order by c1,c2;

http://git-wip-us.apache.org/repos/asf/hive/blob/9bf28a3c/ql/src/test/queries/clientpositive/join_reorder.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/join_reorder.q b/ql/src/test/queries/clientpositive/join_reorder.q
index 5161f1b..8df94f7 100644
--- a/ql/src/test/queries/clientpositive/join_reorder.q
+++ b/ql/src/test/queries/clientpositive/join_reorder.q
@@ -4,70 +4,70 @@ set hive.cbo.enable=false;
 
 
 
-CREATE TABLE T1(key STRING, val STRING) STORED AS TEXTFILE;
-CREATE TABLE T2(key STRING, val STRING) STORED AS TEXTFILE;
-CREATE TABLE T3(key STRING, val STRING) STORED AS TEXTFILE;
+CREATE TABLE T1_n37(key STRING, val STRING) STORED AS TEXTFILE;
+CREATE TABLE T2_n24(key STRING, val STRING) STORED AS TEXTFILE;
+CREATE TABLE T3_n8(key STRING, val STRING) STORED AS TEXTFILE;
 
-LOAD DATA LOCAL INPATH '../../data/files/T1.txt' INTO TABLE T1;
-LOAD DATA LOCAL INPATH '../../data/files/T2.txt' INTO TABLE T2;
-LOAD DATA LOCAL INPATH '../../data/files/T3.txt' INTO TABLE T3;
+LOAD DATA LOCAL INPATH '../../data/files/T1.txt' INTO TABLE T1_n37;
+LOAD DATA LOCAL INPATH '../../data/files/T2.txt' INTO TABLE T2_n24;
+LOAD DATA LOCAL INPATH '../../data/files/T3.txt' INTO TABLE T3_n8;
 
 -- SORT_QUERY_RESULTS
 
-EXPLAIN FROM T1 a JOIN src c ON c.key+1=a.key
+EXPLAIN FROM T1_n37 a JOIN src c ON c.key+1=a.key
 SELECT a.key, a.val, c.key;
 
-EXPLAIN FROM T1 a JOIN src c ON c.key+1=a.key
+EXPLAIN FROM T1_n37 a JOIN src c ON c.key+1=a.key
 SELECT /*+ STREAMTABLE(a) */ a.key, a.val, c.key;
 
-FROM T1 a JOIN src c ON c.key+1=a.key
+FROM T1_n37 a JOIN src c ON c.key+1=a.key
 SELECT a.key, a.val, c.key;
 
-FROM T1 a JOIN src c ON c.key+1=a.key
+FROM T1_n37 a JOIN src c ON c.key+1=a.key
 SELECT /*+ STREAMTABLE(a) */ a.key, a.val, c.key;
 
-EXPLAIN FROM T1 a
-  LEFT OUTER JOIN T2 b ON (b.key=a.key)
-  RIGHT OUTER JOIN T3 c ON (c.val = a.val)
+EXPLAIN FROM T1_n37 a
+  LEFT OUTER JOIN T2_n24 b ON (b.key=a.key)
+  RIGHT OUTER JOIN T3_n8 c ON (c.val = a.val)
 SELECT a.key, b.key, a.val, c.val;
 
-EXPLAIN FROM T1 a
-  LEFT OUTER JOIN T2 b ON (b.key=a.key)
-  RIGHT OUTER JOIN T3 c ON (c.val = a.val)
+EXPLAIN FROM T1_n37 a
+  LEFT OUTER JOIN T2_n24 b ON (b.key=a.key)
+  RIGHT OUTER JOIN T3_n8 c ON (c.val = a.val)
 SELECT /*+ STREAMTABLE(a) */ a.key, b.key, a.val, c.val;
 
-FROM T1 a
-  LEFT OUTER JOIN T2 b ON (b.key=a.key)
-  RIGHT OUTER JOIN T3 c ON (c.val = a.val)
+FROM T1_n37 a
+  LEFT OUTER JOIN T2_n24 b ON (b.key=a.key)
+  RIGHT OUTER JOIN T3_n8 c ON (c.val = a.val)
 SELECT a.key, b.key, a.val, c.val;
 
-FROM T1 a
-  LEFT OUTER JOIN T2 b ON (b.key=a.key)
-  RIGHT OUTER JOIN T3 c ON (c.val = a.val)
+FROM T1_n37 a
+  LEFT OUTER JOIN T2_n24 b ON (b.key=a.key)
+  RIGHT OUTER JOIN T3_n8 c ON (c.val = a.val)
 SELECT /*+ STREAMTABLE(a) */ a.key, b.key, a.val, c.val;
 
 EXPLAIN FROM UNIQUEJOIN
-  PRESERVE T1 a (a.key, a.val),
-  PRESERVE T2 b (b.key, b.val),
-  PRESERVE T3 c (c.key, c.val)
+  PRESERVE T1_n37 a (a.key, a.val),
+  PRESERVE T2_n24 b (b.key, b.val),
+  PRESERVE T3_n8 c (c.key, c.val)
 SELECT a.key, b.key, c.key;
 
 EXPLAIN FROM UNIQUEJOIN
-  PRESERVE T1 a (a.key, a.val),
-  PRESERVE T2 b (b.key, b.val),
-  PRESERVE T3 c (c.key, c.val)
+  PRESERVE T1_n37 a (a.key, a.val),
+  PRESERVE T2_n24 b (b.key, b.val),
+  PRESERVE T3_n8 c (c.key, c.val)
 SELECT /*+ STREAMTABLE(b) */ a.key, b.key, c.key;
 
 FROM UNIQUEJOIN
-  PRESERVE T1 a (a.key, a.val),
-  PRESERVE T2 b (b.key, b.val),
-  PRESERVE T3 c (c.key, c.val)
+  PRESERVE T1_n37 a (a.key, a.val),
+  PRESERVE T2_n24 b (b.key, b.val),
+  PRESERVE T3_n8 c (c.key, c.val)
 SELECT a.key, b.key, c.key;
 
 FROM UNIQUEJOIN
-  PRESERVE T1 a (a.key, a.val),
-  PRESERVE T2 b (b.key, b.val),
-  PRESERVE T3 c (c.key, c.val)
+  PRESERVE T1_n37 a (a.key, a.val),
+  PRESERVE T2_n24 b (b.key, b.val),
+  PRESERVE T3_n8 c (c.key, c.val)
 SELECT /*+ STREAMTABLE(b) */ a.key, b.key, c.key;
 
 

http://git-wip-us.apache.org/repos/asf/hive/blob/9bf28a3c/ql/src/test/queries/clientpositive/join_reorder2.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/join_reorder2.q b/ql/src/test/queries/clientpositive/join_reorder2.q
index 93a0dba..924b2ee 100644
--- a/ql/src/test/queries/clientpositive/join_reorder2.q
+++ b/ql/src/test/queries/clientpositive/join_reorder2.q
@@ -4,39 +4,39 @@ set hive.cbo.enable=false;
 
 
 
-CREATE TABLE T1(key STRING, val STRING) STORED AS TEXTFILE;
-CREATE TABLE T2(key STRING, val STRING) STORED AS TEXTFILE;
-CREATE TABLE T3(key STRING, val STRING) STORED AS TEXTFILE;
-CREATE TABLE T4(key STRING, val STRING) STORED AS TEXTFILE;
+CREATE TABLE T1_n49(key STRING, val STRING) STORED AS TEXTFILE;
+CREATE TABLE T2_n30(key STRING, val STRING) STORED AS TEXTFILE;
+CREATE TABLE T3_n10(key STRING, val STRING) STORED AS TEXTFILE;
+CREATE TABLE T4_n3(key STRING, val STRING) STORED AS TEXTFILE;
 
-LOAD DATA LOCAL INPATH '../../data/files/T1.txt' INTO TABLE T1;
-LOAD DATA LOCAL INPATH '../../data/files/T2.txt' INTO TABLE T2;
-LOAD DATA LOCAL INPATH '../../data/files/T3.txt' INTO TABLE T3;
-LOAD DATA LOCAL INPATH '../../data/files/T1.txt' INTO TABLE T4;
+LOAD DATA LOCAL INPATH '../../data/files/T1.txt' INTO TABLE T1_n49;
+LOAD DATA LOCAL INPATH '../../data/files/T2.txt' INTO TABLE T2_n30;
+LOAD DATA LOCAL INPATH '../../data/files/T3.txt' INTO TABLE T3_n10;
+LOAD DATA LOCAL INPATH '../../data/files/T1.txt' INTO TABLE T4_n3;
 
 EXPLAIN
 SELECT /*+ STREAMTABLE(a) */ *
-FROM T1 a JOIN T2 b ON a.key = b.key
-          JOIN T3 c ON b.key = c.key
-          JOIN T4 d ON c.key = d.key;
+FROM T1_n49 a JOIN T2_n30 b ON a.key = b.key
+          JOIN T3_n10 c ON b.key = c.key
+          JOIN T4_n3 d ON c.key = d.key;
 
 SELECT /*+ STREAMTABLE(a) */ *
-FROM T1 a JOIN T2 b ON a.key = b.key
-          JOIN T3 c ON b.key = c.key
-          JOIN T4 d ON c.key = d.key;
+FROM T1_n49 a JOIN T2_n30 b ON a.key = b.key
+          JOIN T3_n10 c ON b.key = c.key
+          JOIN T4_n3 d ON c.key = d.key;
 
 
 EXPLAIN
 SELECT /*+ STREAMTABLE(a) */ *
-FROM T1 a JOIN T2 b ON a.key = b.key
-          JOIN T3 c ON a.val = c.val
-          JOIN T4 d ON a.key + 1 = d.key + 1;
+FROM T1_n49 a JOIN T2_n30 b ON a.key = b.key
+          JOIN T3_n10 c ON a.val = c.val
+          JOIN T4_n3 d ON a.key + 1 = d.key + 1;
 
 
 SELECT /*+ STREAMTABLE(a) */ *
-FROM T1 a JOIN T2 b ON a.key = b.key
-          JOIN T3 c ON a.val = c.val
-          JOIN T4 d ON a.key + 1 = d.key + 1;
+FROM T1_n49 a JOIN T2_n30 b ON a.key = b.key
+          JOIN T3_n10 c ON a.val = c.val
+          JOIN T4_n3 d ON a.key + 1 = d.key + 1;
 
 
 

http://git-wip-us.apache.org/repos/asf/hive/blob/9bf28a3c/ql/src/test/queries/clientpositive/join_reorder3.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/join_reorder3.q b/ql/src/test/queries/clientpositive/join_reorder3.q
index 7d5a95d..1ebf242 100644
--- a/ql/src/test/queries/clientpositive/join_reorder3.q
+++ b/ql/src/test/queries/clientpositive/join_reorder3.q
@@ -3,39 +3,39 @@ set hive.cbo.enable=false;
 
 
 
-CREATE TABLE T1(key STRING, val STRING) STORED AS TEXTFILE;
-CREATE TABLE T2(key STRING, val STRING) STORED AS TEXTFILE;
-CREATE TABLE T3(key STRING, val STRING) STORED AS TEXTFILE;
-CREATE TABLE T4(key STRING, val STRING) STORED AS TEXTFILE;
+CREATE TABLE T1_n92(key STRING, val STRING) STORED AS TEXTFILE;
+CREATE TABLE T2_n57(key STRING, val STRING) STORED AS TEXTFILE;
+CREATE TABLE T3_n21(key STRING, val STRING) STORED AS TEXTFILE;
+CREATE TABLE T4_n10(key STRING, val STRING) STORED AS TEXTFILE;
 
-LOAD DATA LOCAL INPATH '../../data/files/T1.txt' INTO TABLE T1;
-LOAD DATA LOCAL INPATH '../../data/files/T2.txt' INTO TABLE T2;
-LOAD DATA LOCAL INPATH '../../data/files/T3.txt' INTO TABLE T3;
-LOAD DATA LOCAL INPATH '../../data/files/T1.txt' INTO TABLE T4;
+LOAD DATA LOCAL INPATH '../../data/files/T1.txt' INTO TABLE T1_n92;
+LOAD DATA LOCAL INPATH '../../data/files/T2.txt' INTO TABLE T2_n57;
+LOAD DATA LOCAL INPATH '../../data/files/T3.txt' INTO TABLE T3_n21;
+LOAD DATA LOCAL INPATH '../../data/files/T1.txt' INTO TABLE T4_n10;
 
 EXPLAIN
 SELECT /*+ STREAMTABLE(a,c) */ *
-FROM T1 a JOIN T2 b ON a.key = b.key
-          JOIN T3 c ON b.key = c.key
-          JOIN T4 d ON c.key = d.key;
+FROM T1_n92 a JOIN T2_n57 b ON a.key = b.key
+          JOIN T3_n21 c ON b.key = c.key
+          JOIN T4_n10 d ON c.key = d.key;
 
 SELECT /*+ STREAMTABLE(a,c) */ *
-FROM T1 a JOIN T2 b ON a.key = b.key
-          JOIN T3 c ON b.key = c.key
-          JOIN T4 d ON c.key = d.key;
+FROM T1_n92 a JOIN T2_n57 b ON a.key = b.key
+          JOIN T3_n21 c ON b.key = c.key
+          JOIN T4_n10 d ON c.key = d.key;
 
 
 EXPLAIN
 SELECT /*+ STREAMTABLE(a,c) */ *
-FROM T1 a JOIN T2 b ON a.key = b.key
-          JOIN T3 c ON a.val = c.val
-          JOIN T4 d ON a.key + 1 = d.key + 1;
+FROM T1_n92 a JOIN T2_n57 b ON a.key = b.key
+          JOIN T3_n21 c ON a.val = c.val
+          JOIN T4_n10 d ON a.key + 1 = d.key + 1;
 
 
 SELECT /*+ STREAMTABLE(a,c) */ *
-FROM T1 a JOIN T2 b ON a.key = b.key
-          JOIN T3 c ON a.val = c.val
-          JOIN T4 d ON a.key + 1 = d.key + 1;
+FROM T1_n92 a JOIN T2_n57 b ON a.key = b.key
+          JOIN T3_n21 c ON a.val = c.val
+          JOIN T4_n10 d ON a.key + 1 = d.key + 1;
 
 
 

http://git-wip-us.apache.org/repos/asf/hive/blob/9bf28a3c/ql/src/test/queries/clientpositive/join_reorder4.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/join_reorder4.q b/ql/src/test/queries/clientpositive/join_reorder4.q
index 265980b..16ed71c 100644
--- a/ql/src/test/queries/clientpositive/join_reorder4.q
+++ b/ql/src/test/queries/clientpositive/join_reorder4.q
@@ -1,18 +1,18 @@
-CREATE TABLE T1(key1 STRING, val1 STRING) STORED AS TEXTFILE;
-CREATE TABLE T2(key2 STRING, val2 STRING) STORED AS TEXTFILE;
-CREATE TABLE T3(key3 STRING, val3 STRING) STORED AS TEXTFILE;
+CREATE TABLE T1_n134(key1 STRING, val1 STRING) STORED AS TEXTFILE;
+CREATE TABLE T2_n80(key2 STRING, val2 STRING) STORED AS TEXTFILE;
+CREATE TABLE T3_n32(key3 STRING, val3 STRING) STORED AS TEXTFILE;
 
-LOAD DATA LOCAL INPATH '../../data/files/T1.txt' INTO TABLE T1;
-LOAD DATA LOCAL INPATH '../../data/files/T2.txt' INTO TABLE T2;
-LOAD DATA LOCAL INPATH '../../data/files/T3.txt' INTO TABLE T3;
+LOAD DATA LOCAL INPATH '../../data/files/T1.txt' INTO TABLE T1_n134;
+LOAD DATA LOCAL INPATH '../../data/files/T2.txt' INTO TABLE T2_n80;
+LOAD DATA LOCAL INPATH '../../data/files/T3.txt' INTO TABLE T3_n32;
 
 set hive.cbo.enable=false;
 
-explain select /*+ STREAMTABLE(a) */ a.*, b.*, c.* from T1 a join T2 b on a.key1=b.key2 join T3 c on a.key1=c.key3;
-select /*+ STREAMTABLE(a) */ a.*, b.*, c.* from T1 a join T2 b on a.key1=b.key2 join T3 c on a.key1=c.key3;
+explain select /*+ STREAMTABLE(a) */ a.*, b.*, c.* from T1_n134 a join T2_n80 b on a.key1=b.key2 join T3_n32 c on a.key1=c.key3;
+select /*+ STREAMTABLE(a) */ a.*, b.*, c.* from T1_n134 a join T2_n80 b on a.key1=b.key2 join T3_n32 c on a.key1=c.key3;
 
-explain select /*+ STREAMTABLE(b) */ a.*, b.*, c.* from T1 a join T2 b on a.key1=b.key2 join T3 c on a.key1=c.key3;
-select /*+ STREAMTABLE(b) */ a.*, b.*, c.* from T1 a join T2 b on a.key1=b.key2 join T3 c on a.key1=c.key3;
+explain select /*+ STREAMTABLE(b) */ a.*, b.*, c.* from T1_n134 a join T2_n80 b on a.key1=b.key2 join T3_n32 c on a.key1=c.key3;
+select /*+ STREAMTABLE(b) */ a.*, b.*, c.* from T1_n134 a join T2_n80 b on a.key1=b.key2 join T3_n32 c on a.key1=c.key3;
 
-explain select /*+ STREAMTABLE(c) */ a.*, b.*, c.* from T1 a join T2 b on a.key1=b.key2 join T3 c on a.key1=c.key3;
-select /*+ STREAMTABLE(c) */ a.*, b.*, c.* from T1 a join T2 b on a.key1=b.key2 join T3 c on a.key1=c.key3;
+explain select /*+ STREAMTABLE(c) */ a.*, b.*, c.* from T1_n134 a join T2_n80 b on a.key1=b.key2 join T3_n32 c on a.key1=c.key3;
+select /*+ STREAMTABLE(c) */ a.*, b.*, c.* from T1_n134 a join T2_n80 b on a.key1=b.key2 join T3_n32 c on a.key1=c.key3;

http://git-wip-us.apache.org/repos/asf/hive/blob/9bf28a3c/ql/src/test/queries/clientpositive/join_reordering_no_stats.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/join_reordering_no_stats.q b/ql/src/test/queries/clientpositive/join_reordering_no_stats.q
index 3ea9f0c..e848c9f 100644
--- a/ql/src/test/queries/clientpositive/join_reordering_no_stats.q
+++ b/ql/src/test/queries/clientpositive/join_reordering_no_stats.q
@@ -40,24 +40,24 @@ explain select count(1) from part_nostats,supplier_nostats,lineitem_nostats wher
 set hive.stats.estimate=false;
 explain select count(1) from part_nostats,supplier_nostats,lineitem_nostats where p_partkey = l_partkey and s_suppkey = l_suppkey;
 
-CREATE TABLE Employee_Part(employeeID int, employeeName String) partitioned by (employeeSalary double, country string)
+CREATE TABLE Employee_Part_n1(employeeID int, employeeName String) partitioned by (employeeSalary double, country string)
 row format delimited fields terminated by '|'  stored as textfile;
 
-LOAD DATA LOCAL INPATH "../../data/files/employee.dat"  INTO TABLE Employee_Part partition(employeeSalary='2000.0', country='USA');
-LOAD DATA LOCAL INPATH "../../data/files/employee2.dat" INTO TABLE Employee_Part partition(employeeSalary='2000.0', country='UK');
-LOAD DATA LOCAL INPATH "../../data/files/employee2.dat" INTO TABLE Employee_Part partition(employeeSalary='3000.0', country='USA');
-LOAD DATA LOCAL INPATH "../../data/files/employee2.dat" INTO TABLE Employee_Part partition(employeeSalary='4000.0', country='USA');
-LOAD DATA LOCAL INPATH "../../data/files/employee2.dat" INTO TABLE Employee_Part partition(employeeSalary='3500.0', country='UK');
-LOAD DATA LOCAL INPATH "../../data/files/employee.dat"  INTO TABLE Employee_Part partition(employeeSalary='3000.0', country='UK');
+LOAD DATA LOCAL INPATH "../../data/files/employee.dat"  INTO TABLE Employee_Part_n1 partition(employeeSalary='2000.0', country='USA');
+LOAD DATA LOCAL INPATH "../../data/files/employee2.dat" INTO TABLE Employee_Part_n1 partition(employeeSalary='2000.0', country='UK');
+LOAD DATA LOCAL INPATH "../../data/files/employee2.dat" INTO TABLE Employee_Part_n1 partition(employeeSalary='3000.0', country='USA');
+LOAD DATA LOCAL INPATH "../../data/files/employee2.dat" INTO TABLE Employee_Part_n1 partition(employeeSalary='4000.0', country='USA');
+LOAD DATA LOCAL INPATH "../../data/files/employee2.dat" INTO TABLE Employee_Part_n1 partition(employeeSalary='3500.0', country='UK');
+LOAD DATA LOCAL INPATH "../../data/files/employee.dat"  INTO TABLE Employee_Part_n1 partition(employeeSalary='3000.0', country='UK');
 
 -- partitioned table
 set hive.stats.estimate=true;
-explain select count(1) from Employee_Part,supplier_nostats,lineitem_nostats where employeeID= l_partkey and s_suppkey = l_suppkey;
+explain select count(1) from Employee_Part_n1,supplier_nostats,lineitem_nostats where employeeID= l_partkey and s_suppkey = l_suppkey;
 
 set hive.stats.estimate=false;
-explain select count(1) from Employee_Part,supplier_nostats,lineitem_nostats where employeeID= l_partkey and s_suppkey = l_suppkey;
+explain select count(1) from Employee_Part_n1,supplier_nostats,lineitem_nostats where employeeID= l_partkey and s_suppkey = l_suppkey;
 
-drop table Employee_Part;
+drop table Employee_Part_n1;
 drop table supplier_nostats;
 drop table lineitem_nostats;
 drop table part_nostats;

http://git-wip-us.apache.org/repos/asf/hive/blob/9bf28a3c/ql/src/test/queries/clientpositive/json_serde_tsformat.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/json_serde_tsformat.q b/ql/src/test/queries/clientpositive/json_serde_tsformat.q
index c00450c..3cbd91d 100644
--- a/ql/src/test/queries/clientpositive/json_serde_tsformat.q
+++ b/ql/src/test/queries/clientpositive/json_serde_tsformat.q
@@ -1,12 +1,12 @@
 
 add jar ${system:maven.local.repository}/org/apache/hive/hcatalog/hive-hcatalog-core/${system:hive.version}/hive-hcatalog-core-${system:hive.version}.jar;
 
-CREATE TABLE t1 (c1 int, c2 string, c3 timestamp)
+CREATE TABLE t1_n156 (c1 int, c2 string, c3 timestamp)
 ROW FORMAT SERDE 'org.apache.hive.hcatalog.data.JsonSerDe'
 WITH SERDEPROPERTIES ('timestamp.formats'='yyyy-MM-dd\'T\'HH:mm:ss')
 ;
-LOAD DATA LOCAL INPATH "../../data/files/tsformat.json" INTO TABLE t1;
+LOAD DATA LOCAL INPATH "../../data/files/tsformat.json" INTO TABLE t1_n156;
 select a.c1, a.c2, b.c3
-from t1 a join t1 b on a.c1 = b.c1;
+from t1_n156 a join t1_n156 b on a.c1 = b.c1;
 
-drop table t1;
+drop table t1_n156;

http://git-wip-us.apache.org/repos/asf/hive/blob/9bf28a3c/ql/src/test/queries/clientpositive/keyword_2.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/keyword_2.q b/ql/src/test/queries/clientpositive/keyword_2.q
index e8be760..9a33dc9 100644
--- a/ql/src/test/queries/clientpositive/keyword_2.q
+++ b/ql/src/test/queries/clientpositive/keyword_2.q
@@ -1,13 +1,14 @@
-drop table varchar_udf_1;
+--! qt:dataset:src
+drop table varchar_udf_1_n1;
 
-create table varchar_udf_1 (c1 string, c2 string, c3 varchar(10), c4 varchar(20));
-insert overwrite table varchar_udf_1
+create table varchar_udf_1_n1 (c1 string, c2 string, c3 varchar(10), c4 varchar(20));
+insert overwrite table varchar_udf_1_n1
   select key, value, key, value from src where key = '238' limit 1;
 
 select
   c2 regexp 'val',
   c4 regexp 'val',
   (c2 regexp 'val') = (c4 regexp 'val')
-from varchar_udf_1 limit 1;
+from varchar_udf_1_n1 limit 1;
 
-drop table varchar_udf_1;
+drop table varchar_udf_1_n1;

http://git-wip-us.apache.org/repos/asf/hive/blob/9bf28a3c/ql/src/test/queries/clientpositive/lateral_view_multi_lateralviews.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/lateral_view_multi_lateralviews.q b/ql/src/test/queries/clientpositive/lateral_view_multi_lateralviews.q
index 29e026a..f5050e0 100644
--- a/ql/src/test/queries/clientpositive/lateral_view_multi_lateralviews.q
+++ b/ql/src/test/queries/clientpositive/lateral_view_multi_lateralviews.q
@@ -1,7 +1,7 @@
-CREATE TABLE t1(x5 STRUCT<x4: STRUCT<x3: ARRAY<STRUCT<x1: STRING, x2: ARRAY<STRING>>>> >);
-INSERT INTO t1 SELECT NAMED_STRUCT('x4', NAMED_STRUCT('x3', ARRAY(NAMED_STRUCT('x1', 'x1_1', 'x2', ARRAY('x2_1', 'x2_2')))));
+CREATE TABLE t1_n135(x5 STRUCT<x4: STRUCT<x3: ARRAY<STRUCT<x1: STRING, x2: ARRAY<STRING>>>> >);
+INSERT INTO t1_n135 SELECT NAMED_STRUCT('x4', NAMED_STRUCT('x3', ARRAY(NAMED_STRUCT('x1', 'x1_1', 'x2', ARRAY('x2_1', 'x2_2')))));
 SELECT c1,c2,c3,c4,c5,c6,c7,c8,c9,c10,c11,c12,c13,c14,c15,c16
-FROM t1
+FROM t1_n135
 LATERAL VIEW EXPLODE(x5.x4.x3) lv as c1
 LATERAL VIEW EXPLODE(c1.x2) lv as c2
 LATERAL VIEW EXPLODE(x5.x4.x3) lv as c3

http://git-wip-us.apache.org/repos/asf/hive/blob/9bf28a3c/ql/src/test/queries/clientpositive/lateral_view_onview.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/lateral_view_onview.q b/ql/src/test/queries/clientpositive/lateral_view_onview.q
index fa559f4..4d23634 100644
--- a/ql/src/test/queries/clientpositive/lateral_view_onview.q
+++ b/ql/src/test/queries/clientpositive/lateral_view_onview.q
@@ -1,7 +1,8 @@
-CREATE TABLE lv_table( c1 STRING,  c2 ARRAY<INT>, c3 INT, c4 CHAR(1));
-INSERT OVERWRITE TABLE lv_table SELECT 'abc  ', array(1,2,3), 100, 't' FROM src;
+--! qt:dataset:src
+CREATE TABLE lv_table_n0( c1 STRING,  c2 ARRAY<INT>, c3 INT, c4 CHAR(1));
+INSERT OVERWRITE TABLE lv_table_n0 SELECT 'abc  ', array(1,2,3), 100, 't' FROM src;
 
-CREATE OR REPLACE VIEW lv_view AS SELECT * FROM lv_table; 
+CREATE OR REPLACE VIEW lv_view AS SELECT * FROM lv_table_n0; 
 
 EXPLAIN SELECT * FROM lv_view LATERAL VIEW explode(array(1,2,3)) myTable AS myCol SORT BY c1 ASC, myCol ASC LIMIT 1;
 EXPLAIN SELECT myTable.* FROM lv_view LATERAL VIEW explode(array(1,2,3)) myTable AS myCol LIMIT 3;
@@ -22,9 +23,9 @@ SELECT SIZE(c2),c3,TRIM(c1),c4,myCol from lv_view LATERAL VIEW explode(array(1,2
 
 SELECT SIZE(c2),c3,TRIM(c1),c4,myCol from lv_view LATERAL VIEW explode(array(1,2,3)) myTab as myCol limit 3;
 
-CREATE TABLE lv_table1( c1 STRING,  c3 INT, c4 CHAR(1), c5 STRING, c6 STRING, c7 STRING, c8 STRING, c9 STRING, c10 STRING, c11 STRING, c12 STRING, c13 STRING);
+CREATE TABLE lv_table1_n0( c1 STRING,  c3 INT, c4 CHAR(1), c5 STRING, c6 STRING, c7 STRING, c8 STRING, c9 STRING, c10 STRING, c11 STRING, c12 STRING, c13 STRING);
 CREATE TABLE lv_table2( c1 STRING,  c2 ARRAY<INT>);
-INSERT OVERWRITE TABLE lv_table1 SELECT 'abc  ', 100, 't', 'test', 'test', 'test', 'test', 'test', 'test', 'test', 'test', 'test' FROM src;
+INSERT OVERWRITE TABLE lv_table1_n0 SELECT 'abc  ', 100, 't', 'test', 'test', 'test', 'test', 'test', 'test', 'test', 'test', 'test' FROM src;
 INSERT OVERWRITE TABLE lv_table2 SELECT 'abc  ', array(1,2,3) FROM src;
-EXPLAIN WITH lv_view1 AS (SELECT lv_table1.*, c2 FROM lv_table1 JOIN lv_table2 ON lv_table1.c1 = lv_table2.c1), lv_view2 AS (SELECT * FROM lv_view1 LATERAL VIEW explode(c2) myTable AS myCol) SELECT * FROM lv_view2 SORT BY c1 ASC, myCol ASC LIMIT 1;
-WITH lv_view1 AS (SELECT lv_table1.*, c2 FROM lv_table1 JOIN lv_table2 ON lv_table1.c1 = lv_table2.c1), lv_view2 AS (SELECT * FROM lv_view1 LATERAL VIEW explode(c2) myTable AS myCol) SELECT * FROM lv_view2 SORT BY c1 ASC, myCol ASC LIMIT 1;
\ No newline at end of file
+EXPLAIN WITH lv_view1 AS (SELECT lv_table1_n0.*, c2 FROM lv_table1_n0 JOIN lv_table2 ON lv_table1_n0.c1 = lv_table2.c1), lv_view2 AS (SELECT * FROM lv_view1 LATERAL VIEW explode(c2) myTable AS myCol) SELECT * FROM lv_view2 SORT BY c1 ASC, myCol ASC LIMIT 1;
+WITH lv_view1 AS (SELECT lv_table1_n0.*, c2 FROM lv_table1_n0 JOIN lv_table2 ON lv_table1_n0.c1 = lv_table2.c1), lv_view2 AS (SELECT * FROM lv_view1 LATERAL VIEW explode(c2) myTable AS myCol) SELECT * FROM lv_view2 SORT BY c1 ASC, myCol ASC LIMIT 1;
\ No newline at end of file

http://git-wip-us.apache.org/repos/asf/hive/blob/9bf28a3c/ql/src/test/queries/clientpositive/lateral_view_onview2.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/lateral_view_onview2.q b/ql/src/test/queries/clientpositive/lateral_view_onview2.q
index c13f754..46e2330 100644
--- a/ql/src/test/queries/clientpositive/lateral_view_onview2.q
+++ b/ql/src/test/queries/clientpositive/lateral_view_onview2.q
@@ -1,6 +1,7 @@
-CREATE TABLE lv_table( c1 STRING,  c2 ARRAY<INT>, c3 INT, c4 CHAR(1));
-INSERT OVERWRITE TABLE lv_table SELECT 'abc  ', array(1,2,3), 100, 't' FROM src;
+--! qt:dataset:src
+CREATE TABLE lv_table_n1( c1 STRING,  c2 ARRAY<INT>, c3 INT, c4 CHAR(1));
+INSERT OVERWRITE TABLE lv_table_n1 SELECT 'abc  ', array(1,2,3), 100, 't' FROM src;
 
-CREATE OR REPLACE VIEW lv_view AS SELECT * FROM lv_table; 
+CREATE OR REPLACE VIEW lv_view AS SELECT * FROM lv_table_n1; 
 
 EXPLAIN SELECT myTable.myCol, myTable2.myCol2 FROM lv_view LATERAL VIEW explode(array(1,2,3)) myTable AS myCol LATERAL VIEW explode(array('a', 'b', 'c')) myTable2 AS myCol2 LIMIT 9;

http://git-wip-us.apache.org/repos/asf/hive/blob/9bf28a3c/ql/src/test/queries/clientpositive/lb_fs_stats.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/lb_fs_stats.q b/ql/src/test/queries/clientpositive/lb_fs_stats.q
index 5ad3ff3..a75c411 100644
--- a/ql/src/test/queries/clientpositive/lb_fs_stats.q
+++ b/ql/src/test/queries/clientpositive/lb_fs_stats.q
@@ -7,10 +7,10 @@ set hive.stats.dbclass=fs;
 
 -- INCLUDE_HADOOP_MAJOR_VERSIONS(0.23)
 
-CREATE TABLE test_tab (key STRING, value STRING) PARTITIONED BY (part STRING) STORED AS RCFILE;
+CREATE TABLE test_tab_n0 (key STRING, value STRING) PARTITIONED BY (part STRING) STORED AS RCFILE;
 
-ALTER TABLE test_tab SKEWED BY (key) ON ("484") STORED AS DIRECTORIES;
+ALTER TABLE test_tab_n0 SKEWED BY (key) ON ("484") STORED AS DIRECTORIES;
 
-INSERT OVERWRITE TABLE test_tab PARTITION (part = '1') SELECT * FROM src;
+INSERT OVERWRITE TABLE test_tab_n0 PARTITION (part = '1') SELECT * FROM src;
 
-describe formatted test_tab partition (part='1');
+describe formatted test_tab_n0 partition (part='1');

http://git-wip-us.apache.org/repos/asf/hive/blob/9bf28a3c/ql/src/test/queries/clientpositive/leftsemijoin.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/leftsemijoin.q b/ql/src/test/queries/clientpositive/leftsemijoin.q
index 8974219..07615be 100644
--- a/ql/src/test/queries/clientpositive/leftsemijoin.q
+++ b/ql/src/test/queries/clientpositive/leftsemijoin.q
@@ -1,29 +1,29 @@
 set hive.mapred.mode=nonstrict;
 -- SORT_QUERY_RESULTS
 
-drop table sales;
-drop table things;
+drop table sales_n1;
+drop table things_n1;
 
 set hive.input.format=org.apache.hadoop.hive.ql.io.HiveInputFormat;
 
-CREATE TABLE sales (name STRING, id INT)
+CREATE TABLE sales_n1 (name STRING, id INT)
 ROW FORMAT DELIMITED FIELDS TERMINATED BY '\t';
 
-CREATE TABLE things (id INT, name STRING) partitioned by (ds string)
+CREATE TABLE things_n1 (id INT, name STRING) partitioned by (ds string)
 ROW FORMAT DELIMITED FIELDS TERMINATED BY '\t';
 
-load data local inpath '../../data/files/sales.txt' INTO TABLE sales;
-load data local inpath '../../data/files/things.txt' INTO TABLE things partition(ds='2011-10-23');
-load data local inpath '../../data/files/things2.txt' INTO TABLE things partition(ds='2011-10-24');
+load data local inpath '../../data/files/sales.txt' INTO TABLE sales_n1;
+load data local inpath '../../data/files/things.txt' INTO TABLE things_n1 partition(ds='2011-10-23');
+load data local inpath '../../data/files/things2.txt' INTO TABLE things_n1 partition(ds='2011-10-24');
 
-SELECT name,id FROM sales;
+SELECT name,id FROM sales_n1;
 
-SELECT id,name FROM things;
+SELECT id,name FROM things_n1;
 
-SELECT name,id FROM sales LEFT SEMI JOIN things ON (sales.id = things.id);
+SELECT name,id FROM sales_n1 LEFT SEMI JOIN things_n1 ON (sales_n1.id = things_n1.id);
 
-drop table sales;
-drop table things;
+drop table sales_n1;
+drop table things_n1;
 
 -- HIVE-15458
 explain select part.p_type from part join (select p1.p_name from part p1, part p2 group by p1.p_name) pp ON pp.p_name = part.p_name;

http://git-wip-us.apache.org/repos/asf/hive/blob/9bf28a3c/ql/src/test/queries/clientpositive/leftsemijoin_mr.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/leftsemijoin_mr.q b/ql/src/test/queries/clientpositive/leftsemijoin_mr.q
index c9ebe0e..c1304d9 100644
--- a/ql/src/test/queries/clientpositive/leftsemijoin_mr.q
+++ b/ql/src/test/queries/clientpositive/leftsemijoin_mr.q
@@ -1,20 +1,20 @@
-CREATE TABLE T1(key INT);
-LOAD DATA LOCAL INPATH '../../data/files/leftsemijoin_mr_t1.txt' INTO TABLE T1;
-CREATE TABLE T2(key INT);
-LOAD DATA LOCAL INPATH '../../data/files/leftsemijoin_mr_t2.txt' INTO TABLE T2;
+CREATE TABLE T1_n43(key INT);
+LOAD DATA LOCAL INPATH '../../data/files/leftsemijoin_mr_t1.txt' INTO TABLE T1_n43;
+CREATE TABLE T2_n27(key INT);
+LOAD DATA LOCAL INPATH '../../data/files/leftsemijoin_mr_t2.txt' INTO TABLE T2_n27;
 
 -- Run this query using TestMinimrCliDriver
 
-SELECT * FROM T1;
-SELECT * FROM T2;
+SELECT * FROM T1_n43;
+SELECT * FROM T2_n27;
 
 set hive.auto.convert.join=false;
 set mapred.reduce.tasks=2;
 
 set hive.join.emit.interval=100;
 
-SELECT T1.key FROM T1 LEFT SEMI JOIN (SELECT key FROM T2 SORT BY key) tmp ON (T1.key=tmp.key);
+SELECT T1_n43.key FROM T1_n43 LEFT SEMI JOIN (SELECT key FROM T2_n27 SORT BY key) tmp ON (T1_n43.key=tmp.key);
 
 set hive.join.emit.interval=1;
 
-SELECT T1.key FROM T1 LEFT SEMI JOIN (SELECT key FROM T2 SORT BY key) tmp ON (T1.key=tmp.key);
+SELECT T1_n43.key FROM T1_n43 LEFT SEMI JOIN (SELECT key FROM T2_n27 SORT BY key) tmp ON (T1_n43.key=tmp.key);

http://git-wip-us.apache.org/repos/asf/hive/blob/9bf28a3c/ql/src/test/queries/clientpositive/lineage2.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/lineage2.q b/ql/src/test/queries/clientpositive/lineage2.q
index 9ceafc0..ac8b0a4 100644
--- a/ql/src/test/queries/clientpositive/lineage2.q
+++ b/ql/src/test/queries/clientpositive/lineage2.q
@@ -8,45 +8,45 @@ create table src2 as select key key2, value value2 from src1;
 select * from src1 where key is not null and value is not null limit 3;
 select * from src1 where key > 10 and value > 'val' order by key limit 5;
 
-drop table if exists dest1;
-create table dest1 as select * from src1;
-insert into table dest1 select * from src2;
+drop table if exists dest1_n56;
+create table dest1_n56 as select * from src1;
+insert into table dest1_n56 select * from src2;
 
-select key k, dest1.value from dest1;
+select key k, dest1_n56.value from dest1_n56;
 select key from src1 union select key2 from src2 order by key;
 select key k from src1 union select key2 from src2 order by k;
 
-select key, count(1) a from dest1 group by key;
-select key k, count(*) from dest1 group by key;
-select key k, count(value) from dest1 group by key;
-select value, max(length(key)) from dest1 group by value;
-select value, max(length(key)) from dest1 group by value order by value limit 5;
-
-select key, length(value) from dest1;
-select length(value) + 3 from dest1;
-select 5 from dest1;
-select 3 * 5 from dest1;
-
-drop table if exists dest2;
-create table dest2 as select * from src1 JOIN src2 ON src1.key = src2.key2;
-insert overwrite table dest2 select * from src1 JOIN src2 ON src1.key = src2.key2;
-insert into table dest2 select * from src1 JOIN src2 ON src1.key = src2.key2;
-insert into table dest2
+select key, count(1) a from dest1_n56 group by key;
+select key k, count(*) from dest1_n56 group by key;
+select key k, count(value) from dest1_n56 group by key;
+select value, max(length(key)) from dest1_n56 group by value;
+select value, max(length(key)) from dest1_n56 group by value order by value limit 5;
+
+select key, length(value) from dest1_n56;
+select length(value) + 3 from dest1_n56;
+select 5 from dest1_n56;
+select 3 * 5 from dest1_n56;
+
+drop table if exists dest2_n11;
+create table dest2_n11 as select * from src1 JOIN src2 ON src1.key = src2.key2;
+insert overwrite table dest2_n11 select * from src1 JOIN src2 ON src1.key = src2.key2;
+insert into table dest2_n11 select * from src1 JOIN src2 ON src1.key = src2.key2;
+insert into table dest2_n11
   select * from src1 JOIN src2 ON length(src1.value) = length(src2.value2) + 1;
 
 select * from src1 where length(key) > 2;
 select * from src1 where length(key) > 2 and value > 'a';
 
-drop table if exists dest3;
-create table dest3 as
+drop table if exists dest3_n0;
+create table dest3_n0 as
   select * from src1 JOIN src2 ON src1.key = src2.key2 WHERE length(key) > 1;
-insert overwrite table dest2
+insert overwrite table dest2_n11
   select * from src1 JOIN src2 ON src1.key = src2.key2 WHERE length(key) > 3;
 
-drop table if exists dest_l1;
-CREATE TABLE dest_l1(key INT, value STRING) STORED AS TEXTFILE;
+drop table if exists dest_l1_n0;
+CREATE TABLE dest_l1_n0(key INT, value STRING) STORED AS TEXTFILE;
 
-INSERT OVERWRITE TABLE dest_l1
+INSERT OVERWRITE TABLE dest_l1_n0
 SELECT j.*
 FROM (SELECT t1.key, p1.value
       FROM src1 t1
@@ -59,16 +59,16 @@ FROM (SELECT t1.key, p1.value
       ON (t2.key = p2.key)) j;
 
 drop table if exists emp;
-drop table if exists dept;
-drop table if exists project;
-drop table if exists tgt;
+drop table if exists dept_n10;
+drop table if exists project_n10;
+drop table if exists tgt_n10;
 create table emp(emp_id int, name string, mgr_id int, dept_id int);
-create table dept(dept_id int, dept_name string);
-create table project(project_id int, project_name string);
-create table tgt(dept_name string, name string,
+create table dept_n10(dept_id int, dept_name string);
+create table project_n10(project_id int, project_name string);
+create table tgt_n10(dept_name string, name string,
   emp_id int, mgr_id int, proj_id int, proj_name string);
 
-INSERT INTO TABLE tgt
+INSERT INTO TABLE tgt_n10
 SELECT emd.dept_name, emd.name, emd.emp_id, emd.mgr_id, p.project_id, p.project_name
 FROM (
   SELECT d.dept_name, em.name, em.emp_id, em.mgr_id, em.dept_id
@@ -76,8 +76,8 @@ FROM (
     SELECT e.name, e.dept_id, e.emp_id emp_id, m.emp_id mgr_id
     FROM emp e JOIN emp m ON e.emp_id = m.emp_id
     ) em
-  JOIN dept d ON d.dept_id = em.dept_id
-  ) emd JOIN project p ON emd.dept_id = p.project_id;
+  JOIN dept_n10 d ON d.dept_id = em.dept_id
+  ) emd JOIN project_n10 p ON emd.dept_id = p.project_id;
 
 drop table if exists dest_l2;
 create table dest_l2 (id int, c1 tinyint, c2 int, c3 bigint) stored as textfile;
@@ -105,8 +105,8 @@ group by a.c1, a.c2, a.id, b.c1, b.c2, b.c3
 having count(a.c2) > 0
 order by b.c3 limit 5;
 
-drop table if exists t;
-create table t as
+drop table if exists t_n10;
+create table t_n10 as
 select distinct a.c2, a.c3 from dest_l2 a
 inner join dest_l3 b on (a.id = b.id)
 where a.id > 0 and b.c3 = 15;

http://git-wip-us.apache.org/repos/asf/hive/blob/9bf28a3c/ql/src/test/queries/clientpositive/lineage3.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/lineage3.q b/ql/src/test/queries/clientpositive/lineage3.q
index 735b63f..4633964 100644
--- a/ql/src/test/queries/clientpositive/lineage3.q
+++ b/ql/src/test/queries/clientpositive/lineage3.q
@@ -6,33 +6,33 @@ drop table if exists d1;
 create table d1(a int);
 
 from (select a.ctinyint x, b.cstring1 y
-from alltypesorc a join alltypesorc b on a.cint = b.cbigint) t
+from alltypesorc a join alltypesorc b on a.cint = b.cbigint) t_n20
 insert into table d1 select x + length(y);
 
 drop table if exists d2;
 create table d2(b varchar(128));
 
 from (select a.ctinyint x, b.cstring1 y
-from alltypesorc a join alltypesorc b on a.cint = b.cbigint) t
+from alltypesorc a join alltypesorc b on a.cint = b.cbigint) t_n20
 insert into table d1 select x where y is null
 insert into table d2 select y where x > 0;
 
-drop table if exists t;
-create table t as
+drop table if exists t_n20;
+create table t_n20 as
 select * from
   (select * from
      (select key from src1 limit 1) v1) v2;
 
-drop table if exists dest_l1;
-create table dest_l1(a int, b varchar(128))
+drop table if exists dest_l1_n2;
+create table dest_l1_n2(a int, b varchar(128))
   partitioned by (ds string) clustered by (a) into 2 buckets;
 
-insert into table dest_l1 partition (ds='today')
+insert into table dest_l1_n2 partition (ds='today')
 select cint, cast(cstring1 as varchar(128)) as cs
 from alltypesorc
 where cint is not null and cint < 0 order by cint, cs limit 5;
 
-insert into table dest_l1 partition (ds='tomorrow')
+insert into table dest_l1_n2 partition (ds='tomorrow')
 select min(cint), cast(min(cstring1) as varchar(128)) as cs
 from alltypesorc
 where cint is not null and cboolean1 = true
@@ -115,10 +115,10 @@ where not exists
    where a.key = b.ctinyint + 300)
 and key > 300;
 
-with t as (select key x, value y from src1 where key > '2')
-select x, y from t where y > 'v' order by x, y limit 5;
+with t_n20 as (select key x, value y from src1 where key > '2')
+select x, y from t_n20 where y > 'v' order by x, y limit 5;
 
-from (select key x, value y from src1 where key > '2') t
+from (select key x, value y from src1 where key > '2') t_n20
 select x, y where y > 'v' order by x, y limit 5;
 
 drop view if exists dest_v1;
@@ -174,7 +174,7 @@ alter view dest_v3 as
     where a.cboolean2 = true and b.cfloat > 0
     group by a.ctinyint, a.csmallint, b.cboolean1
     having count(a.cint) > 10
-    order by a, x, b.cboolean1 limit 10) t;
+    order by a, x, b.cboolean1 limit 10) t_n20;
 
 select * from dest_v3 limit 2;
 

http://git-wip-us.apache.org/repos/asf/hive/blob/9bf28a3c/ql/src/test/queries/clientpositive/list_bucket_dml_1.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/list_bucket_dml_1.q b/ql/src/test/queries/clientpositive/list_bucket_dml_1.q
index e38930f..5c25760 100644
--- a/ql/src/test/queries/clientpositive/list_bucket_dml_1.q
+++ b/ql/src/test/queries/clientpositive/list_bucket_dml_1.q
@@ -11,7 +11,7 @@ set mapred.input.dir.recursive=true;
 -- INCLUDE_HADOOP_MAJOR_VERSIONS(0.23)
 
 -- create a skewed table
-create table list_bucketing_dynamic_part (key String, value String) 
+create table list_bucketing_dynamic_part_n0 (key String, value String) 
 partitioned by (ds String, hr String) 
 skewed by (key) on ("484")
 stored as DIRECTORIES
@@ -19,22 +19,22 @@ stored as DIRECTORIES
 
 -- list bucketing DML
 explain extended
-insert overwrite table list_bucketing_dynamic_part partition (ds='2008-04-08', hr) select key, value, hr from srcpart where ds='2008-04-08';
-insert overwrite table list_bucketing_dynamic_part partition (ds='2008-04-08', hr) select key, value, hr from srcpart where ds='2008-04-08';
+insert overwrite table list_bucketing_dynamic_part_n0 partition (ds='2008-04-08', hr) select key, value, hr from srcpart where ds='2008-04-08';
+insert overwrite table list_bucketing_dynamic_part_n0 partition (ds='2008-04-08', hr) select key, value, hr from srcpart where ds='2008-04-08';
 
 -- check DML result
-desc formatted list_bucketing_dynamic_part partition (ds='2008-04-08', hr='11');
-desc formatted list_bucketing_dynamic_part partition (ds='2008-04-08', hr='12');
+desc formatted list_bucketing_dynamic_part_n0 partition (ds='2008-04-08', hr='11');
+desc formatted list_bucketing_dynamic_part_n0 partition (ds='2008-04-08', hr='12');
 
 select count(1) from srcpart where ds='2008-04-08';
-select count(1) from list_bucketing_dynamic_part where ds='2008-04-08';
+select count(1) from list_bucketing_dynamic_part_n0 where ds='2008-04-08';
 
 select key, value from srcpart where ds='2008-04-08' and hr='11' and key = "484";
 set hive.optimize.listbucketing=true;
 explain extended
-select key, value from list_bucketing_dynamic_part where ds='2008-04-08' and hr='11' and key = "484";
-select key, value from list_bucketing_dynamic_part where ds='2008-04-08' and hr='11' and key = "484";
+select key, value from list_bucketing_dynamic_part_n0 where ds='2008-04-08' and hr='11' and key = "484";
+select key, value from list_bucketing_dynamic_part_n0 where ds='2008-04-08' and hr='11' and key = "484";
 
 -- clean up resources
-drop table list_bucketing_dynamic_part;
+drop table list_bucketing_dynamic_part_n0;
 

http://git-wip-us.apache.org/repos/asf/hive/blob/9bf28a3c/ql/src/test/queries/clientpositive/list_bucket_dml_11.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/list_bucket_dml_11.q b/ql/src/test/queries/clientpositive/list_bucket_dml_11.q
index 8ac1627..d3f6ba4 100644
--- a/ql/src/test/queries/clientpositive/list_bucket_dml_11.q
+++ b/ql/src/test/queries/clientpositive/list_bucket_dml_11.q
@@ -10,7 +10,7 @@ set hive.merge.mapredfiles=false;
 -- list bucketing DML: static partition. multiple skewed columns.
 
 -- create a skewed table
-create table list_bucketing_static_part (key String, value String) 
+create table list_bucketing_static_part_n3 (key String, value String) 
     partitioned by (ds String, hr String) 
     skewed by (value) on ('val_466','val_287','val_82')
     stored as DIRECTORIES
@@ -18,19 +18,19 @@ create table list_bucketing_static_part (key String, value String)
 
 -- list bucketing DML without merge. use bucketize to generate a few small files.
 explain extended
-insert overwrite table list_bucketing_static_part partition (ds = '2008-04-08',  hr = '11')
+insert overwrite table list_bucketing_static_part_n3 partition (ds = '2008-04-08',  hr = '11')
 select key, value from src;
 
-insert overwrite table list_bucketing_static_part partition (ds = '2008-04-08', hr = '11')
+insert overwrite table list_bucketing_static_part_n3 partition (ds = '2008-04-08', hr = '11')
 select key, value from src;
 
 -- check DML result
-show partitions list_bucketing_static_part;
-desc formatted list_bucketing_static_part partition (ds='2008-04-08', hr='11');	
+show partitions list_bucketing_static_part_n3;
+desc formatted list_bucketing_static_part_n3 partition (ds='2008-04-08', hr='11');	
 
 set hive.optimize.listbucketing=true;
 explain extended
-select key, value from list_bucketing_static_part where ds='2008-04-08' and hr='11' and value = "val_466";
-select key, value from list_bucketing_static_part where ds='2008-04-08' and hr='11' and value = "val_466";
+select key, value from list_bucketing_static_part_n3 where ds='2008-04-08' and hr='11' and value = "val_466";
+select key, value from list_bucketing_static_part_n3 where ds='2008-04-08' and hr='11' and value = "val_466";
 
-drop table list_bucketing_static_part;
+drop table list_bucketing_static_part_n3;

http://git-wip-us.apache.org/repos/asf/hive/blob/9bf28a3c/ql/src/test/queries/clientpositive/list_bucket_dml_12.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/list_bucket_dml_12.q b/ql/src/test/queries/clientpositive/list_bucket_dml_12.q
index 9facfa5..f929a5e 100644
--- a/ql/src/test/queries/clientpositive/list_bucket_dml_12.q
+++ b/ql/src/test/queries/clientpositive/list_bucket_dml_12.q
@@ -9,7 +9,7 @@ set hive.merge.mapredfiles=false;
 -- SORT_QUERY_RESULTS
 
 -- test where the skewed values are more than 1 say columns no. 2 and 4 in a table with 5 columns
-create table list_bucketing_mul_col (col1 String, col2 String, col3 String, col4 String, col5 string) 
+create table list_bucketing_mul_col_n0 (col1 String, col2 String, col3 String, col4 String, col5 string) 
     partitioned by (ds String, hr String) 
     skewed by (col2, col4) on (('466','val_466'),('287','val_287'),('82','val_82'))
     stored as DIRECTORIES
@@ -17,27 +17,27 @@ create table list_bucketing_mul_col (col1 String, col2 String, col3 String, col4
 
 -- list bucketing DML 
 explain extended
-insert overwrite table list_bucketing_mul_col partition (ds = '2008-04-08',  hr = '11')
+insert overwrite table list_bucketing_mul_col_n0 partition (ds = '2008-04-08',  hr = '11')
 select 1, key, 1, value, 1 from src;
 
-insert overwrite table list_bucketing_mul_col partition (ds = '2008-04-08', hr = '11')
+insert overwrite table list_bucketing_mul_col_n0 partition (ds = '2008-04-08', hr = '11')
 select 1, key, 1, value, 1 from src;
 
 -- check DML result
-show partitions list_bucketing_mul_col;
-desc formatted list_bucketing_mul_col partition (ds='2008-04-08', hr='11');	
+show partitions list_bucketing_mul_col_n0;
+desc formatted list_bucketing_mul_col_n0 partition (ds='2008-04-08', hr='11');	
 
 set hive.optimize.listbucketing=true;
 explain extended
-select * from list_bucketing_mul_col 
+select * from list_bucketing_mul_col_n0 
 where ds='2008-04-08' and hr='11' and col2 = "466" and col4 = "val_466";
-select * from list_bucketing_mul_col 
+select * from list_bucketing_mul_col_n0 
 where ds='2008-04-08' and hr='11' and col2 = "466" and col4 = "val_466";
 
 explain extended
-select * from list_bucketing_mul_col 
+select * from list_bucketing_mul_col_n0 
 where ds='2008-04-08' and hr='11' and col2 = "382" and col4 = "val_382";
-select * from list_bucketing_mul_col 
+select * from list_bucketing_mul_col_n0 
 where ds='2008-04-08' and hr='11' and col2 = "382" and col4 = "val_382";
 
-drop table list_bucketing_mul_col;
+drop table list_bucketing_mul_col_n0;

http://git-wip-us.apache.org/repos/asf/hive/blob/9bf28a3c/ql/src/test/queries/clientpositive/list_bucket_dml_2.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/list_bucket_dml_2.q b/ql/src/test/queries/clientpositive/list_bucket_dml_2.q
index c6dceab..587fec9 100644
--- a/ql/src/test/queries/clientpositive/list_bucket_dml_2.q
+++ b/ql/src/test/queries/clientpositive/list_bucket_dml_2.q
@@ -23,7 +23,7 @@ set hive.stats.reliable=true;
 -- 87 000001_0
 
 -- create a skewed table
-create table list_bucketing_static_part (key String, value String) 
+create table list_bucketing_static_part_n4 (key String, value String) 
     partitioned by (ds String, hr String) 
     skewed by (key, value) on (('484','val_484'),('51','val_14'),('103','val_103'))
     stored as DIRECTORIES
@@ -31,42 +31,42 @@ create table list_bucketing_static_part (key String, value String)
 
 -- list bucketing DML without merge. use bucketize to generate a few small files.
 explain extended
-insert overwrite table list_bucketing_static_part partition (ds = '2008-04-08',  hr = '11')
+insert overwrite table list_bucketing_static_part_n4 partition (ds = '2008-04-08',  hr = '11')
 select key, value from srcpart where ds = '2008-04-08';
 
-insert overwrite table list_bucketing_static_part partition (ds = '2008-04-08', hr = '11')
+insert overwrite table list_bucketing_static_part_n4 partition (ds = '2008-04-08', hr = '11')
 select key, value from srcpart where ds = '2008-04-08';
 
 -- check DML result
-show partitions list_bucketing_static_part;
-desc formatted list_bucketing_static_part partition (ds='2008-04-08', hr='11');	
+show partitions list_bucketing_static_part_n4;
+desc formatted list_bucketing_static_part_n4 partition (ds='2008-04-08', hr='11');	
 
 select count(1) from srcpart where ds = '2008-04-08';
-select count(*) from list_bucketing_static_part;
+select count(*) from list_bucketing_static_part_n4;
 
 set hive.input.format=org.apache.hadoop.hive.ql.io.HiveInputFormat;
 set hive.optimize.listbucketing=true;
 explain extended
-select * from list_bucketing_static_part where ds = '2008-04-08' and  hr = '11' and key = '484' and value = 'val_484';
-select * from list_bucketing_static_part where ds = '2008-04-08' and  hr = '11' and key = '484' and value = 'val_484';
+select * from list_bucketing_static_part_n4 where ds = '2008-04-08' and  hr = '11' and key = '484' and value = 'val_484';
+select * from list_bucketing_static_part_n4 where ds = '2008-04-08' and  hr = '11' and key = '484' and value = 'val_484';
 select * from srcpart where ds = '2008-04-08' and key = '484' and value = 'val_484';
 
 -- 51 and val_51 in the table so skewed data for 51 and val_14 should be none
 -- but query should succeed for 51 or 51 and val_14
 select * from srcpart where ds = '2008-04-08' and key = '51';
-select * from list_bucketing_static_part where key = '51';
+select * from list_bucketing_static_part_n4 where key = '51';
 select * from srcpart where ds = '2008-04-08' and key = '51' and value = 'val_14';
-select * from list_bucketing_static_part where key = '51' and value = 'val_14';
+select * from list_bucketing_static_part_n4 where key = '51' and value = 'val_14';
 
 -- queries with < <= > >= should work for skewed test although we don't benefit from pruning
 select count(1) from srcpart where ds = '2008-04-08' and key < '51';
-select count(1) from list_bucketing_static_part where key < '51';
+select count(1) from list_bucketing_static_part_n4 where key < '51';
 select count(1) from srcpart where ds = '2008-04-08' and key <= '51';
-select count(1) from list_bucketing_static_part where key <= '51';
+select count(1) from list_bucketing_static_part_n4 where key <= '51';
 select count(1) from srcpart where ds = '2008-04-08' and key > '51';
-select count(1) from list_bucketing_static_part where key > '51';
+select count(1) from list_bucketing_static_part_n4 where key > '51';
 select count(1) from srcpart where ds = '2008-04-08' and key >= '51';
-select count(1) from list_bucketing_static_part where key >= '51';
+select count(1) from list_bucketing_static_part_n4 where key >= '51';
 
 -- clean up
-drop table list_bucketing_static_part;
+drop table list_bucketing_static_part_n4;

http://git-wip-us.apache.org/repos/asf/hive/blob/9bf28a3c/ql/src/test/queries/clientpositive/list_bucket_dml_3.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/list_bucket_dml_3.q b/ql/src/test/queries/clientpositive/list_bucket_dml_3.q
index ef01529..d4f9fc9 100644
--- a/ql/src/test/queries/clientpositive/list_bucket_dml_3.q
+++ b/ql/src/test/queries/clientpositive/list_bucket_dml_3.q
@@ -11,23 +11,23 @@ set mapred.input.dir.recursive=true;
 -- INCLUDE_HADOOP_MAJOR_VERSIONS(0.23)
 
 -- create a skewed table
-create table list_bucketing_static_part (key String, value String) partitioned by (ds String, hr String) skewed by (key) on ("484") stored as DIRECTORIES;
+create table list_bucketing_static_part_n1 (key String, value String) partitioned by (ds String, hr String) skewed by (key) on ("484") stored as DIRECTORIES;
 
 -- list bucketing DML
 explain extended
-insert overwrite table list_bucketing_static_part partition (ds='2008-04-08', hr='11') select key, value from srcpart where ds='2008-04-08';
-insert overwrite table list_bucketing_static_part partition (ds='2008-04-08', hr='11') select key, value from srcpart where ds='2008-04-08';
+insert overwrite table list_bucketing_static_part_n1 partition (ds='2008-04-08', hr='11') select key, value from srcpart where ds='2008-04-08';
+insert overwrite table list_bucketing_static_part_n1 partition (ds='2008-04-08', hr='11') select key, value from srcpart where ds='2008-04-08';
 
 -- check DML result
-desc formatted list_bucketing_static_part partition (ds='2008-04-08', hr='11');
+desc formatted list_bucketing_static_part_n1 partition (ds='2008-04-08', hr='11');
 
 select count(1) from srcpart where ds='2008-04-08';
-select count(1) from list_bucketing_static_part where ds='2008-04-08';
+select count(1) from list_bucketing_static_part_n1 where ds='2008-04-08';
 
 select key, value from srcpart where ds='2008-04-08' and hr='11' and key = "484";
 set hive.optimize.listbucketing=true;
 explain extended
-select key, value from list_bucketing_static_part where ds='2008-04-08' and hr='11' and key = "484";
-select key, value from list_bucketing_static_part where ds='2008-04-08' and hr='11' and key = "484";
+select key, value from list_bucketing_static_part_n1 where ds='2008-04-08' and hr='11' and key = "484";
+select key, value from list_bucketing_static_part_n1 where ds='2008-04-08' and hr='11' and key = "484";
 -- clean up resources
-drop table list_bucketing_static_part;
+drop table list_bucketing_static_part_n1;

http://git-wip-us.apache.org/repos/asf/hive/blob/9bf28a3c/ql/src/test/queries/clientpositive/list_bucket_dml_4.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/list_bucket_dml_4.q b/ql/src/test/queries/clientpositive/list_bucket_dml_4.q
index 950409d..48452a3 100644
--- a/ql/src/test/queries/clientpositive/list_bucket_dml_4.q
+++ b/ql/src/test/queries/clientpositive/list_bucket_dml_4.q
@@ -26,7 +26,7 @@ set hive.merge.mapredfiles=false;
 -- 118 000001_0
 
 -- create a skewed table
-create table list_bucketing_static_part (key String, value String) 
+create table list_bucketing_static_part_n2 (key String, value String) 
     partitioned by (ds String, hr String) 
     skewed by (key, value) on (('484','val_484'),('51','val_14'),('103','val_103'))
     stored as DIRECTORIES
@@ -34,39 +34,39 @@ create table list_bucketing_static_part (key String, value String)
 
 -- list bucketing DML without merge. use bucketize to generate a few small files.
 explain extended
-insert overwrite table list_bucketing_static_part partition (ds = '2008-04-08',  hr = '11')
+insert overwrite table list_bucketing_static_part_n2 partition (ds = '2008-04-08',  hr = '11')
 select key, value from srcpart where ds = '2008-04-08';
 
-insert overwrite table list_bucketing_static_part partition (ds = '2008-04-08', hr = '11')
+insert overwrite table list_bucketing_static_part_n2 partition (ds = '2008-04-08', hr = '11')
 select key, value from srcpart where ds = '2008-04-08';
 
 -- check DML result
-show partitions list_bucketing_static_part;
-desc formatted list_bucketing_static_part partition (ds='2008-04-08', hr='11');	
+show partitions list_bucketing_static_part_n2;
+desc formatted list_bucketing_static_part_n2 partition (ds='2008-04-08', hr='11');	
 
 set hive.merge.mapfiles=true;	
 set hive.merge.mapredfiles=true; 
 -- list bucketing DML with merge. use bucketize to generate a few small files.
 explain extended
-insert overwrite table list_bucketing_static_part partition (ds = '2008-04-08',  hr = '11')
+insert overwrite table list_bucketing_static_part_n2 partition (ds = '2008-04-08',  hr = '11')
 select key, value from srcpart where ds = '2008-04-08';
 
-insert overwrite table list_bucketing_static_part partition (ds = '2008-04-08',  hr = '11')
+insert overwrite table list_bucketing_static_part_n2 partition (ds = '2008-04-08',  hr = '11')
 select key, value from srcpart where ds = '2008-04-08';
 
 -- check DML result
-show partitions list_bucketing_static_part;
-desc formatted list_bucketing_static_part partition (ds='2008-04-08', hr='11');	
+show partitions list_bucketing_static_part_n2;
+desc formatted list_bucketing_static_part_n2 partition (ds='2008-04-08', hr='11');	
 
 select count(1) from srcpart where ds = '2008-04-08';
-select count(*) from list_bucketing_static_part;
+select count(*) from list_bucketing_static_part_n2;
 
 set hive.input.format=org.apache.hadoop.hive.ql.io.HiveInputFormat;
 set hive.optimize.listbucketing=true;
 explain extended
-select * from list_bucketing_static_part where ds = '2008-04-08' and  hr = '11' and key = '484' and value = 'val_484';
-select * from list_bucketing_static_part where ds = '2008-04-08' and  hr = '11' and key = '484' and value = 'val_484';
+select * from list_bucketing_static_part_n2 where ds = '2008-04-08' and  hr = '11' and key = '484' and value = 'val_484';
+select * from list_bucketing_static_part_n2 where ds = '2008-04-08' and  hr = '11' and key = '484' and value = 'val_484';
 select * from srcpart where ds = '2008-04-08' and key = '484' and value = 'val_484';
 
 -- clean up
-drop table list_bucketing_static_part;
+drop table list_bucketing_static_part_n2;

http://git-wip-us.apache.org/repos/asf/hive/blob/9bf28a3c/ql/src/test/queries/clientpositive/list_bucket_dml_5.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/list_bucket_dml_5.q b/ql/src/test/queries/clientpositive/list_bucket_dml_5.q
index fce8e2e..47f55cb 100644
--- a/ql/src/test/queries/clientpositive/list_bucket_dml_5.q
+++ b/ql/src/test/queries/clientpositive/list_bucket_dml_5.q
@@ -12,28 +12,28 @@ set mapred.input.dir.recursive=true;
 -- SORT_QUERY_RESULTS
 
 -- create a skewed table
-create table list_bucketing_dynamic_part (key String, value String) 
+create table list_bucketing_dynamic_part_n1 (key String, value String) 
 partitioned by (ds String, hr String) 
 skewed by (key, value) on (('484','val_484'),('51','val_14'),('103','val_103')) 
 stored as DIRECTORIES;
 
 -- list bucketing DML
 explain extended
-insert overwrite table list_bucketing_dynamic_part partition (ds='2008-04-08', hr) select key, value, hr from srcpart where ds='2008-04-08';
-insert overwrite table list_bucketing_dynamic_part partition (ds='2008-04-08', hr) select key, value, hr from srcpart where ds='2008-04-08';
+insert overwrite table list_bucketing_dynamic_part_n1 partition (ds='2008-04-08', hr) select key, value, hr from srcpart where ds='2008-04-08';
+insert overwrite table list_bucketing_dynamic_part_n1 partition (ds='2008-04-08', hr) select key, value, hr from srcpart where ds='2008-04-08';
 
 -- check DML result
-desc formatted list_bucketing_dynamic_part partition (ds='2008-04-08', hr='11');
-desc formatted list_bucketing_dynamic_part partition (ds='2008-04-08', hr='12');
+desc formatted list_bucketing_dynamic_part_n1 partition (ds='2008-04-08', hr='11');
+desc formatted list_bucketing_dynamic_part_n1 partition (ds='2008-04-08', hr='12');
 
 select count(1) from srcpart where ds='2008-04-08';
-select count(1) from list_bucketing_dynamic_part where ds='2008-04-08';
+select count(1) from list_bucketing_dynamic_part_n1 where ds='2008-04-08';
 
 select key, value from srcpart where ds='2008-04-08' and key = "103" and value ="val_103";
 set hive.optimize.listbucketing=true;
 explain extended
-select key, value, ds, hr from list_bucketing_dynamic_part where ds='2008-04-08' and key = "103" and value ="val_103";
-select key, value, ds, hr from list_bucketing_dynamic_part where ds='2008-04-08' and key = "103" and value ="val_103";
+select key, value, ds, hr from list_bucketing_dynamic_part_n1 where ds='2008-04-08' and key = "103" and value ="val_103";
+select key, value, ds, hr from list_bucketing_dynamic_part_n1 where ds='2008-04-08' and key = "103" and value ="val_103";
 
 -- clean up resources
-drop table list_bucketing_dynamic_part;
+drop table list_bucketing_dynamic_part_n1;

http://git-wip-us.apache.org/repos/asf/hive/blob/9bf28a3c/ql/src/test/queries/clientpositive/list_bucket_dml_6.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/list_bucket_dml_6.q b/ql/src/test/queries/clientpositive/list_bucket_dml_6.q
index 631c938..0678602 100644
--- a/ql/src/test/queries/clientpositive/list_bucket_dml_6.q
+++ b/ql/src/test/queries/clientpositive/list_bucket_dml_6.q
@@ -49,7 +49,7 @@ set hive.merge.mapredfiles=false;
 -- SORT_QUERY_RESULTS
 
 -- create a skewed table
-create table list_bucketing_dynamic_part (key String, value String) 
+create table list_bucketing_dynamic_part_n3 (key String, value String) 
     partitioned by (ds String, hr String) 
     skewed by (key, value) on (('484','val_484'),('51','val_14'),('103','val_103'))
     stored as DIRECTORIES
@@ -57,42 +57,42 @@ create table list_bucketing_dynamic_part (key String, value String)
 
 -- list bucketing DML without merge. use bucketize to generate a few small files.
 explain extended
-insert overwrite table list_bucketing_dynamic_part partition (ds = '2008-04-08', hr)
+insert overwrite table list_bucketing_dynamic_part_n3 partition (ds = '2008-04-08', hr)
 select key, value, if(key % 100 == 0, 'a1', 'b1') from srcpart where ds = '2008-04-08';
 
-insert overwrite table list_bucketing_dynamic_part partition (ds = '2008-04-08', hr)
+insert overwrite table list_bucketing_dynamic_part_n3 partition (ds = '2008-04-08', hr)
 select key, value, if(key % 100 == 0, 'a1', 'b1') from srcpart where ds = '2008-04-08';
 
 -- check DML result
-show partitions list_bucketing_dynamic_part;
-desc formatted list_bucketing_dynamic_part partition (ds='2008-04-08', hr='a1');	
-desc formatted list_bucketing_dynamic_part partition (ds='2008-04-08', hr='b1');
+show partitions list_bucketing_dynamic_part_n3;
+desc formatted list_bucketing_dynamic_part_n3 partition (ds='2008-04-08', hr='a1');	
+desc formatted list_bucketing_dynamic_part_n3 partition (ds='2008-04-08', hr='b1');
 
 set hive.merge.mapfiles=true;	
 set hive.merge.mapredfiles=true; 
 -- list bucketing DML with merge. use bucketize to generate a few small files.
 explain extended
-insert overwrite table list_bucketing_dynamic_part partition (ds = '2008-04-08', hr)
+insert overwrite table list_bucketing_dynamic_part_n3 partition (ds = '2008-04-08', hr)
 select key, value, if(key % 100 == 0, 'a1', 'b1') from srcpart where ds = '2008-04-08';
 
-insert overwrite table list_bucketing_dynamic_part partition (ds = '2008-04-08', hr)
+insert overwrite table list_bucketing_dynamic_part_n3 partition (ds = '2008-04-08', hr)
 select key, value, if(key % 100 == 0, 'a1', 'b1') from srcpart where ds = '2008-04-08';
 
 -- check DML result
-show partitions list_bucketing_dynamic_part;
-desc formatted list_bucketing_dynamic_part partition (ds='2008-04-08', hr='a1');	
-desc formatted list_bucketing_dynamic_part partition (ds='2008-04-08', hr='b1');
+show partitions list_bucketing_dynamic_part_n3;
+desc formatted list_bucketing_dynamic_part_n3 partition (ds='2008-04-08', hr='a1');	
+desc formatted list_bucketing_dynamic_part_n3 partition (ds='2008-04-08', hr='b1');
 
 select count(1) from srcpart where ds = '2008-04-08';
-select count(*) from list_bucketing_dynamic_part;
+select count(*) from list_bucketing_dynamic_part_n3;
 
 set hive.input.format=org.apache.hadoop.hive.ql.io.HiveInputFormat;
 set hive.optimize.listbucketing=true;
 explain extended
-select * from list_bucketing_dynamic_part where key = '484' and value = 'val_484';
-select * from list_bucketing_dynamic_part where key = '484' and value = 'val_484';
+select * from list_bucketing_dynamic_part_n3 where key = '484' and value = 'val_484';
+select * from list_bucketing_dynamic_part_n3 where key = '484' and value = 'val_484';
 select * from srcpart where ds = '2008-04-08' and key = '484' and value = 'val_484';
 
 -- clean up
-drop table list_bucketing_dynamic_part;
+drop table list_bucketing_dynamic_part_n3;